-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Polygraphy [HostToDeviceCopy]requires bool I/O but node can not be handled by Myelin. #1698
Comments
Can you run the model with trtexec and attach the log? a log with --verbose is preferred. |
Are you getting errors when running with polygraphy? your trtexec log seems running successfully. try running with the same config as you did with trtexec and don't mark any outputs. |
Yes, I only get errors when running with polygraphy. Running it without mark is ok. But I need to find the tensors that are not right compared with onnx model. |
Marking all tensors as output will break layer fusion hence may affect final output, I recommend making output with bisect until you find the problematic layer. |
Thank you! I finally find the issue layers. Will file another bug for the accuraccy problem. |
Hello,I want to know how to make output with bisect ? Can I directly use polygraphy or use onnx-graphsurgeon to modify the .onnx ? |
I got the way from google |
Thank you!在 2022年9月27日,15:06,safehumeng ***@***.***> 写道:
Marking all tensors as output will break layer fusion hence may affect final output, I recommend making output with bisect until you find the problematic layer.
Hello,I want to know how to make output with bisect ? Can I directly use polygraphy or use onnx-graphsurgeon to modify the .onnx ?
I got the way from google
polygraphy debug precision net_bs8.onnx --fp16 --tactic-sources cublas --check polygraphy run polygraphy_debug.engine --trt --load-outputs onnx_res.json --abs 1e-1
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you commented.Message ID: ***@***.***>
|
@zerollzeng, How can I get difference for all layers?
It detects a big difference in the output layer, which is logical and understandable, but why doesn't it go on to check the other layers? Log
When I set precision to int8, I also get diff only for 'output' layer, and I don't why it tell that precision of both float32? Command:
Log
|
@pranavm-nvidia I haven't sue |
after the debug process, how should we use the reply.json file to build our mixed-precision engine? |
Description
Environment
TensorRT Version: 8.2.1
NVIDIA GPU: V100
NVIDIA Driver Version:
CUDA Version: 11.4
CUDNN Version:
Operating System: ubuntu
Python Version (if applicable):
Tensorflow Version (if applicable):
PyTorch Version (if applicable): 1.10
Baremetal or Container (if so, version):
Relevant Files
Model: https://nvidia-my.sharepoint.com/:u:/p/slyned/ESC-DizzoztLhNsFrQAW_F4BRwGMj7YZAYwteGhvOxxi3A?e=A2Gsb1
Please use nvidia account to get access to it.
Steps To Reproduce
The text was updated successfully, but these errors were encountered: