Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

tts onnx gpu inference time problems #70

Open
1nlplearner opened this issue Dec 2, 2022 · 2 comments
Open

tts onnx gpu inference time problems #70

1nlplearner opened this issue Dec 2, 2022 · 2 comments

Comments

@1nlplearner
Copy link

hi, i met a problem when onnx inference on gpu

  1. onnx inference on gpu slower than onnx cpu inference much time and sometimes faster than gpu pt inference(2 times acceleration)
  2. when i inference same text twice or more, inference achieves 2 time acceleration compare to gpu pt inference
    any advicec?
    thanks
@Masao-Someki
Copy link
Collaborator

@1nlplearner
At the first execution, onnxruntime takes a longer time for inference. So please skip the first execution.

@1nlplearner
Copy link
Author

@1nlplearner At the first execution, onnxruntime takes a longer time for inference. So please skip the first execution.
i think this is the reason, [Some nodes were not assigned to the preferred execution providers which may or may not have an negative impact on performance. e.g. ORT explicitly assigns shape related ops to CPU to improve perf.]
how to solve it?
thanks

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants