GPU Coder vs. ONNXRuntime, is there a difference in inference speed?
2 ビュー (過去 30 日間)
古いコメントを表示
Since I can export from Matlab to ONNX format, why can't I just import my model into TensorRT etc.? Will I get significant speed increases or is the benefit of GPU Coder more about being able to compile all my other Matlab code into optimized Cuda?
Thanks in advance.
0 件のコメント
回答 (1 件)
Joss Knight
2021 年 4 月 2 日
You can compile your network for TensorRT using GPU Coder if that's your intended target, no need to go through ONNX.
I don't believe MathWorks have any published benchmarks against ONNX runtime specifically. GPU Coder on the whole outperforms other frameworks, although it does depend on the network.
2 件のコメント
Matti Kaupenjohann
2022 年 1 月 7 日
Could you show/link the benchmark which includes the performance of gpucoder against other frameworks (which one?).
Joss Knight
2022 年 1 月 7 日
編集済み: Joss Knight
2022 年 1 月 7 日
We don't publish the competitive benchmarks, you'll have to make a request through your sales agent. we can provide some numbers for MATLAB.
参考
カテゴリ
Help Center および File Exchange で Deep Learning Toolbox についてさらに検索
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!