RTX 3090 vs A100 in deep learning.

57 ビュー (過去 30 日間)
지환 오
지환 오 2022 年 5 月 4 日
コメント済み: Joss Knight 2022 年 5 月 10 日
I ran ResNet on RTX 3090 and A100
Performance is better in RTX 3090 about 1.2 times than A100
I searched and found out that GPU Coder helps use TensorCore
So, I want to be sure if I use GPU Coder, A100's performance is going to better before I purchasing GPU Coder
Thanks.

採用された回答

Joss Knight
Joss Knight 2022 年 5 月 6 日
According to the spec as documented on Wikipedia, the RTX 3090 has about 2x the maximum speed at single precision than the A100, so I would expect it to be faster. The A100 is much faster in double precision than the GeForce card.
Both will be using Tensor Cores for deep learning in MATLAB.
  4 件のコメント
Kyle Lee
Kyle Lee 2022 年 5 月 10 日
For understanding, I ask you to confirm.
Then, If other models like Alexnet, Googlenet etc.. are used , do these models automatically use Tensor Core?
Joss Knight
Joss Knight 2022 年 5 月 10 日
Yes.

サインインしてコメントする。

その他の回答 (1 件)

David Willingham
David Willingham 2022 年 5 月 4 日
Hi,
What version of MATLAB did you test ResNet out on? I'd recommend running benchmarks on the latest version of MATLAB.
Was it for inference or training?
AS an FYI, you can contact MathWorks and receive a trial of GPU Coder to test out the performance first hand.

カテゴリ

Help Center および File ExchangeImage Data Workflows についてさらに検索

タグ

製品

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by