Solution of large sparse matrix systems using GPU MLDIVIDE
古いコメントを表示
I have a sparse 1 million by 1 million matrix system that I wish to solve repeatedly in a loop, (I use MLDIVIDE or the ' \ '). My CPU takes about 250 s to solve this once and the RAM goes up to 40 GB during the process. This has made me doubt that, Is it possible to solve this system on a 4 GB GPU, using the GPU MLDIVIDE? Will solving it on a GPU make it faster? Or it does not make sense? I have read that GPU is good for highly parallel operations. I have GeForce GTX 1050 Ti 4 GB GPU. My CPU is i7-9700 with 3 GHz, 8 cores, and 64 GB RAM.
採用された回答
その他の回答 (1 件)
Edric Ellis
2020 年 1 月 16 日
0 投票
A couple of suggestions:
- On the CPU, if you're repeatedly solving the same system, you might be able to benefit from the recently-introduced decomposition object.
- On the GPU, it's hard to say without knowing the exact details whether or not the GPU will be of benefit in this case, so perhaps it would be best to get a Parallel Computing Toolbox trial licence to enable you to experiment.
1 件のコメント
Yash Agrawal
2020 年 1 月 16 日
編集済み: Yash Agrawal
2020 年 1 月 16 日
カテゴリ
ヘルプ センター および File Exchange で Parallel Computing についてさらに検索
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!