What to do when you really ARE out of memory?
情報
この質問は閉じられています。 編集または回答するには再度開いてください。
古いコメントを表示
What is the solution for optimizing code when you really are just working with too large of a dataset?
Currently I need to perform triscatteredinterp using 3 vectors all (100,000,000 x 1).
scatteredInterpolant does not work any better in this instance.
0 件のコメント
回答 (3 件)
the cyclist
2015 年 8 月 4 日
編集済み: the cyclist
2015 年 8 月 4 日
2 投票
For very large datasets, processing a random sample of the data will often give satisfactory results.
0 件のコメント
Walter Roberson
2015 年 8 月 4 日
1 投票
Store the data in hierarchies such as octrees that allow you to extract a subset that fits within working memory to do the fine-grained work on.
0 件のコメント
この質問は閉じられています。
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!