开发者

needs guidance on distributing data on multiple GPUs

I'm currently developing a machine learning toolkit for GPU clusters. I tested logistic regression classifier on multiple GPUs.

I'm using a Master-Worker approach , where a master CPU creates several POSIX threads and matrices are divided among GPUs.

But the problem I have is how to store large matrices which can't be stored on a single machine. Ar开发者_StackOverflow中文版e there any libraries or approaches to share data among nodes?


I'm not sure how big are your matrices but you should check CUDA 4.0 that was released a couple of weeks ago. One of the main features is shared memory across multiple CUDA devices/GPUs

0

上一篇:

下一篇:

精彩评论

暂无评论...
验证码 换一张
取 消

最新问答

问答排行榜