Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Multiple GPU support #26

Open
vikram-gupta opened this issue May 16, 2016 · 8 comments
Open

Multiple GPU support #26

vikram-gupta opened this issue May 16, 2016 · 8 comments

Comments

@vikram-gupta
Copy link

vikram-gupta commented May 16, 2016

Hi Friends,

Do we have the support to run training on multiple GPUs to save time? I am having a machine with 4 GPUs but looks like only one of the GPU is being utilised.

Also, when i am trying to train over complete dataset using the following command, it is taking around 9 hours per epoch. Is this time expected or am i doing something wrong here?

th train.lua --cuda --dataset 0 --hiddenSize 1000

Thanks

@jamesweb1
Copy link

For your reference, I train for whole dataset about 5 hours per pooch in Nvidia GTX980.

@vikram-gupta
Copy link
Author

vikram-gupta commented May 18, 2016

Thanks @jamesweb1 for that. I am more or less on same timelines !

Do you think that using multiple GPUs would help us in bringing the training time down, considerably. Right now, with 50 epochs per training and 5-6 hours per epoch, its taking too much time (~300 hrs) making it difficult to try out and find the best parameters (batchSize,hidden layer size, dataset size) for the task. Do you have any recommendations here?

@jamesweb1
Copy link

Yes, It takes a lot of time. So that I try to train the small subset only(perhaps dataset = 20000). In these experiments, I can obtain the better parameters, and then extend to the whole dataset. I'd like to train on multiple GPUs, but I don't have another resources now. :(

@svenwoldt
Copy link

Could we not pool experimentation and report results somewhere, as to avoid double work?
Create a todo with to be validated parameter settings and I will pick and contribute findings!

@schwittlick
Copy link

That would be great, did anybody start some statistics/benchmarks already?

@HadleRS
Copy link

HadleRS commented Jun 13, 2016

I am also looking at adding multiple GPU support, has anyone had any progress yet?

@macournoyer
Copy link
Owner

@svenwoldt that's a great idea! However, we don't have a good metric for measuring the quality of the model yet. #38 adds a validation set, maybe adding a tests set would be the best way to do this?

RE multiple GPU support. I'm not sure how this could be done and I only have 1 GPU at my disposal. So I'd need help on this :)

@KristenMoore
Copy link

Did anyone make any progress on this? I'm also looking for a multi-GPU solution.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants