-
Notifications
You must be signed in to change notification settings - Fork 347
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Is there any pre-trained model? #59
Comments
Hi, you can use a EC2 GPU amazon machines to train your models. Anyway, you can start playing on amazon EC2 instances for now, since you guess your machine is not sufficient. |
Hi friends, Any uploads of trained data would be wonderful! Be it torrent, dropbox, googledrive, all works for me. |
well, I have just opened a similar issue. |
Getting OpenCL working is a absolute pain on a mobile hybrid graphics system. I've given up on it and will build a desktop with an Nvidia card later that has proper support unlike my current hardware. Atm I'm just training on my cpu with parameters: th train.lua --dataset 30000 --hiddenSize 1000 --maxEpoch 10 Current terminal output: http://pastebin.com/RfhqKHNd |
Nice but won't that consume alot of power, I think you should've tried some of aws instances, they could've taken much less time i think. |
Thats what I thought too, until I looked at the costs. I signed up with AWS but its going to cost me 30usd for a month's dev subscription, and then $0.65/hr for just one gpu's worth of power. If I want to run that for 3 days thats $76.8 all up. A week would cost 139usd. |
hehe, I never looked at it from that perspective. I thought your pc consumed more power than it would cost on aws . |
Its processing epoch 9 atm, I'm playing around chatting with epoch 8 and its looking rather promising. That thing has absolutely watched too many cop movies that is clear haha.
EDIT: all uploaded. Links as above. Remember to rename the model_epoch_8.y7 to model.t7 (I could rename it but the connection is crap don't really want to reupload) |
Nice, keep up the good work. |
what specs does you pc have btw, it looks like a nice machine ? |
Thanks :-) I'm planning on building a desktop pc with a GTX1060 6GB (the 1070 looks tempting but its a bit pricey). Once I got that I'll see if I can crunch a full set of data and will upload it. This stuff is fascinating. |
i'm also eyeing for a better desktop, am currently on an amd x4 b93 with 2.8ghz, currently it's the best machine i've used so far for my projects. |
You should drop a graphics card into what you have and it should work great. Training really should be done on GPU's rather than CPU. Good reading: http://timdettmers.com/2017/03/19/which-gpu-for-deep-learning/ |
I was dissapointed to discover that am on a propriety powersupply which is limited to 300watts I saw a |
https://www.dropbox.com/s/9xc9fisil1xee69/model.t7.zip?dl=0 I can continue to post good model results if people are interested. Bought a GTX1060 6GB a few days ago, more parts arriving in the mail.. Will be able to grind more data, much much quicker :) |
@TTN- Thank you for sharing the model. I downloaded and start making conversation but it really does not make a good conversation. Following is a snippet of the conversation. Could you please share a snippet of a good conversation that you had? `you> hi you> how are you? you> Which dress are you talking about? you> Are you robot? you> Are you human? you> Who are you? you> Don't die you> What are you talking about dude? |
@bienbinod I didn't have very good conversations with it sorry. It's mostly been the same as what you say there.. Its been trained on too small a data set I think. Once I get that new gaming PC running (atm I'm limited to cpu training on my laptop, takes forever) I'll train it on the full dataset and see how that goes . I'll share that too. |
@TTN- Thank you for your reply. Same here, I am also relying on a laptop and it is going to take ages to train the model on full dataset. Let's keep on posting the models, whoever builds first. |
I'll also have a play around with this: https://github.com/mtanana/torchneuralconvo |
let me download, this thanks guys for sharing and making this available, maybe we should make a repo with trained models (.t7) detaling the number of epochs trained cpu info and other details. |
by the way @TTN- share cpu loadable versions when you can. |
@kenkit Are gpu trained models cpu loadable? I'll continue to share what I make progress on. This pc build is going to be a at least a week maybe 2 away. I'm waiting for the Ryzen 5 cpu launch on april 11 to finish the build. |
Not a bad idea to have a repo. Probably best to share .torrents or even magnet links or put it on the piratebay.The github limit is 1GB and my dropbox only has so much space. Google drive will hold 15GB. |
@TTN- they are loadable you just need to load the model then convert it to cpu loadable with |
Cheers, thanks @kenkit for the tip. I only program in C and python. th is a bit different, lots to learn. I'm still building that pc. Got a new ryzen 5 cpu on the 11th of april on the launch, just waiting for it to arrive in the snail mail (it got delayed for some reason). Should be here tomorrow. More trained data sets to be posted once I get things up and running. I haven't forgotten about this :-) |
@kenkit could you post a file for me to run to convert to float? I'm no good with lua sorry. I'm trying a couple things but the resulting file is 4 bytes in size. Pretty sure the model is float as is. Uploading new trained data now. stats:
Full terminal output paste (for stats n stuff): https://pastebin.com/c6pJcCCP hardware:
Took me a while to get it trained up to this point, my ryzen system was unstable for a while and crashed a bunch of times, but thats fixed now. Interestingly, the program is heavily CPU constrained. It maxes out a single core (of the 12) and gets limited to that while the GPU sits mostly idle, even though I was training with --cuda. Video memory usage sat at around 2.5GB most of the time (I have the full movie dataset loaded with no limits on vocabulary). The perplexity (ppl) was decreasing fast up to this point, from here on, I think more training will just result in over fitting. |
just load the model normally, then convert the loaded model to float and save as you would any other model. If don't manage to convert it. Ping me I'll build a complete working code which you can use. |
Try this, just put it where we have train.lua
|
Sweet. Thanks @kenkit I did that, tested the results, but throws error when testing with th eval.lua:
|
you might want to check if the files actually exist.
You are currently the only hope we have at getting some working files, i've managed to get a laptop that should put me back to programming though it's not fast enough |
I had trained ages ago and acquired some 35mb file
Also u should know that, after changing from cpu to gpu or vice versa u must first delete the generated files as they will not be useable |
did you try my code ? |
My computer is not very powerful , can someone or could give a link where we can download a pre-trained model?
The text was updated successfully, but these errors were encountered: