Recent comments in /f/deeplearning
2thleZ t1_j4ooqax wrote
Reply to comment by myth_drannon in Is doing TensorFlow certificate worth the time and effort? by MACKBULLERZ
I've heard a lot of people saying the same thing but could you explain why PyTorch is better than TF?
elf7979 OP t1_j4o933j wrote
Reply to comment by BellyDancerUrgot in Is 100 mega byte text corpus big enought to train? by elf7979
I will check Gensim documentation. Thank you
elf7979 OP t1_j4o90u3 wrote
Reply to comment by suflaj in Is 100 mega byte text corpus big enought to train? by elf7979
I think trascript from company's conference call includes some certain characterstics since business professionals may use some particular verbs or expressions. I haven't checked out w2v datasets you mentioned yet. Is there existing corpus that's business-oriented?
​
What if dataset size increases to 1 giga bytes. Is it big enough?
BellyDancerUrgot t1_j4nw2ku wrote
Reply to comment by suflaj in Is 100 mega byte text corpus big enought to train? by elf7979
Gensim documentation itself has them highlighted along with the necessary arguments to use to download and use them.
Moises-Tohias t1_j4mw2nc wrote
Reply to comment by agentfuzzy999 in Retrieve voice from noisy audio file by BackgroundPass2082
VAD won't do it, since the speech and the noise overlap
bhargavkartik t1_j4mqdsf wrote
Reply to comment by suflaj in Is 100 mega byte text corpus big enought to train? by elf7979
This.
SometimesZero t1_j4mo6ra wrote
Reply to comment by myth_drannon in Is doing TensorFlow certificate worth the time and effort? by MACKBULLERZ
Thanks for the info. I knew PyTorch was popular, but didn’t know about TF’s waning popularity. Good to know.
myth_drannon t1_j4mmfha wrote
Reply to comment by SometimesZero in Is doing TensorFlow certificate worth the time and effort? by MACKBULLERZ
TF out, Pytorch in. Google always struggled to build a long lasting ecosystems for their products. Angular was first but then Facebook came and swooped in with React, same happened with TF and pytorch
[deleted] t1_j4mho46 wrote
Reply to comment by SometimesZero in Is doing TensorFlow certificate worth the time and effort? by MACKBULLERZ
PyTorch is where it’s at !
SometimesZero t1_j4mgn69 wrote
Reply to comment by myth_drannon in Is doing TensorFlow certificate worth the time and effort? by MACKBULLERZ
Can you say more about “on its way out?” Are other libraries more popular in your opinion?
SupremeChampionOfDi t1_j4lwzff wrote
Reply to comment by Legitimate-Gold-8711 in Help with deep learning project "autocorrection" by Legitimate-Gold-8711
This how a Chinese person I know speaks.
Legitimate-Gold-8711 OP t1_j4l9swg wrote
Reply to comment by SupremeChampionOfDi in Help with deep learning project "autocorrection" by Legitimate-Gold-8711
Which reasons :D
SupremeChampionOfDi t1_j4l4r6u wrote
I read this in a funny Chinese accent for some reason.
suflaj t1_j4l1i5l wrote
Likely not enough, at least not for what is considered good. But I fail to see why you'd want to trian it yourself, there are plenty of readily available w2v weights or vocabularies.
myth_drannon t1_j4l048h wrote
TF is on its way out. But in general get certificates for fundamentals not for frameworks that change every couple of years. Certificates are popular in ops domain, so if you are planning on doing mlops then get something related to that.
Trick2206 t1_j4kkijl wrote
Imo not really, from what I've heard and seen most jobs don't really care for random certs from courses. You can just learn everything the course will teach from free resources so you can just search online for the topics you're interested in.
Legitimate-Gold-8711 OP t1_j4kfxo9 wrote
Reply to comment by thatoneboii in Help with deep learning project "autocorrection" by Legitimate-Gold-8711
I tried levenshtein distance algorithm but it's not works like I want
thatoneboii t1_j4jf8h5 wrote
Do you absolutely need to use deep learning? There are tons of way faster autocorrect implementations that use levenshtein distances and non-DL techniques such as SymSpell or Norvig’s algorithm. DL is complicated, expensive, and requires tons of data to train on - I would stay away from that unless you’re doing it for your own enrichment or a school project.
shmollerup t1_j4htyot wrote
You could try something that works on a character level, like a sequence tobsequence model, or maybe a rnn approach like char2vec. Both approaches should work pretty good if you have enough training data
tsgiannis t1_j4fcuo3 wrote
I meant on the betting site
ivan_kudryavtsev t1_j4c5702 wrote
Reply to comment by Infamous_Age_7731 in Cloud VM GPU is much slower than my local GPU by Infamous_Age_7731
Ram performance also may be affected by meltdown, spectre patches.
vagartha OP t1_j4c3wer wrote
Reply to comment by tsgiannis in Building an NBA game prediction model - failing to improve between epochs by vagartha
Simulate? How would I go about doing that? Sorry if that's a silly/involved question, but I'm not sure how I would simulate NBA games.
tsgiannis t1_j4c3d7c wrote
You can always simulate
vagartha OP t1_j4c36zw wrote
Reply to comment by tsgiannis in Building an NBA game prediction model - failing to improve between epochs by vagartha
Haha, I live in CA so sports gambling so that's out of the question...
I was actually hoping to maybe write a paper or something and submit it to something like the Sloane conference or send it in to 538 as an add-on to my resume?
Also, my model uses data from seasons going back all the way to 2014 as of right now. Larger datasets would make a better model, right? So why not use more historical data?
suflaj t1_j4pdd6o wrote
Reply to comment by elf7979 in Is 100 mega byte text corpus big enought to train? by elf7979
You're closer but not yet quite there - the smaller Google News Dataset W2V is trained on is 10 GB. The full one used is around 300GB IIRC