Recent comments in /f/deeplearning
[deleted] t1_jbhjv7j wrote
I don’t get the last question when V(x_t) = 1 means that beta_t = 1
Why the confusion?
itsnotachickennugget t1_jbhiuv0 wrote
Reply to comment by manoj_mjb in Meta’s LLaMa weights leaked on torrent... and the best thing about it is someone put up a PR to replace the google form in the repo with it 😂 by RandomForests92
It's still working,
I started downloading it 4 hours ago and it has finished a few minutes ago.
Puzzleheaded_You5821 t1_jbh3zy7 wrote
Delta8Girl t1_jbgpyz5 wrote
silva_p t1_jbgl59f wrote
Reply to comment by ThunderGodSeed in Meta’s LLaMa weights leaked on torrent... and the best thing about it is someone put up a PR to replace the google form in the repo with it 😂 by RandomForests92
That link has been working fine for me. You're probably just copying wrong. Here:
magnet:?xt=urn:btih:ZXXDAUWYLRUXXBHUYEMS6Q5CE5WA3LVA&dn=LLaMA
Mods let me know if this is not allowed.
ThunderGodSeed t1_jbg44s0 wrote
Reply to comment by manoj_mjb in Meta’s LLaMa weights leaked on torrent... and the best thing about it is someone put up a PR to replace the google form in the repo with it 😂 by RandomForests92
Let me know if you get a link
Electronic-Clerk868 OP t1_jbfz12k wrote
Reply to comment by cadoi in How can i improve my model in order to get more accuray and less loss?? Thanks by Electronic-Clerk868
thanks!
cadoi t1_jbfvfg8 wrote
Reply to How can i improve my model in order to get more accuray and less loss?? Thanks by Electronic-Clerk868
Work on your ability to communicate?
activatedgeek t1_jbfjafv wrote
From (1), it looks like V(x_t) is the conditional variance of x_t given x_{t-1} (for the forward process defined by q).
mr_birrd t1_jbfe325 wrote
Reply to comment by wutheringsouls in what exactly is Variance(Xt) during the Forward Process in Diffusion model ? by eugene129
yes t is the timestep index
wutheringsouls t1_jbf9vd1 wrote
Reply to comment by mr_birrd in what exactly is Variance(Xt) during the Forward Process in Diffusion model ? by eugene129
Do you mean at different time steps?
mr_birrd t1_jbf8sdi wrote
The variance of your sample x_t, simple as that.
[deleted] t1_jbdgwpj wrote
[deleted]
[deleted] t1_jbd46q1 wrote
[deleted]
LetMeGuessYourAlts t1_jbbw4d4 wrote
Reply to comment by GhostingProtocol in Trying to figure out what GPU to buy... by GhostingProtocol
I just bought a used 3090 for $740 on eBay before tax. I view my GPU as a for-fun expenditure. Part of that is ML stuff. For the cost of a handful of new release videogames, you can go from 10gb to 24gb and do a lot of cool stuff. There's going to be increasingly less state of the art stuff that fits in 10gb comfortably.
MatureKit t1_jbb8bgo wrote
Reply to newby here. looking for help on a MLP for speech recognition. any tips or pointers would be appreciated by alexilas
Check out the WaveNet paper for some ideas about this!
dfcHeadChair t1_jbb2dyi wrote
Reply to comment by alexilas in newby here. looking for help on a MLP for speech recognition. any tips or pointers would be appreciated by alexilas
Yep feel free
alexilas OP t1_jbazarg wrote
Reply to comment by dfcHeadChair in newby here. looking for help on a MLP for speech recognition. any tips or pointers would be appreciated by alexilas
Thanks!! I really appreciate. I really like the ai world and if it's not too much to ask, if you have anything else you would recommend me to go further I would appreciate it. Again thanks!!
dfcHeadChair t1_jbau8dy wrote
Reply to comment by alexilas in newby here. looking for help on a MLP for speech recognition. any tips or pointers would be appreciated by alexilas
If you’re only detecting speech, that is doable with heuristics and some napkin math, or an MLP, for simple cases. However, “detect speech in this audio” is rarely the end of the story in the real world. Next up comes transcription, sentiment analysis, tonal feature flagging, etc. all of which are currently dominated by Transformers. You’ll also see some great work in the RNN space, but Transformer-based architectures are king right now.
Some models for inspiration, https://huggingface.co/models?pipeline_tag=automatic-speech-recognition&sort=downloads
alexilas OP t1_jban3hs wrote
Reply to comment by dfcHeadChair in newby here. looking for help on a MLP for speech recognition. any tips or pointers would be appreciated by alexilas
Thanks for that link!! But out of curiosity, what would you use instead of a MLP?
dfcHeadChair t1_jbaljm8 wrote
Reply to newby here. looking for help on a MLP for speech recognition. any tips or pointers would be appreciated by alexilas
MLP for speech recognition probably isn't a great solution, but if it's for a class and you can only use numpy start here: https://towardsdatascience.com/coding-a-neural-network-from-scratch-in-numpy-31f04e4d605
suflaj t1_jb92dey wrote
Reply to comment by GhostingProtocol in Trying to figure out what GPU to buy... by GhostingProtocol
Well to be honest, unless there's some particular reason why you need the GPUs locally, the most cost effective solution is to just run it in the cloud.
Having GPUs locally is mostly a luxury for when some contract prevents you from using the cloud, or you need to train something every day for several hours over a year or more. For everything else, cloud pay-as-you-go will be cheaper and faster.
GhostingProtocol OP t1_jb928dg wrote
Reply to comment by suflaj in Trying to figure out what GPU to buy... by GhostingProtocol
Yeah, I found out about gpu bus width a few hours ago. At least I’be learned a lot about GPUs haha
Buying a 3090 when my current 6gb laptop gpu stops working for my purposes.
suflaj t1_jb91xiw wrote
The 3060s are pretty gimped for DL. Currently, you either go for a 3090/4090, or used 1080Ti/2080Ti(s) if on a budget.
Ferocious_Armadillo t1_jbhura1 wrote
Reply to How can i improve my model in order to get more accuray and less loss?? Thanks by Electronic-Clerk868
Increase the learning rate of your optimizer. Something else you could try— look into “back propagation” to tune your learning rate after each layer.