Recent comments in /f/deeplearning

LetMeGuessYourAlts t1_jbbw4d4 wrote

I just bought a used 3090 for $740 on eBay before tax. I view my GPU as a for-fun expenditure. Part of that is ML stuff. For the cost of a handful of new release videogames, you can go from 10gb to 24gb and do a lot of cool stuff. There's going to be increasingly less state of the art stuff that fits in 10gb comfortably.

2

dfcHeadChair t1_jbau8dy wrote

If you’re only detecting speech, that is doable with heuristics and some napkin math, or an MLP, for simple cases. However, “detect speech in this audio” is rarely the end of the story in the real world. Next up comes transcription, sentiment analysis, tonal feature flagging, etc. all of which are currently dominated by Transformers. You’ll also see some great work in the RNN space, but Transformer-based architectures are king right now.

Some models for inspiration, https://huggingface.co/models?pipeline_tag=automatic-speech-recognition&sort=downloads

2

suflaj t1_jb92dey wrote

Well to be honest, unless there's some particular reason why you need the GPUs locally, the most cost effective solution is to just run it in the cloud.

Having GPUs locally is mostly a luxury for when some contract prevents you from using the cloud, or you need to train something every day for several hours over a year or more. For everything else, cloud pay-as-you-go will be cheaper and faster.

1