Recent comments in /f/deeplearning
xolotl96 t1_j9izg4o wrote
The proce of 4 rtx4090 is very high. If you plan on spending that much money I suppose it is for business or research. It could make sense to invest in a server grade processor and motherboard with support to many more pcie lanes. Also, servers often use multiple power supplies for redundancy, but in this case they can be helpful for managing the wattage of 4 cards. In my experience limiting the max power does not impact training time in a dramatic way, so I would do that especially if you are planning to air cool them (which is probably the best thing for an offsite server)
suflaj t1_j9iwxuj wrote
You would have to limit the power to 250W. It will overheat without an open case. PCI-E 3 x8 means you are cutting the cards bandwidth in half.
Overall a terrible idea.
suflaj t1_j9iwkp8 wrote
Reply to comment by buzzz_buzzz_buzzz in Bummer: nVidia stopping support for multi-gpu peer to peer access with 4090s by mosalreddit
They removed NVLink to cut down on costs on a feature that is supported by only a handful of games in existence and otherwise useless.
thelibrarian101 t1_j9ivql0 wrote
PCIe extenders always worry me, they tend to catch fire very easilly. There is a large amount of current flowing through them after all
hayAbhay t1_j9i9nfv wrote
Reply to comment by nail_nail in Fine tuning a GPT for text generation by nashcaps2724
If you have the hardware, and if you have a lot of those input-output examples, you can use alternative smaller models in the gpt family.
Should work reasonably well especially if the variance in the input-output isn't too much. (A lot depends on your dataset here)
Definitely tradeoffs here in terms of model dev, inference and maintenance of it. If the expected costs aren't too high, I'd strongly recommend gpt3 as a base.
[deleted] t1_j9i8tj7 wrote
Reply to Bummer: nVidia stopping support for multi-gpu peer to peer access with 4090s by mosalreddit
[deleted]
buzzz_buzzz_buzzz t1_j9hjtqx wrote
Reply to Bummer: nVidia stopping support for multi-gpu peer to peer access with 4090s by mosalreddit
Not surprising given that they removed NVLink to force multi-gpu users to purchase A6000s
nail_nail t1_j9gkc5w wrote
Reply to comment by hayAbhay in Fine tuning a GPT for text generation by nashcaps2724
That means it will need to be paid for each summarization API call, forever, right? Is there an alternative model that one can tune on a couple of high end nvidia cards? Like GPT NeoX?
hayAbhay t1_j9duoda wrote
Reply to Fine tuning a GPT for text generation by nashcaps2724
Create a Corpus C like this
<source text from corpus A> <human generated text from corpus B> . . .
Make sure you add some unique tokens marking the start and end of each example and the input and output within it.
Then, take any pretrained LLM (tuning gpt3 is trivial with ~10-20 lines of code).
For inference, use the tuned model and give it the input and let it complete the output. You can add the "end" marker token to get generation to complete.
[Source: trained/tuned several language models including gpt3]
MugiwarraD t1_j9djpux wrote
Reply to comment by JurgenSchmidthuber in I trained a neural network. Now what? by Ricenaros
lol
pitrucha t1_j9cdugi wrote
Reply to comment by nashcaps2724 in Fine tuning a GPT for text generation by nashcaps2724
more like: given item x from Corpus B, predict item y from corpus A
nashcaps2724 OP t1_j9cdelh wrote
Reply to comment by pitrucha in Fine tuning a GPT for text generation by nashcaps2724
Just to make sure I’m following, steps here could be…
- Finetune GPT2 or GPT3 on Corpus A
- Take the finetuned GPT* and continue as a summarization task?
pitrucha t1_j9ccuun wrote
Reply to Fine tuning a GPT for text generation by nashcaps2724
proceed as with summarization. you can train gpt family using open ai or get yourself t5 and train it. training from scratch will be a waste of resources
[deleted] t1_j9bn3lp wrote
Reply to comment by junetwentyfirst2020 in Entry to a career in deep learning by No-Celebration6994
[removed]
junetwentyfirst2020 t1_j9bn2hn wrote
Reply to comment by No-Celebration6994 in Entry to a career in deep learning by No-Celebration6994
Look into computer vision courses in university websites and you should see the range
No-Celebration6994 OP t1_j9bkj7z wrote
Reply to comment by junetwentyfirst2020 in Entry to a career in deep learning by No-Celebration6994
My instinct says perception, but I can’t say I’m all too familiar with the full pipeline. Do you have any resources where I could read up on it? Thanks so much
Nerveregenerator t1_j99gc4f wrote
Reply to HOW DO I KNOW THE STEP BY STEP PROCESSES INVOLVED IN SOUND CLASSIFICATION USING CNN. Can anyone help me by pointing me towards the rght direction, be it a paid course or whatever.... I need to know how to implement it for my project work. I would really appreciate your help. Thanks by Illustrious-Force-74
You just use mfcc and then it’s just like image detection
mreyesag t1_j91hr5l wrote
Reply to I trained a neural network. Now what? by Ricenaros
Once trained (and happy w the testing results)…make it fail :) find perturbations and difficult inputs so you understand the limits of your models you can now focus on robustness
a_man_of_music t1_j91clm6 wrote
Reply to I trained a neural network. Now what? by Ricenaros
Grid search to reduce the search space on the hyperparameters, then random search on the reduced search space.
PaleontologistDue620 t1_j90z1yo wrote
Reply to comment by [deleted] in My Neural Net is stuck, I've run out of ideas by [deleted]
no you're not annoying at all, thanks for the update :)
[deleted] OP t1_j90ysce wrote
Reply to comment by PaleontologistDue620 in My Neural Net is stuck, I've run out of ideas by [deleted]
Sorry to be annoying but I thought it was nice to give you some news as well. I was confused as to why there isnt yolo in pytorch, here it is why https://github.com/pytorch/vision/issues/6341
Morteriag t1_j90stfo wrote
Reply to I trained a neural network. Now what? by Ricenaros
Put it to test with real data. Putting a lot of effort into tuning a model on a fixed data set that will eventually be deployed is a waste of time. And dont freak out when it fails! Just add more quality data from when it is deployed.
JurgenSchmidthuber t1_j8zf6gn wrote
Reply to I trained a neural network. Now what? by Ricenaros
Rub it in oil and herbs and place it in the oven on broil for 10-12 minutes
Oceanboi t1_j8zdely wrote
Reply to comment by crimson1206 in Physics-Informed Neural Networks by vadhavaniyafaijan
Oh I see, I missed the major point that the training data is basically incomplete to model the entire relationship.
Why embed priors into neural networks, doesn’t Bayesian Modeling using MCMC do pretty much what this is attempting to do? We did something similar to this in one of my courses although we didn’t get to spend enough time on it so forgive me if my questions are stupid. I also would need someone to walk me through a motivating example for a PINN because I’d just get lost in generalities otherwise. I get the example, but am failing to see the larger use case.
jcoffi t1_j9jcsha wrote
Reply to Is my plan for a 4x RTX 3090 Machine Learning rig feasible? by [deleted]
P2P is disabled on the 4090s so if you needed that expect degraded performance