Recent comments in /f/deeplearning

qiltb t1_j3rtr6a wrote

Be sure to check logs (i.e. dmesg for starters). Many A100s on AWS for example suffer from memory corruptions which leads to severe degradation in performance. Also check temps.

A single A100 (even the least capable one - 400W with 40GB) should be more of a level of 3090Ti.

You also need to check memory usage (if it's on a limit - like 78.9/80 - there's a problem somewhere). Also don't exclude drivers.

Those are some common headaches when setting up remote GPU instances for DL...

1

VinnyVeritas t1_j3rrzvr wrote

Actually I've been sort of looking at ML computers (kind of of browsing and dreaming one day I would have one, but it's always going to be out of my means and needs anyway). Anyway, they can put two PSUs in a box, obviously it's made by companies, so the total cost is twice or 3 times the cost of the parts alone (e.g. building yourself would be 2-3x cheaper) but it could inspire you for picking your parts.https://bizon-tech.com/amd-ryzen-threadripper-up-to-64-cores-workstation-pc

https://shop.lambdalabs.com/gpu-workstations/vector/customize

1

ASalvail t1_j3r5kxy wrote

A statistical model. I'm personally partial to using the ETS model (error, trend, seasonality), but a SARIMAX is also another good one. The 'issue' with a stats model will be that you need to do some hand tuning and thus need to understand how the model works (and ETS is a fairly simple one to comprehend).

2

Infamous_Age_7731 OP t1_j3qy6qv wrote

> multithreaded
>
>sysbench
>
> to compare CPU and RAM

Thanks a lot for your input! I checked the CPU %steal it seems optimal ranging from 0.0 to 0.1st. Then, I don't think it's a shard since in the NVIDIA I have the full 80Gb memory at my disposal (unless they do some trickeries). I did a series of `sysbench tests and I found out that the VM's CPU is slightly worse for single-thread performance, but what is more astounding is the RAM speed. For 1 or 8 threads the write is 0.8x slower and the read is 1.5x slower. The Ram speed drop seems to reflect the iteration per second speed drop when I train the model. I guess this might be the fault.

2

trajo123 t1_j3qwftm wrote

13 periods as history to forecast another 13, this seems like a very atypical/extreme TS forecasting problem, do these services actually handle so little data?

First, it's unlikely that this little data is enough for anything but the simplest models. Probably the best you could do in terms of a domain independent model is linear regression. Even so calculating performance metrics - knowing how good the model is - is going to be challenging as that would require you to further reduce the amount of training data in order to have a validation/"out of sample" set.

Getting useful predictions with so little data is probably going to require you to make a model with strong assumptions - e.g. come up with a set of domain-specific parametrized equations that govern the time-series and then fit those parameters to the data.

In any case, Deep Learning is far from the first approach that comes to mind trying to solve this problem. Solving this problem is probably just a few lines of code using R or scipy.stats + sklearn, probably less than calling the cloud API functions. The trick is to use the right mathematical model.

2

ASalvail t1_j3pvlls wrote

You don't have enough data to use AI, you're likely just going to overfit the series. In fact, time series are usually fairly short which led to the whole forecasting community to erroneously think ML could never be used for forecasting (see the M4 competition). Statistical is the way to go in your case.

If you absolutely want ML, use a simple random forest library.

3

ivan_kudryavtsev t1_j3pssyp wrote

Why so? GPUs are passed to VM in a pass-through mode, so no significant performance pitfails must happen. I recommend OP to look at CPU %steal, nvidia-smi (maybe it is A100 1/7 shard, not a full GPU). Run a single and multithreaded sysbench to compare CPU and RAM. Also, PCI-E generation or deficated bandwidth can be outperforming on your hardware if a cloud provider uses a not well-balanced custom build.

2

DustinEwan t1_j3pnwfz wrote

Time series is an extremely complex problem unless there's clear periodicity to the data.

In this case linear regression may be a better approach than deep learning.

3

shrshk7 t1_j3pf6di wrote

I’ve been wanting to do the same, I did some research over past few days and seems like this can be achieved in tensorflow js, for the amount of data you mentioned I think you can train the model in the browser

1

yesterdaymee OP t1_j3pdlfo wrote

Yes, I only need to detect persons with some tracking id to them and if there are no persons in the frame "no persons in the room" must be printed. And also if the person if the person wears a mask print "with mask" else "without" mask. I know if no classes were detected then i need to make some change. I just couldnt understand where the class is being finalised. Could you help me ? (For a video input)

1