Recent comments in /f/deeplearning

Constant-Cranberry29 OP t1_irreagz wrote

Do you mean I need size up in this part?

def ResNet50Regression():

`Res_input = layers.Input(shape=(178,))`

# 128

`width = 128`



`x = dens_block(Res_input,width)`

`x = identity_block(x,width)`

`x = identity_block(x,width)`



`x = dens_block(x,width)`

`x = identity_block(x,width)`

`x = identity_block(x,width)`



`x = dens_block(x,width)`

`x = identity_block(x,width)`

`x = identity_block(x,width)`



`x = layers.BatchNormalization()(x)`

`x = layers.Dense(1,activation='linear')(x)`

`model = models.Model(inputs=Res_input, outputs=x)`



`return model`
1

_Arsenie_Boca_ t1_irrcmwh wrote

If all you want to see is the two curves close to each other, I guess you could size up the model, so that it overfits terribly. But is that really desirable?

If my assumption that you predict the whole graph autoregressively is correct, then I believe it works just fine. You should check the forecast horizon and think about what it is you want to achieve in the end

1

_Arsenie_Boca_ t1_irr9bb4 wrote

No this is not a modelling issue. It actually isnt a real issue at all. Predicting a very long trajectory is simply very hard. At each timestep, a slight error will occur which will exponentiate, even if the error per timestep is marginal. Imagine being asked to predict a certain stock price. Given some expertise and current information, you might be able to do it for tomorrow, but can you do it precisely for the next year?

1

Constant-Cranberry29 OP t1_irr8eo4 wrote

>I guess this is timeseries forecasting. You should think about the lookahead. Probably, during training, the model only has to predict the next point, while during testing, it has to predict many values autoregressively

what should I do? should I change the model structure or use another model?

1

braintampon t1_irow1qf wrote

Reply to comment by DaltonSC2 in 1080 vs 2060 for deeplearning by ccppoo0

I agree. I was joking myself when i built myself a pc for deep learning couple years ago. Selling it off within this month to get a macbook and will invest in colab/sagemaker.

dm if any1 is lookin to buy lol

0

suflaj t1_irmuzmw wrote

2060 if it's the 8 GB version, otherwise 1080. But obviously the best would be to not waste your money and just buy cloud compute... With these cards being 200-300$, that is roughly 200-300 hours of A100 training, which is much faster and enables you to train much larger models.

EDIT: I see you've gotten a 110$ offer for 1080, I'd say go with that. You'll be severely limited in what you can run but after you learn your lesson you can still sell it for 100$, or even more.

2

zero_one_memrisor t1_irmq3x3 wrote

While I agree with the general statement that tensor core are better to have, memory on device is a bigger limiting factor in the long run.

Depending on the model, deep learning can require a significant amount of memory for training. I would suggest you save a little more and go for a 2070 (8GB) or if you can swing it the 3070 with 8GB.

(Edited with correction on the amount of memory available for the 3070)

6