Recent comments in /f/deeplearning

sweeetscience t1_iv0q0xr wrote

This should fail since the original work is not being redistributed. To wholly recreate a repo on which Codex was trained you’d have to literally start typing the original code, and even then the contextual suggestions would likely yield a different result from the original anyways. I could be mistaken but I remember reading about some litigation in this space concerning a model trained on copyrighted data. The court ruled in favor of the defendant because the resulting model couldn’t possibly reproduce the original work. It’s tricker here because technically you could recreate the original work, but you would have to know very well what the original work was to begin with to actually recreate it, and if that’s the case what’s the point of using copilot to begin with. I could be (and probably am) wrong.

Imagine trying to recreate PyTorch from scratch using Codex or copilot. IF, and that’s a big if, one did so the author of the recreation would still have to attribute it.

Not legal advice

5

px05j OP t1_iv0hl5e wrote

I believe there could be other models which will fall in this category, image generation models specially.

This particular lawsuit is interesting as it says it violates github's own terms.

0

suflaj t1_iuyg2am wrote

Well I couldn't understand what your task was when you didn't say what it was until now.

Other than that, skimming through the paper it quite clearly says the following:

> Our present results do not indicate our procedure can generalize to motifs that are not present in the training set

Because what they're doing doesn't generalize, I think the starting assumptions (that there will be imprevements with a larger model) are wrong, and so the question is unnecessary... The issue is with the method or the data, they do not elaborate more than that.

2

suflaj t1_iuybshu wrote

That seems very bad. You want your train-dev-test to be different samples of the same distribution, so, not very different sets.

Furthermore, if you're using test for model validation, that means you will have no dataset to finally evaluate your model on. Reconsider your process.

Finally, again, I urge you to evaluate your dataset on an established evaluation metric for the task, not the loss you use to train the model. What is the exact task?

2

suflaj t1_iuya2f9 wrote

It can be seen as an approximation of the variance between the noise and the noise predicted conditioned on some data.

If it's on the training set it is not even usable as a metric, and if it is not directly related to the performance it is not a good metric. You want to see how it acts on unseen data.

1

Current-Basket3920 t1_iuxezmx wrote

The code is freely available and on Youtube you find videos on „How to use Stylegan“.

But:

  • As already mentioned you‘ll need more data. 5-10k are already at the low end. You need more like 50-100k I think.
  • You need some serious hardware to train something like this. For Stylegan they used 8 high end GPUs for a week. I guess for Stylegan2/3 even more. You might be happy with less than NVIDIA, but it‘s no quick thing to do.
  • And even then there‘s no guarantee that it will turn out perfectly. It might need some fine-tuning, it‘s not a simple algorithm.
2