Recent comments in /f/MachineLearning

i_am__not_a_robot t1_je4fv3w wrote

Also, demanding that "AI research and development should be refocused on making today's powerful, state-of-the-art systems more accurate, safe, interpretable, transparent, robust, aligned, trustworthy, and loyal" is a defamatory implication that researchers do not currently have these goals.

1

i_am__not_a_robot t1_je4er0i wrote

The demand for corporate self-restraint the face of enormous profit opportunities is naïve and doomed to fail. And this is without even addressing the merits of the open letter's demands, which I - and many other computer scientists - do not support, by the way.

1

Calamero t1_je4doo0 wrote

It will enable creative people to bring their ideas to reality. It won’t make people less creative. AI technology democratizes the execution part, making it easier for people from all walks of life to transform their visions into reality. It will augment human creativity rather than stifling it.

1

sdmat t1_je4bgwh wrote

Exactly, it's bizarre to point to revealing failure cases for a universal approximator then claim that fixing those failure cases in later versions would be irrelevant.

Entirely possible that GPT3 only did interpolation and fails horribly out of domain and that GPT5 will infer the laws of nature, language, psychology, logic, etc and be able to apply them to novel material.

It certainly looks like GPT4 is somewhere in between.

4

DragonForg t1_je4ascu wrote

This is a scam, or something else. I really do not know. But I don't know how all these famous people can get together in like one day, and state we need to slow progress of the next technological craze. Even if it leads to our doom, I doubt this many tech people would even realize it.

2

Narabedla t1_je4abcc wrote

Quite frankly, yes all jobs that can be reliably automated, should be.

But with that should come an universal basic income funded from the increase in productivity. The gain from automation shouldnt be capital gain for the few at top, but a freedom gain for the general population.

1

Haycart t1_je4923c wrote

>Yes, ChatGPT is doing much more than querying text! It is not just a query engine on a giant corpus of text. … Duh! I do not think you should only think of ChatGPT as a query engine on a giant corpus of text. There can be a lot of value in reasoning about ChatGPT anthropomorphically or in other ways. RLHF also complicates the story, as over time it weighs responses away from the initial training data. But “query engine on a giant corpus of text” should be a non-zero part of your mental model because, without it, you cannot explain many of the things ChatGPT does.

The author seems to present this bizarre dichotomy, that either you have to think of ChatGPT as a query engine or you have to think of it in magical/mystical/anthropomorphic terms.

(They also touch on viewing ChatGPT as a function on the space of "billion dimensional" embeddings. This is closer to the mark but seems to conflate the model's parameter count with the dimensionality of its latent space, which doesn't exactly inspire confidence in the author's level of understanding.)

Why not just think of ChatGPT as what it is--a very large transformer?

The fact that a model like ChatGPT is able to do what it does is not at all surprising, IMO, when you consider the following facts:

  1. Transformers (and neural networks in general) are universal approximators. A sufficiently large neural network can approximate any function to arbitrary precision (with a few minor caveats).
  2. Neural networks trained with stochastic gradient descent benefit from implicit regularization -- SGD naturally tends to seek out simple solutions that generalize well. Furthermore, larger neural networks appear to generalize better than smaller ones.
  3. The recent GPTs have been trained on a non-trivial fraction of the entire internet's text content.
  4. Text on the internet (and language data in general) arises from human beings interacting with the world--reasoning, thinking, and emoting about those interactions--and attempting to communicate the outcome of this process to one another.

Is it really crazy to imagine that the simplest possible function capable of fitting a dataset as vast as ChatGPT's, might resemble the function that produced it? A function that subsumes, among other things, human creativity and reasoning?

In another world, GPT 3 or 4 might have turned out to be incapable of approximating that function to any notable degree of fidelity. But even then, it wouldn't be outlandish to imagine that one of the later members of the GPT family could eventually succeed.

5

fripperML OP t1_je43ftw wrote

Yes, I don't know what to think honestly. I've read with amusement this paper (well, some of the examples, not all because I did not have time to finish it):

https://arxiv.org/abs/2303.12712

It's very optimistic, and alligned with what you say (not an incremental improvement from previous models).

But then, besides the article I shared, I've read this thread:

https://www.reddit.com/r/MachineLearning/comments/124eyso/n_openai_may_have_benchmarked_gpt4s_coding/

So I don't know... Probably we will see soon, when access to GPT-4 is more spread.

Thanks for commenting :)

3

sdmat t1_je42icd wrote

> So long as it’s a transformer model, GPT-4 will also be a query engine on a giant corpus of text, just with more of the holes patched up, so it’d be harder to see the demonstrative examples of it being that.

This claim has a strong scent of sophistry about it - any and all signs of intelligence can be handwaved away as interpolating to plausible text.

The explanations of failures are convincing, but the theory needs to go further and explain why larger models like GPT4 (and in some cases 3.5) are so much more effective at answering out-of-domain queries with explicit reasoning proceding from information that it does have. E.g 4 correctly answers the weights question and gives a clear explanation of its reasoning. And that isn't an isolated example.

It's not just an incremental improvement, there is a clear difference in kind.

13