Recent comments in /f/MachineLearning

olmec-akeru t1_jck223w wrote

Yeah, totally right—and I understand that the specifics really matter in some cases (for example calculating a starship trajectory).

What intrigues me, is that in ideas of concept, of logic, this specific error isn't meaningful. i.e. if the sum of three primes was initially correct the approach wouldn't be invalid. There is something in this.

1

ml_head t1_jcjyon2 wrote

I'm sure that it does. And would beca better demo of the technology. Maybe, keep the Cinderella story too, since some people wouldn't read your original story and wouldn't be able to tell if the summary is good. You might want to add an image with your original story in a format that wouldn't be easy to OCR, like using weird font on noisy background. In this way you are making the story available to humans but taking measures to hide it from any web crawler used by language models.

1

JaCraig t1_jcjx7lx wrote

Genuine question: Why are you trying to use a language model to do something that you could write a basic app to calculate?

Like I would have asked it to write an app in JavaScript, Java, C#, etc. Some popular language to calculate four perfect cubes to represent a number. That'd probably get me 90% of the way there then I'm just fixing a couple bugs. That seems like the more intuitive use case to me but I'm also a dev by trade.

0

SQLGene t1_jcjx3a6 wrote

The title is "[D] GPT-4 is really dumb" and what would be accurate is "[D] GPT-4 is bad at math problems". This was a known issue with GPT 3.5 and I expect it to continue to be an issue. but I think it's a mischaracterization to say it's "dumb" when there are a number of non-mathematical applications where it's impressive so far.

So while my statement was a simplification, I stand by the intention. You are evaluating a tool based on an application I don't think it's meant for.

2

olmec-akeru t1_jcjw333 wrote

Right, so ignoring the specific error and thinking about the general approach: adding a^3 is a fourth term; and it happens that a = 0.

Sneaky, but not illogical.

Edit: the above is wrong, read the thread below for OPs insights.

1

Single_Blueberry t1_jcjvh6o wrote

Again, can't find a reliable source for that.

I personally doubt that GPT-4 is significantly larger than GPT 3.x, simply because that would also further inflate inference cost, which you generally want to avoid in a product (as opposed to a research feat).

Better architecture, better RLHF, more and better train data, more train compute? Seems all reasonable.

Orders of magnitudes larger again? Don't think so.

2