Recent comments in /f/MachineLearning
andreichiffa t1_jdvojfg wrote
Reply to comment by shanereid1 in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
It's a common result from the domain of flat minima - to train the model needs to be overparametrized to avoid getting stuck in local minima and to smooth the loss landscape.
However the overparameterization at the training stage can be trimmed at the inference stage.
WokeAssBaller t1_jdvodfn wrote
Reply to comment by lambertb in [D] GPT4 and coding problems by enryu42
Yeah I don’t buy a survey, could be heavily biased
tvetus t1_jdvo16b wrote
Reply to comment by brierrat in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
The most dangerous liar is the one that tells mostly truths.
alexmin93 t1_jdvo0ls wrote
Reply to My ChatGPT Chrome Extension that saves conversations in .md files is finally approved by the Chrome Web Store. It's still and so will continue to be Open Source. [P] by ThePogromist
How can you do it without openai plugin access? Or you mean it just logs the chat into a file and nothing else? Can it provide context to the gpt?
alexmin93 t1_jdvnm53 wrote
Reply to comment by i_am__not_a_robot in My ChatGPT Chrome Extension that saves conversations in .md files is finally approved by the Chrome Web Store. It's still and so will continue to be Open Source. [P] by ThePogromist
It sounds similar to programmer (programist) in russian
lambertb t1_jdvngk3 wrote
WokeAssBaller t1_jdvmmfp wrote
Reply to comment by lambertb in [D] GPT4 and coding problems by enryu42
I don’t even roll yet but that 40% number, I would love to see how they calculated it.
I’ve tried gpt 4 on a lot of problems and it fails 9/10 times and I would be faster just googling it.
This stuff will be amazing it’s just not quite yet
LifeScientist123 t1_jdvmkkx wrote
Reply to comment by WarmSignificance1 in [D] GPT4 and coding problems by enryu42
>Part of intelligence is the ability to learn in an efficient manner.
Agree to disagree here.
A young deer (foal?) learns to walk 15 minutes after birth. Human babies on average take 8-12 months. Are humans dumber than deer? Or maybe human babies are dumber than foals?
Intelligence is extremely poorly defined. If you look at the scientific literature it's a hot mess. I would argue that intelligence isn't as much about efficiency as it's about two things,
- Absolute performance on complex tasks
AND
- Generalizability to novel situations
If you look at LLMs, they perform pretty well on both these axes.
-
GPT-4 has human level performance in 20+ coding languages AND 20+ human languages on top of being human level/super human in some legal exams, medical exams, AP chemistry, biology, physics etc etc. I don't know many humans that can do all of this.
-
GPT-4 is also a one-shot/ few-shot learner on many tasks.
Avastor_Neretal t1_jdvmgow wrote
Reply to comment by sEi_ in My ChatGPT Chrome Extension that saves conversations in .md files is finally approved by the Chrome Web Store. It's still and so will continue to be Open Source. [P] by ThePogromist
Your mental problems and inability of understanding concept of the other cultures memes and jargonism isn't my problem.
You've proven that you're not only unable to follow wiki links, but also can't use translator which is integrated into a damn browser.
And all of that for the purpose of... of what? To prove me that reddit is full of vocal minorities which are being "oppressed" even by their own reflection? Lol, like if it's not well known fact.
KungFuScubaMaster t1_jdvmcyw wrote
Just adding, I'm also very interested in this!
Ok-Hunt-5902 t1_jdvm7kp wrote
Reply to comment by SkinnyJoshPeck in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
It’s as much an encyclopedia as any.. outdated/incorrect info is ubiquitous in them. What op shows here is ChatGPTs potential to show more accuracy now and in future iterations.
[deleted] t1_jdvlyva wrote
[deleted]
tt54l32v t1_jdvlsg3 wrote
Reply to comment by WarAndGeese in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
Ok so how does one simplify that? Also why does it have to be separate? Genuinely curious.
Ok-Hunt-5902 t1_jdvklga wrote
Reply to comment by Cool_Abbreviations_9 in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
Wild. Nice work!
jabowery OP t1_jdvkjt0 wrote
Reply to comment by 1stuserhere in [D] Definitive Test For AGI by jabowery
Imputation can make interpolation appear to be extrapolation.
So, to fake AGI's capacity for accurate extrapolation (data efficiency), one may take a big pile of money and throw it at expanding the training set to infinity and expanding the matrix multiplication hardware to infinity. This permits more datapoints within which one may interpolate over a larger knowledge space.
But it is fake.
If, on the other hand, you actually understand the content of Wikipedia (the Hutter Prize's very limited, high quality corpus), you may deduce (extrapolate) the larger knowledge space through the best current mathematical definition of AGI: AIXI's where the utility function of the sequential decision theoretic engine is to minimize the algorithmic description of the training data (Solomonoff Induction) used as the prediction oracle in the AGI.
tamilupk OP t1_jdvk3xs wrote
Reply to comment by LifeScientist123 in [D] Will prompting the LLM to review it's own answer be any helpful to reduce chances of hallucinations? I tested couple of tricky questions and it seems it might work. by tamilupk
Yeah humans tend to do that, but llms seems to be a bit better than humans in this. As someone replied to this post even OpenAI used this kind of technique to reduce toxicity/ hallucinations.
SkinnyJoshPeck t1_jdvk16j wrote
Reply to comment by Cool_Abbreviations_9 in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
This is an important thing I've been telling everyone I can about - people talk about how GPT kills education because someone can just ask for a paper and never do the work themselves to learn.
This is a language model, not an encyclopedia, or a quantitative machine, or some other use. It fakes sources; it has no concept of right/wrong or truth vs untruth. It doesn't reason between sources.
The beauty of it is, frankly, it's ability to mimic (at this point) a pseudo-intellectual, haha. Kids are going to turn in papers sourced like they talked to their conspiracy theory uncle, and it will be the "watermark" of AI written papers. It can't reason, it can't generate opinions, thus it can't write a paper. We're long from that (if we could ever get there anyways).
jabowery OP t1_jdvjoi8 wrote
Reply to comment by Deep-Station-1746 in [D] Definitive Test For AGI by jabowery
Information quality may be measured in terms of its signal to noise ratio. Now, agreed, too dense a signal may appear to be noise to some audiences and this is part of the art of writing. However, an advantage of interactive media as opposed to, say, a book, is that the audience is present -- hence [D] is possible. What I've presented to you is, while not understandable to the general audience as signal, is, nevertheless, profoundly true. It may therefore be a good starting point for [D].
[deleted] OP t1_jdvjggt wrote
Reply to comment by Specific-Arrival-127 in [R] BLLR "Budd´s Logistic Linear Regression" - A hybrid algorithm for neural network by [deleted]
[removed]
Specific-Arrival-127 t1_jdvip42 wrote
Reply to comment by [deleted] in [R] BLLR "Budd´s Logistic Linear Regression" - A hybrid algorithm for neural network by [deleted]
When a new approach is proposed, it's common to evaluate its performance against already established algorithms, like simple linear or lasso regression in this case. I'm sorry, but I don't see how your algorithm fares against the baselines, I don't even see any established real-world regression datasets. I would advise you to look into that and test your model in a more controlled manner.
SmellElectronic6656 t1_jdvigcv wrote
Reply to [D] Can we train a decompiler? by vintergroena
Sorry if this is a very basic question. But what are the applications of such a model? What can we do with the decompiled code?
killerfridge t1_jdvid7z wrote
Reply to comment by tamilupk in [D] Will prompting the LLM to review it's own answer be any helpful to reduce chances of hallucinations? I tested couple of tricky questions and it seems it might work. by tamilupk
Yeah, I tried the "France" prompt in both ChatGPT4 and Bard, and both failed in the same way (ferret). Bard failed to adjust on review, but in a different way - it claimed whilst it was wrong about the letter, there were no animals that began with the letter 'P', which I did not expect!
[deleted] OP t1_jdviava wrote
Reply to comment by Specific-Arrival-127 in [R] BLLR "Budd´s Logistic Linear Regression" - A hybrid algorithm for neural network by [deleted]
[removed]
sEi_ t1_jdvi7f9 wrote
Reply to comment by ThePogromist in My ChatGPT Chrome Extension that saves conversations in .md files is finally approved by the Chrome Web Store. It's still and so will continue to be Open Source. [P] by ThePogromist
Try a search in any search engine for this: "Погромщик" or "ThePogromist" (OP's username).
You will not get any other definitions unless you specific ask for it by saying "jargon" or the like.
Why evade the fact that your username and you without reasonable doubt is glorifying the pogrom issue?
With a 5 day old reddit account, maybe ditch it and pick a more suitable name next time.
I mean no harm but your nick (still) worries me.
Btw. I can do GPT-4 too:
>The Russian word "погромщик" (pogromshchik) can be translated to English as "pogromist" or "rioter." It refers to someone who takes part in a pogrom, which is a violent attack or riot directed against a specific ethnic or religious group, often resulting in the destruction of property and the persecution or killing of individuals. Pogroms have historically been associated with the persecution of Jews in Eastern Europe, particularly in Russia and Ukraine.
vintergroena OP t1_jdvons9 wrote
Reply to comment by SmellElectronic6656 in [D] Can we train a decompiler? by vintergroena
Ultimately, you could force open-sourcing of any software (if you can run it on your device, not SaaS)