Recent comments in /f/MachineLearning
andreichiffa t1_jdu5wmj wrote
Reply to [D] Will prompting the LLM to review it's own answer be any helpful to reduce chances of hallucinations? I tested couple of tricky questions and it seems it might work. by tamilupk
Yes, that’s the mechanism GPT-4 paper showed they were using for a bunch of things in the annex. It was initially discovered in the toxicity detection domain (RealToxicPrompts paper I believe)
MysteryInc152 t1_jdu4v0n wrote
In the gpt-4 technical paper, we see base gpt-4 have really good calibration. That is confidence directly correlated with ability to solve problems. But apparently the RlHF they did knocked that out some.
MysteryInc152 t1_jdu4sl2 wrote
Reply to comment by [deleted] in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
In the gpt-4 technical paper, we see base gpt-4 have really good calibration. That is confidence directly correlated with ability to solve problems. But apparently the RlHF they did knocked that out some.
blose1 t1_jdu4cln wrote
Reply to comment by visarga in [D] GPT4 and coding problems by enryu42
What? Where I'm exactly mistaken? Because both of my statements are true. And there is 0% chance you can pass olympiad task without knowledge, human with all the knowledge WILL reason and come up with a solution BASED on the knowledge he has AND experience of others that is part of that knowledge, if that weren't true then no human would solve any Olympiad. Sorry, but what you wrote in context of my comment is just ridiculous, and looks like a reply to something I didn't write.
super_deap t1_jdu3zan wrote
Reply to comment by artsybashev in [D] GPT4 and coding problems by enryu42
It is fine if you disagree and I believe a lot more people will disagree with this philosophical position as it is not very popular these days.
Near-death experiences, out-of-body experiences, contact with 'immaterial entities' and so on hint towards an existence beyond our material reality. Since there is no way one could 'scientifically' test these does not mean these things simply do not exist.
Testimony widely used yet mostly dismissed method of knowledge acquisition establishes all of the above:
A patient being operated on while in a complete medical comma explaining the things happening in clear details in a nearby room after the operation that there is no way they could have known that, one such testimony by a reliable person is sufficient to establish that our current understanding of the world is insufficient. And there are so many of these.
I am not saying u have to change your worldview just because I am saying so. do your research. the world is much bigger than what is out there on the internet. (pun intended)
Cool_Abbreviations_9 OP t1_jdu3xlj wrote
Reply to comment by [deleted] in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
It does appear to have some calibration capabilities
currentscurrents t1_jdu3vwq wrote
Reply to comment by Smallpaul in [D] Can we train a decompiler? by vintergroena
Yeah, but they're hand-crafted algorithms and produce code that's hard to read.
lvvy t1_jdu3qg4 wrote
Reply to [D] GPT4 and coding problems by enryu42
It would be interesting to see if ChatGPT can solve these problems not with code, but with a text instruction, that would allow a human to solve these problems? So if you force it to write giant text wall of actions, would a human with calculator solve these confident? Also, is code that it generates cannot be corrected at all by discussing or discussing would take too long?
sdmat t1_jdu3imb wrote
Reply to [D] Can we train a decompiler? by vintergroena
GPT4 will do this to an extent out of the box, feed it some assembly and it will hypothesise a corresponding program in the language of your choice. For me it still has that disassembler character of over-specificity, but I didn't try very hard to get idiomatic result.
It can give detailed analysis of assembly too, including assessing what it does at a high level in plain english. Useful!
Edit: Of course it's going to fail hopelessly for large/complex programs.
Majestic_Food_4190 t1_jdu3fod wrote
Reply to comment by cegras in [D] GPT4 and coding problems by enryu42
It amuses me that people always mentions things of this nature. If the answer is simply, yes.... Then it's still doing it far faster than you are. Making it a better developer than most others.
It's like Watson beating the top people at jeopardy. Was it just searching the internet? Pretty much. Did it in turn win jeopardy? Yes.
So does the how matter?
Smallpaul t1_jdu38fb wrote
Reply to [D] Can we train a decompiler? by vintergroena
Decompilers already exist though.
Disastrous_Elk_6375 t1_jdu2tzp wrote
Reply to comment by rePAN6517 in [D] GPT4 and coding problems by enryu42
badum-tsss
artsybashev t1_jdu2hjs wrote
Reply to comment by super_deap in [D] GPT4 and coding problems by enryu42
The physical world that we know is very different from the virtual twin that we see. The human mind lives in a virtual existence created by the material human brain. This virtual world creates nonexisting things like pain, colors, feelings and also the feeling of existence.
The virtual world that each of our brain creates is the wonderful world where a soul can emerge. Virtual worlds can also be created by computers. There is no third magical place besides these two in my view.
Borrowedshorts t1_jdu1o78 wrote
So if you're using this for academic research, you can put in your original prompt and then tell it to only return references with a confidence score > .5. Neat little trick.
[deleted] t1_jdu1mz6 wrote
[deleted]
visarga t1_jdu1fgf wrote
Reply to comment by LifeScientist123 in [D] GPT4 and coding problems by enryu42
> Does this mean developers/humans don't have AGI?
The intellect of our species isn't universal, we're merely experts at self-preservation and propagation. Take, for instance, chess – it isn't our forte, and even a small calculator could outperform us. Our minds are incapable of 5-D visualization, and we struggle to maintain over 10 unrelated items in our immediate memory. Generally, we falter when addressing problems where the initial move relies on the final steps, or situations that don't allow for linear progression, such as chess or mathematical quandaries. It took us centuries to decipher many of these enigmas. Our specialization lies in tackling human-centric challenges, rather than all-encompassing ones. Evolution simply hasn't had sufficient time to adapt our cerebral cortex for mathematical prowess.
super_deap t1_jdu0w8f wrote
Reply to comment by artsybashev in [D] GPT4 and coding problems by enryu42
Hard disagree with Materialism. I know I might get a lot of -ve votes, but this has to be said:
A large portion of the world (especially outside of the west) does not believe in 'consciousness "emerging" from electrical impulses of the brain.' While the west has progressed a lot materially, bringing us to modernity (and now post-modernity), people outside of the west believe in an immaterial soul that cannot be captured by definition by the scientific method and it transcends our material body.
While I believe we will reach general human-level intelligence (and may go beyond this) because intelligence has a purely material component that we can replicate in computers, consciousness will never ever arise in these systems. There are very strong philosophical arguments to support this case.
mil24havoc t1_jdu05qo wrote
Reply to [D] Can we train a decompiler? by vintergroena
There's already research on this. For example, see "DIRECT: A Transformer Model for Decompiled Variable Name Recovery" by Nitin et al.
Cool_Abbreviations_9 OP t1_jdu04a2 wrote
The first two papers exist but the others don't
vintergroena OP t1_jdtzyuo wrote
Reply to comment by Fit-Recognition9795 in [D] Can we train a decompiler? by vintergroena
Yeah, GPT sucks on tasks which require actual thinking and personally I am kind of skeptical about it's actual usefulness tbh. But my impression is that despite being primarily built to work with natural language, it actually does work better with computer code. Probably because computer code has much simpler structure. This got me thinking that building something more specialized that would be required to only work with computer code would actually be an easier task - more similar to automated translation perhaps, which is already working pretty well using ML.
addition t1_jdtzh4k wrote
Reply to comment by COMPEWTER_adminisp in [D] GPT4 and coding problems by enryu42
Clearly I’m not the first person to think this by a long shot. I was just pointing out that a new trend has been forming recently.
visarga t1_jdtz3wx wrote
Reply to comment by boaking69 in [D] GPT4 and coding problems by enryu42
The original title of the "Sparks of AGI" paper was "First Contact With an AGI System" (line 8). If you carefully read the paper it suggests GPT-4 is stronger than what seems to be our consensus.
Fit-Recognition9795 t1_jdtz20q wrote
Reply to [D] Can we train a decompiler? by vintergroena
You think we are far ahead of where we are... and I also wish we were there, but we are not.
Not saying one day will not be possible but have you tried asking gpt4 to multiply two 3 digits numbers?
sdmat t1_jdtytyy wrote
Reply to comment by yaosio in [D] GPT4 and coding problems by enryu42
> Like coding, even if you use chain of thought and self reflection GPT-4 will try to write the entire program in one go. Once something is written it can't go back and change it if it turns out to be a bad idea, it is forced to incorporate it. It would be amazing if a model can predict how difficult a task will be and then break it up into manageable pieces rather than trying to do everything at once.
I've had some success leading it through this in coding with careful prompting - have it give a high level outline, check its work, implement each part, check its work, then put the thing together. It will even revise the high level idea if you ask it to and update a corresponding implementation in the context window.
But it definitely can't do so natively. Intuitively it seems unlikely that we can get similar results to GPT4+human with GPT4+GPT4 regardless of how clever the prompting scheme is. But the emergent capabilities seen already are highly surprising, so who knows.
Really looking forward to trying these schemes with a 32K context window.
Add code execution to check results and browsing to to get library usage right and it seems all the pieces are there for an incredible level of capability even it still needs human input in some areas.
Haycart t1_jdu7hlp wrote
Reply to comment by visarga in [D] GPT4 and coding problems by enryu42
Oh, you are probably correct. So it'd be O(N^2) overall for autoregressive decoding. Which still exceeds the O(n log n) that the linked post says is required for multiplication, though.