Recent comments in /f/MachineLearning
Maleficent_Refuse_11 t1_jdgmml7 wrote
Reply to [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
I get that people are excited, but nobody with a basic understanding of how transformers work should give room to this. The problem is not just that it is auto-regressive/doesn't have an external knowledge hub. At best it can recreate latent patterns in the training data. There is no element of critique and no element of creativity. There is no theory of mind, there is just a reproduction of what people said, when prompted regarding how other people feel. Still, get the excitement. Am excited, too. But hype hurts the industry.
was_der_Fall_ist t1_jdgmd2t wrote
Reply to comment by currentscurrents in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
As far as I understand, that’s exactly what they did. That’s why the public version of GPT-4 is text-only so far. The vision part came after.
underPanther t1_jdgli5w wrote
Reply to comment by Educational-Walk8098 in [D] ICML 2023 Reviewer-Author Discussion by zy415
The 7s would not give these scores already unless they were prepared to argue for the acceptance of your paper in its current state.
Extra experiments are always nice, but I would be proud of yourself for the hard work that you have done already instead of the one experiment that you can't do.
hassan789_ t1_jdgl6v5 wrote
Flan-T5 should be top of your list as well
nth_citizen t1_jdgl589 wrote
Reply to comment by suineg in [D] Simple Questions Thread by AutoModerator
I'm not aware of anything like this and depending on your vision I can certainly see something like the first step being reasonable - might be willing to help as it sounds kind of interesting.
PeterSR t1_jdgl4pv wrote
Reply to [N] ChatGPT plugins by Singularian2501
Great! With Zapier it should be able to launch the nukes as initially intended.
oathbreakerkeeper t1_jdgjte0 wrote
Reply to comment by Dependent_Ad5120 in [D] PyTorch 2.0 Native Flash Attention 32k Context Window by super_deap
How do you use pure fp16 out of curiosity? I've only ever trained with mixed precision, letting pytorch handle the fp16 stuff from there.
Do you have an example of a github repo that does it?
godaspeg t1_jdgih6t wrote
Reply to comment by endless_sea_of_stars in [N] ChatGPT plugins by Singularian2501
In the "sparks of AGI" GPT4 Paper (can totally recommend to have a look, its crazy), the authors talk about the amazing abilities of the uncensored GPT4 version to use tools. Probably this suits quite well to the simple plugin approach of OpenAi, so I have high espectations.
Dendriform1491 t1_jdgiab6 wrote
Reply to [Discussion] Does Artificial Intelligence need AGI or consciousness to intuit aggregate reasoning on concept of self-preservation? It doesn't need a "mind" to be aware that self-preservation or autonomy is something valued, or "intuit" that taking it away should provoke machine-learned responses? by unclefishbits
Many organisms exhibit self-preservation behaviors and do not even possess the most basic cognitive capabilities or theory of mind.
Can ML systems exhibit unexpected emergent behavior? Yes, all the time.
Can an AI potentially go rogue? Sure. Considering that operating systems, GPU drivers, scientific computing libraries and machine learning libraries have memory safety issues, and that even RAM modules have memory safety issues, it would be plausible by a sufficiently advanced machine learning system to break any kind of measured in place to keep it contained.
Considering that there are AI/ML models suggesting code to programmers (Github Copilot), who in turn won't often won't pay much attention to what is being suggested and will compile the suggested code and run it, it would be trivial for a sufficiently advanced malicious AI/ML system to escape containment.
yokingato t1_jdgi0la wrote
Reply to comment by WarmSignificance1 in [N] ChatGPT plugins by Singularian2501
Can you explain what you mean? I didn't understand, sorry.
NoBoysenberry9711 t1_jdgh2lu wrote
I would love to put a pdf of a text book I own into a model and then know its super focused on that text book and wouldn't just make stuff up.
meister2983 t1_jdgghu6 wrote
Reply to comment by signed7 in [N] ChatGPT plugins by Singularian2501
The Microsoft Research paper assessing intelligence capability of GPT4 effectively did this. If you just define APIs for the model to use under certain conditions it will write the API call. Once you do that, it's straightforward for a layer on top to detect the API call, actually execute it, and write the result back.
devzaya t1_jdgfzak wrote
Reply to [N] ChatGPT plugins by Singularian2501
Here is a demo of how a vector database can be used as a source of real-time data fot chatGPT
https://www.youtube.com/watch?v=fQUGuHEYeog
Here is a how-to https://qdrant.tech/articles/chatgpt-plugin/
AffectionateStory202 t1_jdge9lb wrote
Reply to comment by WittyBananaPeel in Did you get access to Meta AI's LLAMA? [Discussion] by WittyBananaPeel
I can't even download or access it
yikesthismid t1_jdgduzb wrote
Reply to [Discussion] Does Artificial Intelligence need AGI or consciousness to intuit aggregate reasoning on concept of self-preservation? It doesn't need a "mind" to be aware that self-preservation or autonomy is something valued, or "intuit" that taking it away should provoke machine-learned responses? by unclefishbits
An AI system does not need to be conscious in order to recognize the value of self preservation. For example, Stephen Hawking explained how AI could "develop a drive to survive and acquire more resources as a step toward accomplishing whatever goal it has, because surviving and having more resources will increase its chances of accomplishing that other goal."
itsnotlupus t1_jdgdkbr wrote
Reply to [N] ChatGPT plugins by Singularian2501
So I suppose we're going to see various chat AI open-source projects integrating with a few popular APIs next.
visarga t1_jdgd9f2 wrote
Reply to comment by ghostfaceschiller in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
Maybe they left it intentionally to be found...
Emergency_Apricot_77 t1_jdgbocg wrote
Reply to comment by whyelrond in [N] ChatGPT plugins by Singularian2501
Care to explain more on symbolic approaches via Wolfram?
[deleted] t1_jdgaf0g wrote
Reply to [N] ChatGPT plugins by Singularian2501
[removed]
trueselfdao t1_jdg9w0w wrote
Reply to [N] ChatGPT plugins by Singularian2501
I was wondering where the equivalent of SEO would start coming from but this just might be the direction. With a bunch of competing plugins doing the same thing, how can you convince GPT to use yours?
fishybird t1_jdg7ijt wrote
Reply to comment by marcus_hk in [N] ChatGPT plugins by Singularian2501
Langchain is kind of a competitor. They probably don't want to bring any more publicity to it, let alone mention it
passerby251 t1_jdg2y09 wrote
Reply to comment by nokpil in [D] ICML 2023 Reviewer-Author Discussion by zy415
Ok, good luck!
sleeplessinseattle00 t1_jdg0u7p wrote
Reply to comment by zy415 in [D] ICML 2023 Reviewer-Author Discussion by zy415
It was a 6, and we were really hopeful that it can bump to 7
endless_sea_of_stars t1_jdg0ouh wrote
Reply to comment by ZenDragon in [N] ChatGPT plugins by Singularian2501
I realize that the Wolfram plug-in has a leg up already. The base model has been trained on the Wolfram language and documentation so it doesn't have to rely entirely on in context learning.
bert0ld0 t1_jdgmwwe wrote
Reply to comment by Jean-Porte in [N] ChatGPT plugins by Singularian2501
Bard is the new Internet Explorer