Recent comments in /f/MachineLearning
SmLnine t1_jdmftzs wrote
Reply to comment by sweatierorc in [R] Reflexion: an autonomous agent with dynamic memory and self-reflection - Noah Shinn et al 2023 Northeastern University Boston - Outperforms GPT-4 on HumanEval accuracy (0.67 --> 0.88)! by Singularian2501
I said "effectively" because a blanked statement would be unwarranted. There has probably been at least one naked mole rate in the history of the universe that got cancer.
https://www.cam.ac.uk/research/news/secrets-of-naked-mole-rat-cancer-resistance-unearthed
Small-Fall-6500 t1_jdmftp5 wrote
Art10001 t1_jdmff0b wrote
Reply to comment by sweatierorc in [R] Reflexion: an autonomous agent with dynamic memory and self-reflection - Noah Shinn et al 2023 Northeastern University Boston - Outperforms GPT-4 on HumanEval accuracy (0.67 --> 0.88)! by Singularian2501
More intelligence, more time (AIs are at different time scales) = faster rate of discoveries
devl82 t1_jdmf6b3 wrote
Reply to comment by Yardanico in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
no they is no overhype, you just don't understand what Alpaca is trying to do & I am sure others will also reply similar
Daveboi7 t1_jdme4qf wrote
Can we just download the model?
SDRealist t1_jdmdwkl wrote
Reply to comment by visarga in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
Users could certainly donate their questions, but I believe the TOS for ChatGPT forbid using the generated output to train competing models (at least for commercial purposes).
machineko t1_jdmdvst wrote
Reply to comment by light24bulbs in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
We are working on adding that as well. Keep an eye out on our repo.
Rejg t1_jdmdspx wrote
Reply to comment by michaelthwan_ai in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
I think you are potentially missing Claude 1.0 and Claude 1.2, the Co:Here Suite, and Google Flan models.
ZestyData t1_jdmdjrd wrote
Reply to comment by DarkTarantino in [N] March 2023 - Recent Instruction/Chat-Based Models and their parents by michaelthwan_ai
Well.. you can just create these graphs if its important for your current task.
There isn't a role called "Chief graph maker" who makes graphs for people when they need them.
ambient_temp_xeno t1_jdmdh2i wrote
Reply to comment by Nyanraltotlapun in [R] Reflexion: an autonomous agent with dynamic memory and self-reflection - Noah Shinn et al 2023 Northeastern University Boston - Outperforms GPT-4 on HumanEval accuracy (0.67 --> 0.88)! by Singularian2501
There is work done on how to even start interacting with an extraterrestrial civilization, and it would probably be a vast amount harder than whatever intelligence is contained in a human-data-filled, human-trained model. https://www.nasa.gov/connect/ebooks/archaeology_anthropology_and_interstellar_communication.html
That said, it is the closest we have to that so you're not 'wrong'.
harharveryfunny t1_jdmd38s wrote
Reply to comment by alrunan in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
>You should read the LLaMA paper.
OK - will do. What specifically did you find interesting (related to scaling or not) ?
Zealousideal_Low1287 t1_jdmcuti wrote
Reply to [D] Do you use a website or program to organise and annotate your papers? by who_here_condemns_me
I print them out and put them in piles. I write on them with pens.
[deleted] t1_jdmcma9 wrote
Reply to comment by soggy_mattress in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
[deleted]
SpaceCadetIowa t1_jdmcfga wrote
Reply to [R] Reflexion: an autonomous agent with dynamic memory and self-reflection - Noah Shinn et al 2023 Northeastern University Boston - Outperforms GPT-4 on HumanEval accuracy (0.67 --> 0.88)! by Singularian2501
No need, the government makes up new ones to keep the people thinking we need them.
alrunan t1_jdmbv4k wrote
Reply to comment by harharveryfunny in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
The chinchilla scaling laws is just used to calculate the optimal scale for dataset and model size for a particular training budget.
You should read the LLaMA paper.
LazyCheetah42 t1_jdmb2i0 wrote
is there already a dolly.cpp?
light24bulbs t1_jdmad5n wrote
Reply to comment by machineko in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
Hey, I've been looking at this more and it's very cool. One thing I REALLY like is that I see see self-training using dataset generation on your roadmap. This is essentially the technique that Facebook used to train ToolFormer, if I'm reading their paper correctly.
I'd really love to use your library to try to reimplement toolformers approach someday.
EmmyNoetherRing t1_jdma3em wrote
Reply to comment by Crystal-Ammunition in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
Introspection? Cog-sci/classical AI like to use the term, not always in the best justified fashion I think. But when you’re hallucinating your own new training data it seems relevant.
[deleted] t1_jdma10z wrote
Reply to comment by nicku_a in [P] Reinforcement learning evolutionary hyperparameter optimization - 10x speed up by nicku_a
[deleted]
danielbln t1_jdm967m wrote
Daveboi7 t1_jdm8aby wrote
Reply to comment by dreamingleo12 in [R] Hello Dolly: Democratizing the magic of ChatGPT with open models by austintackaberry
What platform are you using for training?
sweatierorc t1_jdm83bv wrote
Reply to comment by SmLnine in [R] Reflexion: an autonomous agent with dynamic memory and self-reflection - Noah Shinn et al 2023 Northeastern University Boston - Outperforms GPT-4 on HumanEval accuracy (0.67 --> 0.88)! by Singularian2501
which one ? do they not get cancer or are they more resistant to it ?
noobgolang t1_jdm7pvm wrote
Big is not always better (˵ ͡° ͜ʖ ͡°˵)
atheist-projector t1_jdm7mmi wrote
Reply to comment by Blacky372 in [D] Do we really need 100B+ parameters in a large language model? by Vegetable-Skill-9700
I love the odea of calling them closedai.
Thats it j am doing it from now on
[deleted] t1_jdmgnya wrote
Reply to [R] Reflexion: an autonomous agent with dynamic memory and self-reflection - Noah Shinn et al 2023 Northeastern University Boston - Outperforms GPT-4 on HumanEval accuracy (0.67 --> 0.88)! by Singularian2501
[removed]