Recent comments in /f/MachineLearning
Matthew2229 t1_jduz7mi wrote
Reply to comment by Kush_McNuggz in [D] Simple Questions Thread by AutoModerator
When you're clustering or classifying, you are predicting something discrete (clusters/classes), so it's unclear what you mean by removing these hard cutoffs. There must be some kind of hard cutoff when doing clustering/classification unless you are okay with something having a fuzzy classification (e.g. 70% class A / 30% class B).
bjj_starter t1_jduz6p7 wrote
Reply to comment by thecodethinker in [D] GPT4 and coding problems by enryu42
I'm not sure if most of them would agree, based on their actions and statements. They certainly think that AI is an existential risk, but that is a different thing from viewing it as conscious. You could definitely be right, I just haven't seen much from them that would indicate it.
That said, the extremely common sense position you just outlined was mainstream among basically all respectable intellectuals who had any position on AI, right up until the rubber hit the road and it looked like AI might actually achieve that goal in the near future. The fact is that if something behaves like a conscious entity in all of the ways that matter, it is conscious for the sake of the social meaning of the term. Provenance shouldn't matter any more than gender.
Matthew2229 t1_jduyuw9 wrote
Reply to comment by Various_Ad7388 in [D] Simple Questions Thread by AutoModerator
I think either is probably fine to learn. Both have roughly the same set of features at this point. TF used to be the pre-dominant framework, but PyTorch has gained popularity over the past few years. Now if it'll stay that way or there will be a new trend in the future, no one can say for sure.
[deleted] t1_jduysn7 wrote
Reply to comment by light24bulbs in [P] Using ChatGPT plugins with LLaMA by balthierwings
[removed]
HatsusenoRin t1_jduyliz wrote
Reply to comment by lhenault in [P] SimpleAI : A self-hosted alternative to OpenAI API by lhenault
Thanks for your help. I think I'll take a look at the /docs.
Matthew2229 t1_jduyi8o wrote
Reply to comment by masterofn1 in [D] Simple Questions Thread by AutoModerator
It's a memory issue. Since the attention matrix scales quadratically (N^2) with sequence length (N), we simply don't have enough memory for long sequences. Most of the development around transformers/attention has been targeting this specific problem.
Deep-Station-1746 t1_jduxvm4 wrote
Reply to comment by [deleted] in [D] Build a ChatGPT from zero by manuelfraile
You know what? provide me with a 2-3 sample "good" responses to the above post, and explain why they make for a better response than what I wrote, and I'll actually use them from now on to respond to low-effort posts from this sub.
Matthew2229 t1_jdux7sh wrote
Reply to [D] Definitive Test For AGI by jabowery
No, this is not a "definitive test for AGI". It just shows that the system is able to solve a single task. What if you give the same model a simple IQ test question and it fails miserably? Clearly it's not an AGI.
ThePogromist OP t1_jdux39f wrote
Reply to comment by -xylon in My ChatGPT Chrome Extension that saves conversations in .md files is finally approved by the Chrome Web Store. It's still and so will continue to be Open Source. [P] by ThePogromist
Sure, when I will have enough of a free time again, but now I'm too busy with university.
Though I will aprecciate someone to make a commits on github for firefox support.
Tr4sHCr4fT t1_jduwqp7 wrote
Reply to [D] GPT4 and coding problems by enryu42
it can only replace bootcamp juniors ^/s
-xylon t1_jduwnkm wrote
Reply to My ChatGPT Chrome Extension that saves conversations in .md files is finally approved by the Chrome Web Store. It's still and so will continue to be Open Source. [P] by ThePogromist
Any chance to have this for Firefox?
light24bulbs t1_jduwgqt wrote
Reply to comment by was_der_Fall_ist in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
Yeah, like it's actually using a huge amount of brain power to figure out what the next word is. Just because that's how it works doesn't mean it's not intelligent.
If you want to be really good at figuring out what the next word is you have to be really smart
FermiAnyon t1_jduvxst wrote
Reply to comment by Matthew2229 in Have deepfakes become so realistic that they can fool people into thinking they are genuine? [D] by [deleted]
Ooh, text... that's a really good point. Okay okay. I'm happy to jog that back then.
thecodethinker t1_jduvi9z wrote
Reply to comment by bjj_starter in [D] GPT4 and coding problems by enryu42
That’s not even to mention that appearing conscious is as good as being conscious as far as the teams behind these LLMs are concerned.
There’s no practical difference
knome t1_jduvetc wrote
after leaning the token-at-a-time seeing its own output nature of the model, I asked it to start printing outputs, and then following them with a judgment on whether they were correct. It usually caught itself.
light24bulbs t1_jduuuep wrote
Reply to comment by alexmin93 in [P] Using ChatGPT plugins with LLaMA by balthierwings
I do, still struggling with it
topcodemangler t1_jduuhcf wrote
Reply to [D] Simple Questions Thread by AutoModerator
Is there any real progress on the JEPA architecture proposed and pushed by LeCun? I see him constantly bashing LLMs and saying how we need JEPA (or something similar) to truly solve intelligence but it has been a long time since the initial proposition (2 years?) and nothing practical has come out of it.
​
It may sound a bit aggressive but that was not my intention - the original paper really sparked my interest and I agree with a lot that he has to say. It's just that I would want to see how those ideas fare in the real world.
[deleted] t1_jdutklr wrote
Reply to [D] ICML 2023 Reviewer-Author Discussion by zy415
[deleted]
Smallpaul t1_jdusjts wrote
Reply to comment by matthkamis in [D] Can we train a decompiler? by vintergroena
Good question.
The generative model might be able to learn with fewer examples? Because if already knows more about coding in various languages?
Just a guess.
Readorn t1_jdureq0 wrote
Reply to [P] nanoT5 - Inspired by Jonas Geiping's Cramming and Andrej Karpathy's nanoGPT, we fill the gap of a repository for pre-training T5-style "LLMs" under a limited budget in PyTorch by korec1234
So like let me get this straight, we can download this repository, train the model, this NanoGPT model and use it?
[deleted] t1_jdur1dt wrote
Reply to comment by nixed9 in [D] GPT4 and coding problems by enryu42
[removed]
sdmat t1_jduqoi1 wrote
Reply to comment by redboundary in [D] GPT4 and coding problems by enryu42
Pity they never made more seasons of that show
[deleted] t1_jduqdc3 wrote
Reply to comment by Deep-Station-1746 in [D] Build a ChatGPT from zero by manuelfraile
[deleted]
matthkamis t1_jduqbsd wrote
Reply to [D] Can we train a decompiler? by vintergroena
Why do you need a generative model for this? Couldn’t this be done with standard supervised learning?
xorbinant_ranchu t1_jduzaeh wrote
Reply to comment by HatsusenoRin in [P] SimpleAI : A self-hosted alternative to OpenAI API by lhenault
Try obsidian with ChatGPT_md plugin
https://github.com/bramses/chatgpt-md