Recent comments in /f/MachineLearning
LanchestersLaw t1_jcf41pg wrote
Reply to comment by Competitive_Dog_6639 in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Bing Chat runs on GPT-4 and a fully version with multimodality is availible as a research preview
1998marcom t1_jcf3er8 wrote
Reply to comment by twilight-actual in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Detail note: to the best of my knowledge, as for what OpenAI is doing right now with their software, they could very well be using GPL code in their stack, and they wouldn't be violating any of the GPL clauses. A stricter licence such as AGPL I guess would be needed to cover as usage cases not only the shipping of software to the customer but also the mere utilization of the software.
ml_head t1_jcf3e64 wrote
Reply to [P] Multimedia GPT: Can ChatGPT/GPT-4 be used for vision / audio tasks just by prompt engineering? by Empty-Revolution7570
So, the model recognized the Cinderella story in the audio. But how do we know that summary was generated from the audio, and not from prior knowledge of the story? I know that those models can do this task. However, for the demo I would use an original story instead.
DamienLasseur t1_jcezbpd wrote
This is actually really cool! I've been demotivated about how much progress is occurring in the field of ML that I would've liked to contribute to. I'll give it a shot!
Additionally, if anyone would like to collaborate on this challenge, feel free to shoot me a PM and I'll set up a Discord or something.
frequenttimetraveler t1_jcewwlh wrote
Reply to comment by MrEloi in [D] ChatGPT responds to criticisms of GPT-4's high test scores. by spiritus_dei
chatGpt is good at writing nice-looking prose. That doesn't mean its content is actually good, just well-ornamented.
In real life, people who are uneducated are bad at both reasoning and language expression. People who are generall well-educated are good at both. chatGPT ain't like us
[deleted] t1_jcewul5 wrote
[removed]
frequenttimetraveler t1_jcewp2n wrote
Chatgpt itself ain't better. One thing chatGPT struggles with is jeopardy-like queries, where you give an answer and ask the question. It failed even with simple answers like "four neutrinos".
Bing however (which I assume is running gpt4) is good at it
suflaj t1_jcevhag wrote
Reply to comment by Eggy-Toast in [Discussion] What happened to r/NaturalLanguageProcessing ? by MadNietzsche
I understand what the user is saying, I do not understand how it relates to anything said before that.
Sadly, while GPT4 may be able to predict the next token based on studying the language's syntax thoroughly, it still fails to actually understand anything. Unless the original commenter is a bot, I would expect them to explain how what they said has anything to do with my comment, or the claims made about NLP researcher's obsolescence due to its release.
wazis t1_jceuzis wrote
Reply to comment by MrEloi in [D] ChatGPT responds to criticisms of GPT-4's high test scores. by spiritus_dei
Yeah because they didn't go to school
SvenAG t1_jceuwcf wrote
Reply to [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
I don't think that we are generally in the era of closed research now - there are still many companies sharing their research, ideas and concepts. If you are interested, we are trying to build an open alternative currently: https://opengpt-x.de/en/
But we are still in early stages
bartturner t1_jcetimy wrote
Reply to comment by twilight-actual in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
> Post mean things?
Not the terminology I would choose. But yes post things that they should not be doing this. Public opinion is a very, very powerful tool to get people to behave.
AccountGotLocked69 t1_jceti2q wrote
Reply to comment by camp4climber in [D] Is there an expectation that epochs/learning rates should be kept the same between benchmark experiments? by TheWittyScreenName
I mean... If this holds true for other benchmarks, it would be a huge shock for the entire community. If someone published a paper showing that AlexNet beats ViT on imagenet if you simply train it for ten million epochs, that would be insane. That would mean all the research into architectures we did in the last ten years can be replaced by a good hyperparameter search and training longer.
Balance- t1_jcete3n wrote
Thanks for organizing and funding this!
sam__izdat t1_jcet79g wrote
Reply to comment by currentscurrents in Modern language models refute Chomsky’s approach to language [R] by No_Draft4778
> Language isn't a structured formal thing
[citation needed]
bartturner t1_jcet1m2 wrote
Reply to comment by Single_Ad_2188 in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Exactly. I do not like that OpenAI looks to be changing the culture of sharing.
AccountGotLocked69 t1_jcesw8m wrote
Reply to comment by MrTacobeans in [D] Is there an expectation that epochs/learning rates should be kept the same between benchmark experiments? by TheWittyScreenName
I assume by hallucinate gaps you mean interpolate? In general it's the opposite, smaller simpler models are better at generalizing. Of course there are a million exceptions to this rule, but in the simple picture of using stable combinations of batch sizes and learning rates, big models will be more prone to overfit the data. Most of this rests on the assumption that the "ground truth" is always a simpler function than memorizing the entire dataset.
MrEloi t1_jces9ka wrote
Reply to comment by fnordit in [D] ChatGPT responds to criticisms of GPT-4's high test scores. by spiritus_dei
Don't be silly.
That response was better than most people in the street could do.
londons_explorer t1_jcer3ci wrote
Seems this can be cleanly split into 'unrolling' and 'ink recognition'.
Unrolling at first seems like the easy bit... But it could be made complex if there are fragments of material which have internally become detached and fallen
Single_Ad_2188 t1_jcer0ba wrote
Reply to comment by bartturner in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
>It seems like the days for open research in AI are gone.
if the Google not released the paper "Attention is all you need". then GPT is not possible to create
[deleted] t1_jceql2d wrote
[removed]
sam__izdat t1_jceowxm wrote
Ridiculously unfounded claim based on a just plain idiotic premise. Children don't learn language by cramming petabytes of text documents to statistically infer the most plausible next word in a sentence, nor do they accept input with arbitrary syntactic rules. Right or wrong, the minimalist program and Merge offer a plausible partial explanation for a recent explosion of material culture -- which did not happen gradually or across multiple species -- consistent with what we can observe in real human beings. GPT, on the other hand, is not a plausible explanation for anything in the natural world, and has basically nothing inherently to do with human language. He's not wrong that it's a bulldozer. It will just as happily accommodate a made-up grammar that has nothing in common with any that a person could ever use, as it would English or Japanese.
> Chomsky et al. 2023 tilt at an imagined version of these models, while ignoring the fact that the real ones so aptly capture syntax, a success Chomsky and others have persistently claimed was impossible.
Exactly the opposite is true. Transformers are general-purpose computers that will gobble up almost anything you can throw at them. His objection was to the "defect" that it will capture any arbitrary syntax, which means it isn't interesting or helpful to cognitive scientists -- just like like a backhoe doesn't offer any insight into how people, in biological terms, are able to lift heavy objects. What he said was impossible, when approached about it decades ago, was to do these things without resorting to brute force in the absence of an actual theoretical framework and computational model for how language works in the brain. That statement is just as correct today as it was in the 1950s and the rigorous theory of "let's keep cramming in data and stirring the big ol' pot of linear algebra until candy comes out" doesn't do anything to to change that picture.
> In particular, we may view each model or set of modeling assumptions as a possible hypothesis about how the mind may work. Testing how well a model matches humanlike behavior then provides a scientific test of that model’s assumptions.
I see. So, by that reasoning, Deep Blue affirmed our hypothesis that Kasporov got really good at chess by evaluating 200,000,000 positions per second. Nonsense on stilts.
> Hosseini et al. 2022 present evidence that models trained on developmentally-plausible amounts of data already capture human neural responses to language in the brain
If you define "developmentally plausible" as "100 million tokens" and "capture human neural responses" as "fasten together a linear regression model to make statistically significant correlations with some coarse fMRI activity."
> After decades of privilege and prominence in linguistics, Noam Chomsky’s approach to the science of language is experiencing a remarkable downfall. The story is, in part, a cautionary tale about what happens when an academic field isolates itself from what should be complementary endeavours.
Sound advice for overconfident doe-eyed futurists guzzling the silicon valley kool aid and assuming they can dispense with common sense and just wing it with anthropology, biology and cog sci.
edit - and, by the way, it's actually kind of impressive for a person to miss this point so completely:
> Generate ten other sentences like [colorless green ideas]
> Purple fluffy clouds dream wildly
> Blue glittery unicorns jump excitedly
> Orange sparkly penguins waddle energetically
> Red shimmering elephants trumpet melodiously
> Yellow dazzling lions roar majestically
> Green glittery monkeys swing energetically
> Pink sparkly giraffes walk gracefully
> Black shiny kangaroos hop playfully
> White glittery bears cuddle affectionately
> Brown shimmering rabbits hop merrily.
Literally none of these cut-and-paste color-texture-animal-verb-adverb responses satisfies the query, and to get even a little bit closer with GPT, you have to explicitly beg for examples of contradictions and category errors.
asdfsflhasdfa t1_jceoofi wrote
Reply to comment by gengarvibes in [D] ChatGPT responds to criticisms of GPT-4's high test scores. by spiritus_dei
I don’t think not mentioning a criticism of standardized testing means the model is necessarily biased... or “the training data is skewed towards the wealthy”. Whatever that means lol
mikonvergence OP t1_jceoitq wrote
Reply to comment by kross00 in [P] ControlNetInpaint: No extra training and you can use 📝text +🌌image + 😷mask to generate new images. by mikonvergence
Hi! Do you think many people would be interested? I've never used Automatic1111 since I work in research so I need direct access to the source code.
If there is enough interest, I will definitely check out how to release it as an extension!
serge_cell t1_jcenssl wrote
Reply to [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Let's fight the fire with gasoline.
lapurita t1_jcf4fs6 wrote
Reply to comment by thecity2 in [D] What do people think about OpenAI not releasing its research but benefiting from others’ research? Should google meta enforce its patents against them? by [deleted]
Ugh I'm not psyched about this at all, it will just protect the big companies from competitors and result in worse products for everyone