Recent comments in /f/MachineLearning
TheEdes t1_je6tweq wrote
Reply to comment by cegras in [N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data by Balance-
Sure but what's being advertised isn't sentience per se, at least with the leetcode part of their benchmarks. The issue here is that they claim that it can do X% on leetcode, but it seems like it's much less on new data. Even if it learned to find previous solutions and replace it with changes it should be able to perform well due to the nature of the problems.
xander76 OP t1_je6tpox wrote
Reply to comment by HangOutWithMyYangOut in [P] Imaginary programming: implementation-free TypeScript functions for GPT-powered web development by xander76
You can check out the code at https://github.com/imaginary-dev/imaginary-dev , and there's code for demo projects as well:
Blog Writing demo from the screencast: Live site: blog-demo.imaginary.dev Code: https://github.com/imaginary-dev/imaginary-dev/tree/main/example-clients/nextjs-blog
Emojifier Live site: emojifier.imaginary.dev Code: https://github.com/imaginary-dev/imaginary-dev/tree/main/example-clients/nextjs-api
I'm also happy to answer any questions you might have!
EvilMegaDroid t1_je6s99k wrote
Reply to comment by Zealousideal-Ice9957 in [D] FOMO on the rapid pace of LLMs by 00001746
Good idea, I'm kinda skeptical if enough users would complete tasks for it to get enough data.
Not impossible though, there are huge open source projects so who knows.
[deleted] t1_je6ra9u wrote
Reply to comment by AdamEgrate in [D] The best way to train an LLM on company data by jaxolingo
[removed]
CyberDainz t1_je6qsbb wrote
The success of generalization for segmentation depends not only on the network configuration, but also on the augmentation and pretrain on non mask target.
try my new project Deep Roto https://iperov.github.io/DeepXTools/
[deleted] t1_je6q2mw wrote
Reply to [N] OpenAI may have benchmarked GPT-4’s coding ability on it’s own training data by Balance-
[removed]
Nobodyet94 t1_je6pqi2 wrote
Reply to [D] Simple Questions Thread by AutoModerator
Can you advise me a Vision Transfomer project to present at university? Thanks!
Intelligent_Rough_21 t1_je6plhz wrote
Reply to [P] Imaginary programming: implementation-free TypeScript functions for GPT-powered web development by xander76
I knew this would happen. Programming as a profession is doomed to laziness lol.
viertys OP t1_je6peyv wrote
Reply to comment by azorsenpai in [D] Improvements/alternatives to U-net for medical images segmentation? by viertys
I started with U-Net, but I'm open to other architectures. I will try out DeepLab V3, thank you!
I believe the data is generally clean. Sadly, I can't get more data as all the datasets used in the research papers that I've read are private.
shitasspetfuckers t1_je6p0z9 wrote
Reply to comment by detached-admin in [D] The best way to train an LLM on company data by jaxolingo
Why not other people's money?
Hands0L0 t1_je6o5t5 wrote
Reply to comment by Im_Unlucky in [D] The best way to train an LLM on company data by jaxolingo
Off topic but I love how underpants gnomes memes are still relevant 25 years later
MadScientist-1214 t1_je6o0st wrote
Most new architectures based on U-Net do not actually work. Researchers need papers to get published, so they introduce leakage or optimize the seed. Segmentation papers in journals like CVPR are of better quality.
reditum t1_je6nzh0 wrote
Reply to [P] Imaginary programming: implementation-free TypeScript functions for GPT-powered web development by xander76
Dude, this is pretty amazing.
My biggest concern with it not writing the code is that it might not perform as well (network latency or connectivity issues) and won't be deterministic and could hallucinate, but I could also imagine a few cases where GPT would generate faster than code would run (and some non-determinism will be desirable sometimes).
The name of your product is super-catchy as well. I can definitely see imaginary programming becoming a trend!
Ifnerite t1_je6m5ke wrote
Reply to [P] Imaginary programming: implementation-free TypeScript functions for GPT-powered web development by xander76
/** This function returns an awesome react from tend. */. (props or something) => {...}.
Hooray! New I don't have to deal with the godawful mess that is fronted development ever again!
HangOutWithMyYangOut t1_je6m0vt wrote
Reply to [P] Imaginary programming: implementation-free TypeScript functions for GPT-powered web development by xander76
This is incredibly creative. Any chance you'll be putting an example project on github
was_der_Fall_ist t1_je6lfl9 wrote
Reply to comment by ntaylor- in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
Why are matrix multiplications mutually exclusive with complicated operations?
A computer just goes through a big series of 0s and 1s, yet through layers of abstraction they accomplish amazing things far more complicated than a naive person would think 0s and 1s could represent and do. Why not the same for a massive neural network trained via gradient descent to maximize a goal by means of matrix multiplication?
visarga t1_je6kqvw wrote
Reply to comment by EverythingGoodWas in [D] The best way to train an LLM on company data by jaxolingo
I'd rather fine-tune the LLM on company documentations than feeding it through retrieval. Does anyone have experience with fine-tuning GPT-3 on a new text? Can it answer questions or freely use information from this text?
visarga t1_je6k74d wrote
Reply to comment by Im_Unlucky in [D] The best way to train an LLM on company data by jaxolingo
Often it seems that the model can't properly synthesise information from a bunch of snippets, it lacks the context of those snippets, so it will combine the information incorrectly or hallucinate an explanation.
Retrieval + loading data in the context is far from solved.
lambertb t1_je6jjps wrote
Reply to comment by WokeAssBaller in [D] GPT4 and coding problems by enryu42
There can’t be a double blind study because the people using the copilot will know they’re using it.
Im_Unlucky t1_je6hn68 wrote
Reply to comment by LetGoAndBeReal in [D] The best way to train an LLM on company data by jaxolingo
This is the way ^
- shove data into vector db
- embed incoming query
- return topk matches from vector db
- inject those into prompt
- make api call to openai
- display response
- ???
- you are now a ML engineer
azorsenpai t1_je6hjpu wrote
Is there any reason you're really restraining to a Unet based model ? I'd recommend testing different architectures such as DeepLab V3 or FPN and see whether stuff improves. If it doesn't I'd recommend looking to your data and the quality of the ground truth as with only 100 data points you should be very much limited by the information contained in your data.
If the data is clean I'd recommend using some kind of ensemble method, this might be overkill, especially with heavy models but having multiple models with random initializations infer on a same input generally gives a few more points of accuracy/dice so if you really need it , this is an option.
Haycart t1_je6grih wrote
Reply to comment by ChuckSeven in [D] Very good article about the current limitations of GPT-n models by fripperML
>The Transformer is not a universal function approximator. This is simply shown by the fact that it cannot process arbitrary long input due to the finite context limitations.
We can be more specific, then: the transformer is a universal function approximator* on the space of sequences that fit within its context. I don't this distinction is necessarily relevant to the point I'm making, though.
*again with caveats regarding continuity etc.
>Your conclusion is not at all obvious or likely given your facts. They seem to be in hindsight given the strong performance of large models.
Guilty as charged, regarding hindsight. I won't claim to have predicted GPT-3's performance a-priori. That said, my point was never that the strong performance we've observed from recent LLMs was obvious or likely--only that it shouldn't be surprising. And, in particular it should not be surprising that a GPT model (not necessarily GPT-3 or 4) trained on a language modeling task would have the abilities we've seen. Everything we've seen falls well within the bounds of what transformers are theoretically capable of doing.
There are, of course, aspects of the current situation specifically that you can be surprised about. Maybe you're surprised that 100 billion-ish parameters is enough, or that the current volume of training data was sufficient. My argument is mostly aimed at claims along the lines of "GPT-n can't do X because transformers lack capability Y" or "GPT-n can't do X because it is only trained to model language".
detached-admin t1_je6gbz7 wrote
Don't do that if you're dealing with other people's money in any way. If it's yours, try langchain.
WokeAssBaller t1_je6fveg wrote
Reply to comment by lambertb in [D] GPT4 and coding problems by enryu42
I doubt it, I do pretty standard engineering, whats more likely is there is selection bias in the survey and people are overestimating it due to hype.
I'd love to see an actual double blind study.
xander76 OP t1_je6uh96 wrote
Reply to comment by reditum in [P] Imaginary programming: implementation-free TypeScript functions for GPT-powered web development by xander76
All great points! I tend to think that it's best suited to problems that you can't solve with traditional programming. If the problem you have is "reverse this array of numbers", then writing the code or having Copilot write the code is a better answer.
But if the problem you want to solve is "come up with good titles for this blog post" or "summarize this user email" or "categorize these customer service complaints by anger level", there really isn't a JavaScript/TypeScript function you can write to do that. In this case, I think the latency is often worth the functionality.
As to the non-determinism, I think that's a real issue. Right now, the state of the art of testing GPT prompts feels very shaky; one person I talked to said that they "change the prompt and then bang the keyboard to try four or five inputs". This clearly isn't ok for serious projects. To help with this, we're currently building some tools to help developers generate test inputs and evaluate how their imaginary functions perform.
ETA: and thanks for the comment about the name! (I can't take credit for it, though; I believe it was first coined by Shane Milligan.)