Recent comments in /f/MachineLearning
robobub t1_jdsria4 wrote
Reply to comment by liqui_date_me in [D] GPT4 and coding problems by enryu42
While GPT-4 is autoregressive, it takes into account the tokens it has chosen to generate incrementally. So it is only limited to O(1) if it attempts to answer with the correct answer immediately. It can in theory take O(m) steps, where m is the number of intermediate tokens it predicts.
BeautifulLazy5257 t1_jdsr09g wrote
Reply to comment by ghostfaceschiller in [D] GPT4 and coding problems by enryu42
I was wondering if you knew the trick to ReAct without langchain.
For instance, memory is just passing the past conversations through the prompt as context. There's nothing programtic about it. You don't need the langchain library, you just have to craft the right prompt
I think that using langchain kind of obscures how the model is actually achieving the desired outputs.
Having models interact with pdfs ultimately is just turning a pdf into a string and passing the string as context while adding a prompt to help prime the model.
I'll look into CoT and look through the ReAct sourcecode, but I'm going to avoid the use of langchain for most stuff or even looking at ReAct documentation, since those docs are only going to tell me how to use those libraries and not tell me how to achieve the effect from scratch.
Edit:
This is a pretty clear overview of CoT. Very compelling as well.
https://ai.googleblog.com/2022/05/language-models-perform-reasoning-via.html?m=1
I guess I'll start AB testing some prompts to breakdown problems and tool selections.
If you have any more input on particular prompts you've used, I'd be grateful.
Edit 2: https://www.youtube.com/watch?v=XV1RXLPIVlw&ab_channel=code_your_own_AI It can't get clearer than this. great video
Cwlrs t1_jdsqyyq wrote
Reply to [D] GPT4 and coding problems by enryu42
It's performing really well for my project. Online web app game in python flask socketio
CobaltAlchemist t1_jdsqr5e wrote
Reply to comment by Hamoodzstyle in [D] GPT4 and coding problems by enryu42
(not op) I've found that asking it directly if it understands helps to bridge any gaps I miss. It's asked me clarifying questions afterward in the past that I hadnt thought about
Alternatively, when I assume it understands sometimes it comes up with some real wild stuff because I wasn't clear
mxby7e t1_jdsqijp wrote
Reply to Have deepfakes become so realistic that they can fool people into thinking they are genuine? [D] by [deleted]
I got bored last night and took a look what is out there. Between diffusion and GAN workflows, you can deepfake almost anything you want in any style with just a little technical background.
You can easily take a real photo and use inpainting to replace any aspect of the image, then run it through a few img2img loops to balance the composition. You can train a subject finetune with a handful of pictures and a few hours of training time.
You can use consumer face swap tools to swap faces into any image you want.
Midjourney v5 can generate images that are hard to differentiate from real photos.
farmingvillein t1_jdspflq wrote
Reply to comment by enryu42 in [D] GPT4 and coding problems by enryu42
So, don't know whether this actually makes a difference, but I'd review the overall post-conversion text.
E.g.: https://github.com/enryu43/llm_coding/blob/main/atcoder_eval/statements/statement_abc293_b.txt
You'll see that it represent "K" and "N" wrong here (in sample 1, 15 versus 5, 12 versus 2).
Certainly, as a human, I would find this confusing. Maybe you could get some automated robustness by telling it how you converted the text (as it might automatically adjust its "expectations" on interpreting the numbers). Obviously, the fairer comparison though would just be to fix this.
> as they require coming up with some ideas before writing the code.
The other thing I'd note--
Not sure whether you're using the API directly, but if I play around with these in ChatGPT, I often run into the context window and have to nurse it along to complete text. I'd make sure that however you're running things, you're giving it enough "space" to iterate (particularly if you use any reflection techniques).
TheStartIs2019 t1_jdspe6e wrote
Reply to comment by ghostfaceschiller in [D] GPT4 and coding problems by enryu42
It gets better! https://arxiv.org/abs/2303.11366
Various_Ad7388 t1_jdsp2q8 wrote
Reply to [D] Simple Questions Thread by AutoModerator
Hey @all if I am just starting off in machine learning what should I learn first Tensorflow or PyTorch or other?? Also once Im more experienced where do I go from there?
[deleted] OP t1_jdsor6g wrote
Reply to comment by addandsubtract in Have deepfakes become so realistic that they can fool people into thinking they are genuine? [D] by [deleted]
[removed]
enryu42 OP t1_jdsokwz wrote
Reply to comment by farmingvillein in [D] GPT4 and coding problems by enryu42
Interesting! Here are the scraped and auto-converted statements (formatting is off sometimes, especially in the sample tests, but understandable). Prefixes are: "abc" for beginner, "arc" for regular, "agc" for "grand".
I do believe that the "Beginner" ones can be improved, but it'll be interesting to see what happens on "Grand" (or even "Regular"), as they require coming up with some ideas before writing the code.
blose1 t1_jdso3tt wrote
Reply to comment by ghostfaceschiller in [D] GPT4 and coding problems by enryu42
I literally told you my use case and it failed on that and it failed on similar problem 1-2 months ago when I was using 3.5 version, for my class of problems nothing changes, it fails the same way. I think you have your eyes shut and not reading what people write. I'm not talking about easy CRUD problems that you can find thousands of solutions online, ChatGPT is doing ok on these kind of tasks and it solved a lot of them for me too.
mcilrain t1_jdsnxkr wrote
Reply to comment by Secure-Fix-6355 in [D] GPT4 and coding problems by enryu42
Who asked?
ghostfaceschiller t1_jdsnev1 wrote
Reply to comment by blose1 in [D] GPT4 and coding problems by enryu42
This line of thinking sounds sillier and sillier every week. Its like talking to someone who has had their eyes shut and fingers in their ears for the last two months.
EDIT: and tbc, i'm not trying to argue that it isn't statistics-based/trained on the internet/etc. I'm saying that it turns out that kind of system is powerful & capable than we ever would have intuitively thought it would be
farmingvillein t1_jdsmsh9 wrote
Reply to comment by E_Snap in [D] GPT4 and coding problems by enryu42
Gotcha. Yeah, that is presumably where the power of inner monologue / step-by-step / reflection come from.
Will be cool to see that (presumably) progressively systematized.
farmingvillein t1_jdsmdt2 wrote
Reply to comment by nixed9 in [D] GPT4 and coding problems by enryu42
- This isn't really an accurate summary of the Reflexion paper. As noted in the other post:
> Eh, I must've misunderstood the paper. It sounded like they were asking GPT4 to create unit tests, execute the code, and then update its answer based on the results of those unit tests.
This version is correct.
- However, if I do the above and I throw in a semi-random Beginner problem that failed in OP's original pass-through, it successfully builds the answer.
u/enryu42 -- if you care to take things forward, I'd try implementing Reflexion (either with the underlying codebase (https://github.com/noahshinn024/reflexion-human-eval/) or just manual prompt work.
Or if you can provide a link to the problems in copy-pastable text form (manually coercing the math notation is a little painful), since you presumably already did this, it would greatly accelerate others hopping on analysis.
The fact that I immediately saw improvement on a randomly-selected (Beginner) problem suggests that there is a bunch of upward room here.
Jeffy29 t1_jdsm90r wrote
Reply to comment by enryu42 in [D] GPT4 and coding problems by enryu42
>But I strongly doubt it'll help much: it's not that the solutions have minor bugs, they're usually just completely wrong
I strongly doubt that it wouldn't help. I haven't tested GPT-4 in coding but from what I've seen GPT-3 makes a number of simple errors, especially in longer complex code it's almost inevitable. But it's able to quickly identify and correct it when you point it out. GPT-4 not being able to compile and test its own code that is a big limitation that humans don't have. It also can't calculate the math, it's essentially guessing the calculation, but both can be addressed with an external compiler and calculator like Wolfram. Something humans also have access to. There would need to be some time limit imposed so it can't brute force the solution after guessing for a few days but even so I think the improvements would be quite large.
Yecuken t1_jdsm4w1 wrote
Reply to comment by currentscurrents in [D] GPT4 and coding problems by enryu42
Tests would not help against optimization, models will just learn how to pass the test. Optimization will always win against any problem with a known solution
farmingvillein t1_jdsm0hw wrote
Reply to comment by muskoxnotverydirty in [D] GPT4 and coding problems by enryu42
No, you didn't misunderstand it--your understanding is correct. OP is giving an answer that is similar to part of the Reflexion paper, but not the entirety.
imaginethezmell t1_jdsksw5 wrote
Reply to comment by addition in [D] GPT4 and coding problems by enryu42
also people keep thinking it is just one thing, but it is actually an infinite thing
you can have a bot for everything all the way down
bot to create the idea + bot that reviews the ideas + bot that finds if the idea exists + bot that adds use cases to each general idea...a bot that decides the best idea
bot to create the outline/write/code + bot that reviews/QA each part
and btw each part doesnt have to be done at once either
you can start with a single bot doing a simple sub task, then another one the next one, an assembling bot adding them together, while the review bot verifies it
with a set of connections to the api, that can be done np today
no human task cannot be cut into enough sub steps that the army of bots cannot do it little by little
some tasks 1 bot can do most in 1 shot
endless_sea_of_stars t1_jdskiit wrote
Reply to comment by light24bulbs in [P] Using ChatGPT plugins with LLaMA by balthierwings
The advantage of in context learning is that it is trivial to add and remove plug-ins.
Training with the plug-ins is more powerful, but you can't really easily add or subtract. In theory training with APIs should result in a smaller model as the main model no longer needs to learn math or trivia (in theory).
blose1 t1_jdskab0 wrote
Reply to comment by enryu42 in [D] GPT4 and coding problems by enryu42
These models have access to all human knowledge, all scientific papers, books etc. If I would have such a knowledge I could solve any Olympiad tasks.
Secure-Fix-6355 t1_jdsjqnn wrote
Reply to comment by E_Snap in [D] GPT4 and coding problems by enryu42
I have no idea what that is and I'm not farming Karma, I'm abusing you
E_Snap t1_jdsjku6 wrote
Reply to comment by Secure-Fix-6355 in [D] GPT4 and coding problems by enryu42
Says the guy with a karma farm account name. Guess you have to get those low effort internet points somehow, huh?
VelveteenAmbush t1_jdsjab4 wrote
Reply to comment by artsybashev in [R] Reflexion: an autonomous agent with dynamic memory and self-reflection - Noah Shinn et al 2023 Northeastern University Boston - Outperforms GPT-4 on HumanEval accuracy (0.67 --> 0.88)! by Singularian2501
Also an LLM to read all of the tldrs and tell me which of them I should pay attention to.
night81 t1_jdsrnvi wrote
Reply to comment by artsybashev in [D] GPT4 and coding problems by enryu42
There are significant challenges to that hypothesis. https://iep.utm.edu/hard-problem-of-conciousness/