Recent comments in /f/MachineLearning
Deep-Station-1746 t1_je8u12c wrote
Reply to [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
This is interesting - compared to LoRa, it allows LLaMA to also accept images as inputs. And, I believe it is orthogonal to using LoRa. Meaning, they possibly can be used together. I'm unsure about the training stability though. I know that LoRa training allows ridiculously high learning rates (1e-5 for Text encoder), especially for dreambooth. Using LoRa for the frozen weights + LLaMA adapter is an interesting thing to explore.
Edit: spelling
3_Thumbs_Up t1_je8tvtc wrote
Reply to comment by sdmat in [R] The Debate Over Understanding in AI’s Large Language Models by currentscurrents
>We can test with things that are highly unlikely to be in the training data.
We can also test things where theres an infinite amount of alternatives so that memorization would be impossible.
If GPT could solve every arithmetic problem thrown at it, then it's obvious that it has developed some understanding of arithmetic, as it's simply impossible to memorize the answer for every possible problem.
However, the fact that it fails on arithmetic of large numbers could be an indication that it doesn't understand, but failure could also be caused by other factors, such as lack of enough working memory or similar (humans would fail at multiplying large numbers in their head as well).
So I think one could prove understanding, but proving lack of understanding seems harder.
CasulaScience t1_je8tqrr wrote
Reply to [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
In terms of practical application, is there any reason why someone would use this over Low Rank Adaptation?
Nous_AI t1_je8tcyp wrote
Fascinating.
idontcareaboutthenam t1_je8rz5a wrote
Reply to comment by -_1_2_3_- in [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
Can you elaborate?
Electric-Gecko t1_je8rtly wrote
I am not a programmer, though I have managed to compile programs on Linux before. Is there any way for me to contribute to this without programming knowledge, or detailed knowledge on the working of this software. It's based on reinforcement learning, right? Will I be able to train it?
fnordstar t1_je8rhex wrote
Reply to [D] Alternatives to fb Hydra? by alyflex
Isn't just using python flexible enough for you?
DigThatData t1_je8pm87 wrote
Reply to comment by dreaming_geometry in [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
I've decided to just lean into it and am literally just giving my ideas away. https://github.com/dmarx/bench-warmers
Cantareus t1_je8pk04 wrote
Reply to comment by yaosio in [Discussion] IsItBS: asking GPT to reflect x times will create a feedback loop that causes it to scrutinize itself x times? by RedditPolluter
There's no self-reflecting happening when the request to self-reflect is in the prompt. The improvement happens because the expected output after asking it to self-reflect is a more thought out response. You can get a type of reflecting by pasting the output back into the prompt with your own feedback.
-_1_2_3_- t1_je8p5xx wrote
Reply to comment by dreaming_geometry in [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
They are using gpt-4 to accelerate their work
sdmat t1_je8p59n wrote
Reply to comment by NotDoingResearch2 in [R] The Debate Over Understanding in AI’s Large Language Models by currentscurrents
I think we can safely say this is:
> it's essential to note that relying solely on jeans for habitat survivability would not be sufficient.
I don't have a degree in exojeanology, but the ideas all seem to at least be at the level of smart generalist brainstorming.
Specifically:
- Contextually appropriate - these are plausibly relevant to the needs of a Martian colony
- Nontrivial - no "wear them as clothing" (GPT3.5 did this)
- Logical and well articulated - each is a clearly expressed and internally consistent concept
- Passes the "common sense" test - no linguistically valid statements that are ridiculous if you have general knowledge of the world. E.g. "Use jeans to signal ships in orbit", or GPT3.5's suggestion to use jeans as radiation shielding because denim is a thick fabric.
They aren't necessarily good ideas in the sense that NASA should be writing this down. But that isn't the point.
I would argue that behaviourally GPT4 demonstrates a great deal of understanding here and a notable lack of the "brittleness and unhumanlike errors" that 3.5 shows on the same question.
elbiot t1_je8ngu2 wrote
Reply to comment by LetGoAndBeReal in [D] The best way to train an LLM on company data by jaxolingo
Huh? Have you never included text in a prompt and asked it to answer questions about the text? Seems like that counts as "new knowledge" by your definition
pseeth t1_je8mold wrote
Reply to [D] Alternatives to fb Hydra? by alyflex
I have a lightweight package that I use that has all the main things I wanted from hydra or gin-config. It's here and it's pretty tiny in terms of lines of code: https://github.com/pseeth/argbind
LetGoAndBeReal t1_je8m6y9 wrote
Reply to comment by light24bulbs in [D] The best way to train an LLM on company data by jaxolingo
Include new factual statements in your training data like “Joe Biden’s cat is named Fluffy.” Ask the model the name of Joe Biden’s cat before and after training and let us know the answers you get back. See if you get reliable answers across a set of data/questions.
Purplekeyboard t1_je8m61n wrote
Reply to comment by currentscurrents in [R] The Debate Over Understanding in AI’s Large Language Models by currentscurrents
> LLMs likely have a type of understanding, and humans have a different type of understanding.
Yes, this is more of a philosophy debate than anything else, hinging on the definition of the word "understanding". LLMs clearly have a type of understanding, but as they aren't conscious it is a different type than ours. Much as a chess program has a functional understanding of chess, but isn't aware and doesn't know that it is playing chess.
disastorm t1_je8lm7w wrote
Reply to [D] Simple Questions Thread by AutoModerator
I have a question about reinforcement learning, or more specifically gym-retro ( i know gym is pretty old now I guess ).
In the case of gym-retro, if you give a reward to the AI, are they actually looking at a set of variables and saying like "oh I pressed this button while all of these variables were these values and got this reward, so I should press it when all these variables are similar" or are they just saying like "oh I pressed this button and got this reward, so I should press it more often"?
NotDoingResearch2 t1_je8l7oz wrote
Reply to comment by sdmat in [R] The Debate Over Understanding in AI’s Large Language Models by currentscurrents
Is this accurate though? Serious question as I’m not an expert on the use of jeans on Mars.
hailfire27 t1_je8l7id wrote
Reply to comment by ghostfaceschiller in [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
I think he's talking about how during conversations, there are different cognitive levels to a conversation. You are basically having a conversation with yourself about what to say and remembering things to talk about, while at the same time considering the context of the situation, such as the environment or activity.
So he's saying for a model like this, would it be possible to tune the model so that it is able to give better answers in a conversation.
Purplekeyboard t1_je8l78y wrote
Reply to comment by currentscurrents in [R] The Debate Over Understanding in AI’s Large Language Models by currentscurrents
The point is that GPT-3 and GPT-4 can synthesize information to produce new information.
One question I like to ask large language models is "If there is a great white shark in my basement, is it safe for me to be upstairs?" This is a question no one has ever asked before, and answering the question requires more than just memorization.
Google Bard answered rather poorly, and said that I should get out of the house or attempt to hide in a closet. It seemed to be under the impression that the house was full of water and that the shark could swim through it.
GPT-3, at least the form of it I used when I asked it, said that I was safe because sharks can't climb stairs. Bing Chat, using GPT-4, was concerned that the shark could burst through the floorboards at me, because great white sharks can weigh as much as 5000 pounds. But all of these models are forced to put together various bits of information on sharks and houses in order to try to answer this entirely novel question.
SigmaSixShooter t1_je8kncb wrote
Reply to [D] Training a 65b LLaMA model by Business-Lead2679
I don’t have an answer for you, but as a fellow noobie, I’d love to hear how you did this. Any tips or resources you want to provide would be greatly appreciated.
_Arsenie_Boca_ t1_je8km8c wrote
Reply to [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
Very interesting work! Though I find the explanation of the concrete approach (how the additional parameters are used by the LM) to be a bit vague. Does anyone have a deeper understanding? Is it using regular adapters?
StudentBrilliant3388 t1_je8kijr wrote
Reply to [N] Predicting Finger Movement and Pressure with Machine Learning and Open Hardware Bracelet by turfptax
This is brilliant! And one of the very few projects I’ve seen here focused on rehabilitation or improving the lives of those struggling with new or chronic conditions, it makes me happy. 💗
I wish you all the best in your journey!
VinceD6 t1_je8kaza wrote
I am currently trying to do the same thing. Take a look at LlamaIndex, build a POC yesterday and it seemed to work really good.
LetGoAndBeReal t1_je8j7hw wrote
Reply to comment by elbiot in [D] The best way to train an LLM on company data by jaxolingo
The key word in that OpenAI link is “examples”. It says “more examples” and not “more knowledge”, because it’s referring to few shot training, which is about conditioning rather than providing new data.
In other words, if you want to get the model to classify sentiment of user comments as positive or negative, you can provide several examples in the prompt of both positive and negative comments. Fine-tuning allows you to provide many more such examples to the model than can fit in a prompt.
The key point is that through fine-tuning these examples can condition the model to classify sentiment but do not cause new facts to be absorbed by the model. You cannot get new facts to be readily absorbed through fine-tuning, which is why the OP should not look to fine-tuning to endow the model with the external dataset they want to use for question answering.
ahm_rimer t1_je8u2bi wrote
Reply to [R] LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention by floppy_llama
LoRA + PEFT + Zero-init attention adapter = 🤯