Recent comments in /f/MachineLearning
Quail-That t1_jcsgkc4 wrote
Reply to comment by username001999 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Not knowing and not being allowed to know are radically different things. If you want to conflate the two, you are acting in bad faith.
frownGuy12 t1_jcsfnh7 wrote
Reply to comment by ThatInternetGuy in [P] The next generation of Stanford Alpaca by [deleted]
If OpenAI wants people to respect their IP they should take the word “open” out of their name. They scraped our data to train their models after all, it’s not like OpenAI themselves aren’t pushing the boundaries of what’s acceptable when it comes to copyright law.
Legally it’s questionable, but ethically speaking I think it’s a fine idea.
NotARedditUser3 t1_jcsc9lp wrote
Reply to [D] LLama model 65B - pay per prompt by MBle
You can get llama running on consumer grade hardware. There's 4 and 8 bit quantization for it i believe where it fits in a normal gpu's vram, i saw floating around here
360macky t1_jcsanpu wrote
Reply to comment by Taenk in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
Thanks!
Smallpaul t1_jcsah9r wrote
Reply to comment by RoyalCities in [P] The next generation of Stanford Alpaca by [deleted]
I think the new model gets most of its knowledge from its original model and the training is mostly about how to act like a RLHF model.
Riboflavius t1_jcs7afw wrote
Reply to comment by the320x200 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Pretty sure whoever knows what happened to Jimmy Hoffa made sure they kept their trap shut in public… ;)
Taenk t1_jcs5eon wrote
Reply to comment by legendofbrando in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
A proper port to the neural engine would be especially interesting. There was one by Apple for Stable Diffusion.
Taenk t1_jcs53iw wrote
Reply to comment by starstruckmon in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
The results for LLaMA-33B quantised to 3bit are rather interesting. That would be an extremely potent LLM capable of running on consumer hardware. Pity that there are no test results for the 2bit version.
mattrobs t1_jcs3vvo wrote
Reply to comment by v_krishna in [R] Stanford-Alpaca 7B model (an instruction tuned version of LLaMA) performs as well as text-davinci-003 by dojoteef
Have you tried GPT4? It’s been quite resilient in my testing
[deleted] t1_jcs3lrc wrote
[deleted] t1_jcs3icv wrote
ReasonablyBadass t1_jcs32ea wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
Careful. That MIT license won't work, I think, thanks to ClosedAIs licences
ThatInternetGuy t1_jcs253z wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
You know ChatGPT and GPT4 licenses forbid using their output data for training competing AI models. What Stanford did was to show proof of concept for their paper, not to open-source the model, at all.
yehiaserag t1_jcrz2qt wrote
Reply to comment by yehiaserag in [P] The next generation of Stanford Alpaca by [deleted]
seems very similar
MysteryInc152 OP t1_jcrz16i wrote
Reply to comment by MisterManuscript in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Yeah I'm wrong it seems. Read a few articles using bootstrapping in the definition I used so I assumed that was generally it.
MisterManuscript t1_jcry6cj wrote
Reply to comment by MysteryInc152 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
That's not what bootstrapping is, it is a resampling technique used to create multiple datasets of the same size from the original dataset using random sampling with replacement. It is done to get the estimate of the standard deviation of a desired variable.
Here's the link to the ISLR textbook. The bootstrap chapter will verify what it is.
RoyalCities t1_jcrxlvr wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
I was talking to GPT 4 about this and it said that it seems plausible and can dramatically bring down costs.
It called it "knowledge distillation"
It also mentioned that if we had access to the weights from open ai you can use a process called model compression to scale down the hardware and put it on less powerful gpus or distributed gpus (like how render farms work)
This also explains why open ai is so cagey on releasing weights - the initial training cost is where the money sink is but once weights are out their is ways to make it run on cheaper hardware.
But Im wondering does this mean the smaller model can ONLY respond to the questions your generating or will it have latent knowledge outside of just the knowledge transfer? Like would say the smaller model thats trained off this approach also be able to answer questions on topics that are "restricted" in open ais view that you couldnt ask it or do you absolutely must need to get an initial answer for such restricted content for it to be able to produce a responce?
Talking about things like writing malicious code or what not. I dont plan on doing that obviously but Im curious on if this means that these smaller models will basically be totally unrestricted now or if its just trained on say tons of python code it can just create said malicious code from scratch without actually being exposed with examples of "how" to make it (since it has a greater knowledge of the ubderlying principals of python)
Edit: Okay guess it can per GPT 4.
Damn these things are fascinating.
>Yes, the same concerns can apply to a smaller model being trained from a larger one via knowledge distillation. Knowledge distillation is a technique where the smaller model learns to mimic the larger model's behavior by training on a dataset generated using the larger model's outputs. The smaller model effectively learns from the larger model's knowledge and understanding of language patterns and concepts.
>As a result, the smaller model can also gain latent knowledge about various topics and domains, even if it hasn't been explicitly exposed to specific examples during training. This means that the smaller model could potentially generate undesirable content based on its understanding of the relationships between words and concepts, similar to the larger model.
Craiglbl t1_jcrxjy2 wrote
Reply to comment by MysteryInc152 in [P] The next generation of Stanford Alpaca by [deleted]
ChatGLM is really good. I sometimes have a hard time distinguishing its Chinese outputs from those of chatgpt.
Sadly its English could use some improvement as it usually use Chinese adjectives when similar words are lacking in English.
MisterManuscript t1_jcrwvc8 wrote
Reply to comment by Temporary-Warning-34 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
I tried googling it, it's is a nonexistent terminology in the realm of statistics. I know what bootstrapping is, but not this version of it.
It's better to ask the GitHub authors about this to make sure they're not just spitting out pseudostatistical terminology.
Addendum: another guy did query the authors regarding this terminology in the issues tab, they did not respond.
Fender6969 OP t1_jcrw759 wrote
Reply to comment by TheGuywithTehHat in [D] Unit and Integration Testing for ML Pipelines by Fender6969
This makes perfect sense thank you. I’m going to think through this further. If you have any suggestions for verification/sanity testing for any of the components listed above please let me know.
GaggiX t1_jcrvtz6 wrote
Reply to comment by redpandabear77 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
I mean, even the Taiwan flag emoji is banned on Chinese phones lmao
BawkSoup t1_jcsgyoe wrote
Reply to comment by username001999 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Okay, tankie. Keep it about machine learning.