Recent comments in /f/MachineLearning
noobgolang t1_jcrvlfl wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
someone needs to take the plunge and release all of this altogether to the wild rather than this closed source nature
endless_sea_of_stars t1_jcrv26g wrote
Reply to comment by redpandabear77 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Outside of criticizing government or religion can you name an illegal topic anywhere?
yehiaserag t1_jcru1ty wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
Do you have a repo, a website, anywhere we can follow the progress of this?
[deleted] t1_jcrsk06 wrote
Reply to comment by starstruckmon in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
[deleted]
TheGuywithTehHat t1_jcrsjlo wrote
Reply to comment by Fender6969 in [D] Unit and Integration Testing for ML Pipelines by Fender6969
Most of that makes sense. The only thing I would be concerned about is the model training test. Firstly, a unit test should test the smallest possible unit. You should have many unit tests to test your model, and you should focus on those tests being as simple as possible. Nearly every function you write should have its own unit test, and no unit test should test more than one function. Secondly, there is an important difference between verification and validation testing. Verification testing shouldn't test for any particular accuracy threshold or anything like that, it should at most verify things like "model.fit() causes the model to change" or "a linear regression model that is all zeroes produces an output of zero." Verification testing is what you put on your CI pipeline to sanity check your code before it gets merged to master. Validation testing, however, should test model accuracy. It should go on your CD pipeline, and should validate that the model you're trying to push to production isn't low quality.
A1-Delta t1_jcrpd05 wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
Interesting project! I’ve seen many suggest that the training data for transfer learning might actually be the biggest thing holding Alpaca back from a ChatGPT like experience. In other words, that although the OpenAI model allows for the creation of a lot of training data, that data might include a lot of low quality pairs that in an ideal world wouldn’t be included. Do you have any plan to increase the quality of your dataset in addition to the size of it?
I hear your concern about the LLaMA license. It might be bad advice, but personally I wouldn’t worry about it. This is a very popular model people are using for all sorts of things. The chance they are going to come after you seems to me to be small and my understanding is that it’s sort of uncharted legal ground once you’ve done significant fine tuning. That being said, I’m not a lawyer.
LLaMA is a very powerful model and I would hate for you to put all this effort into creating something that ends up being limited and not clearly better than Alpaca simply because of license fears. If I were you though, I’d go with the 13B version. Still small enough to run on many high end consumer GPUs after quantization while providing significantly better baseline performance than the 7B version.
relevantmeemayhere t1_jcrp2rr wrote
Reply to comment by Temporary-Warning-34 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Honestly, really comes off as word salad lol.
I haven’t read the details, but it sounds like resampling in a serial learner?
relevantmeemayhere t1_jcrotun wrote
Reply to comment by MysteryInc152 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Mm, not really.
Bootstrapping is used to determine the standard error of estimates using resampling. From here we can derive tools like confidence intervals, or other interval estimates.
Generally speaking you do not use the bootstrap to tweak the parameters of your model. You use cross validation to do so.
baffo32 t1_jcronvh wrote
Reply to comment by Meddhouib10 in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
- offloading and accelerating (moving some parts to memory mapped disk or gpu ram, this can also make for quicker loading)
- pruning (removing parts of the model that didn’t end up impacting outputs after training)
- further quantization below 4 bits
- distilling to a mixture of experts?
- factoring and distilling parts out into heuristic algorithms?
- finetuning to specific tasks (e.g. distilling/pruning out all information related to non-relevant languages or domains) this would likely make it very small
EDIT:
- numerous techniques published in papers over the past few years
- distilling into an architecture not limited by e.g. a constraint of being feed forward
MysteryInc152 t1_jcro0q5 wrote
Reply to comment by Either-Job-341 in [P] The next generation of Stanford Alpaca by [deleted]
He's talking about the playground which is per token https://platform.openai.com/playground
Fender6969 OP t1_jcrnzzg wrote
Reply to comment by gdpoc in [D] Unit and Integration Testing for ML Pipelines by Fender6969
Many of the clients I support have rather sensitive data and persisting this into a repo would be a security risk. I suppose creating synthetic data would be the next best alternative.
Fender6969 OP t1_jcrnury wrote
Reply to comment by theAbominablySlowMan in [D] Unit and Integration Testing for ML Pipelines by Fender6969
Yeah I am always uncomfortable pushing untested code to Production. I think I have some good ideas for what to add to my CI pipeline regarding unit tests.
MysteryInc152 t1_jcrnqc8 wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
You can try training chatGLM. 6b parameters and initially trained on 1T English/Chinese Tokens. Also completely open source. However, it's already been fine tuned and had RLHF but that was optimized for Chinese Q/A. Could use some English work,
Another option is RWKV. There are 7b and 14b models(I would go with the 14b, it's the better of the two) fine tuned to a context length of 8196 tokens. He plans on increasing context further too.
Fender6969 OP t1_jcrnppi wrote
Reply to comment by TheGuywithTehHat in [D] Unit and Integration Testing for ML Pipelines by Fender6969
Thanks for the response. I think hardcoding things might make the most sense. Ignoring testing the actual data for a minute, let us say I have an ML pipeline with the following units:
- Data Engineering: method that queries data, performs further aggregation in Pandas/PySpark
- Unit test: hardcode an input to pass into this function and leverage Pytest/unittest to check for the exact output'
- Model Training: method that engineers features and passes data into Sklearn pipeline, which scales/encodes data and trains ML model
- Unit test: check for successful predictions on training data to a degree of accuracy based on your evaluation metric
- Model Serving: first method that performs ETL for prediction data and second method that loads Sklearn pipeline object to serve prediction
- Unit test:
- Module 1: same as Data Engineering
- Module 2: check for successful predictions
- Unit test:
Does the above unit tests make sense to add in a CI pipeline?
redpandabear77 t1_jcrng6h wrote
Reply to comment by gronaninjan in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Name one forbidden topic in China that doesn't have to do with criticizing the government.
[deleted] OP t1_jcrnbs3 wrote
Reply to comment by Either-Job-341 in [P] The next generation of Stanford Alpaca by [deleted]
[deleted]
username001999 t1_jcrn1aq wrote
Reply to comment by BalorNG in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
We Americans live in a country where kids are regularly gunned down in school so we make ourselves feel better by making jokes about how much worse other countries are for events that happened over 30 years ago. Or we don’t even know our own history, like the Kent State Massacre.
Either-Job-341 t1_jcrhysc wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
OpenAI API costs based on how many tokens you use, isn't that the case? Afaik, the fixed price (20$) is for the case when you're using it via UI (probably max one session).
RemindMeBot t1_jcrhq4b wrote
Reply to comment by Either-Job-341 in [P] nanoT5 - Inspired by Jonas Geiping's Cramming and Andrej Karpathy's nanoGPT, we fill the gap of a repository for pre-training T5-style "LLMs" under a limited budget in PyTorch by korec1234
I will be messaging you in 3 days on 2023-03-22 00:14:14 UTC to remind you of this link
CLICK THIS LINK to send a PM to also be reminded and to reduce spam.
^(Parent commenter can ) ^(delete this message to hide from others.)
| ^(Info) | ^(Custom) | ^(Your Reminders) | ^(Feedback) |
|---|
RemindMeBot t1_jcrhfug wrote
Reply to comment by Either-Job-341 in [P] The next generation of Stanford Alpaca by [deleted]
I will be messaging you in 2 days on 2023-03-21 00:12:32 UTC to remind you of this link
10 OTHERS CLICKED THIS LINK to send a PM to also be reminded and to reduce spam.
^(Parent commenter can ) ^(delete this message to hide from others.)
| ^(Info) | ^(Custom) | ^(Your Reminders) | ^(Feedback) |
|---|
Either-Job-341 t1_jcrhe3r wrote
Reply to [P] The next generation of Stanford Alpaca by [deleted]
RemindMe! 2 days
DreamMidnight t1_jcrh53z wrote
Reply to comment by LeN3rd in [D] Simple Questions Thread by AutoModerator
Here are some sources:
https://home.csulb.edu/~msaintg/ppa696/696regmx.htm
https://developers.google.com/machine-learning/data-prep/construct/collect/data-size-quality (order of magnitude in this case means 10)
Philpax t1_jcrgxbb wrote
Reply to [D] LLama model 65B - pay per prompt by MBle
As the other commenter said, it's unlikely anyone will advertise a service like this as LLaMA's license terms don't allow for it. In your situation, I'd just rent a cloud GPU server (Lambda Labs etc) and test the models you care about. It'll only end up being a dollar or two if you're quick with your use.
wyhauyeung1 t1_jcrvnxz wrote
Reply to [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
I successfully deployed in my local PC and run. Just wondering, where is the model file stored after install? It seems I could not find any big files under the directory