Recent comments in /f/MachineLearning
squareOfTwo t1_jctkqt2 wrote
Reply to [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
someone had to do this as OSS MIT, great, thx!
danielbln t1_jctk7sy wrote
Reply to comment by currentscurrents in [D] LLama model 65B - pay per prompt by MBle
Pennies per request would be a lot, it's a fraction of a penny per request.
michaelthwan_ai OP t1_jcthy7x wrote
Reply to comment by taci7c0ff33 in [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Thank you! :D
EcstaticStruggle t1_jcthdzz wrote
Reply to comment by josejo9423 in [D] Simple Questions Thread by AutoModerator
Thanks. This was something I tried earlier. I noticed that using the maximum number of estimators almost always lead to the highest cross validation score. I was worried there would be some overfitting as a result.
taci7c0ff33 t1_jcth2g9 wrote
Reply to [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Very cool!
emotionalfool123 t1_jctg58u wrote
Reply to comment by Tr4sHCr4fT in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Thanks for letting me know a better way.
visarga t1_jctfir1 wrote
Reply to comment by relevantmeemayhere in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Human Feedback is being boostsrapped by GPT3 predictions "stolen" against OpenAI's will (for just $500 API bills).
clueless1245 t1_jcteoda wrote
Reply to comment by farmingvillein in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
Open source doesn't mean freely used. That's the whole reason there's an F in FOSS.
starstruckmon t1_jcte34d wrote
Reply to comment by Taenk in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
🤷
Sometimes models just come out crap. Like BLOOM which has almost the same number of parameters as GPT3, but is absolute garbage in any practical use case. Like a kid from two smart parents that turns out dumb. Just blind chance.
Or they could be wrong. 🤷
Taenk t1_jctdmvi wrote
Reply to comment by starstruckmon in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
I haven’t tried the larger models unfortunately. However I wonder how the model could be „shockingly bad“ despite having almost three times the parameter count.
tomd_96 t1_jctddsu wrote
Reply to [D] LLama model 65B - pay per prompt by MBle
You can do this using replicate: https://github.com/replicate/cog-llama
RoyalCities t1_jctcu1m wrote
Reply to comment by starstruckmon in [P] The next generation of Stanford Alpaca by [deleted]
Couldnt it be possible to set up a large community Q/A repositiry then? Just crowdsource whatever it outputs and document collectively.
philipgutjahr t1_jctbs35 wrote
Reply to comment by Smallpaul in [P] The next generation of Stanford Alpaca by [deleted]
which can make a huge difference: GPT-3 + RLHF = Chat-GPT
schorhr t1_jctb6tz wrote
Reply to comment by simpleuserhere in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
Okay. I don't have the capacity right now (old laptop, disk too small to really use a second OS). I appreciate the help! I will once I get a new computer.
luaks1337 t1_jctagcz wrote
Reply to comment by emotionalfool123 in [R] ChatGLM-6B - an open source 6.2 billion parameter Eng/Chinese bilingual LLM trained on 1T tokens, supplemented by supervised fine-tuning, feedback bootstrap, and RLHF. Runs on consumer grade GPUs by MysteryInc152
In German this command could be interpreted as "you son of a whore"
simpleuserhere OP t1_jct9btk wrote
Reply to comment by schorhr in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
For Android build please use Linux ( tested with Ubuntu 20.04)
philipgutjahr t1_jct7e53 wrote
Reply to comment by [deleted] in [P] The next generation of Stanford Alpaca by [deleted]
LGTM
LeN3rd t1_jct6arv wrote
Reply to comment by DreamMidnight in [D] Simple Questions Thread by AutoModerator
Ok, so all of these are linear ( logistics) regression models, for which it makes sense to have more data points, because the weights aren't as constraint as in a convolutional layer I.e. but it is still a rule of thumb, not exactly a proof.
[deleted] OP t1_jct6a1x wrote
Reply to comment by VertexMachine in [P] The next generation of Stanford Alpaca by [deleted]
Provide a clear and formal way for the community to express its opinion. You know, as opposed to tweeting at one person who does not have absolute control over Meta AI. Notable people have brought attention to that pull request and it is currently gaining traction.
schorhr t1_jct58nc wrote
Reply to comment by simpleuserhere in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
Thanks!
Both of the instructions (for Android which I'm attempting, but also the Windows instructions) result with the > C:/MinGW-32/bin/make.exe: invalid option -- ? error. I can't seem to figure out what make version I should use instead, or how to edit that.
michaelthwan_ai OP t1_jct4sdj wrote
Reply to [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Demo page: https://searchgpt-demo.herokuapp.com/
Github : https://github.com/michaelthwan/searchGPT
searchGPT is a search engine or question-answer bot based on LLM to give natural language answers. You may see the footnote which is the reference of sources from the web. Below there is a explainability view to show how the response is related to the sources.
Why Grounded though?
Because it is impossible for the LLM to learn everything during the training, thus real-time factual information is needed for reference.
This project tried to reproduce work like Bing and perplexity AI which have external references to support the answer of LLM.
Some examples of good grounded answer from searchGPT and wrong ungrounded answer from ChatGPT is mentioned in the github.
simpleuserhere OP t1_jct4k2z wrote
Reply to comment by schorhr in [Research] Alpaca 7B language model running on my Pixel 7 by simpleuserhere
I have updated readme with Windows build instructions,please check https://github.com/rupeshs/alpaca.cpp#windows
veonua t1_jct419t wrote
Reply to comment by currentscurrents in [D] LLama model 65B - pay per prompt by MBle
Creating a monopoly on AI can be extremely risky. Although OpenAI was founded to prevent it, recent actions by the company suggest that they may be contributing to monopolization by reducing prices.
ProbablyBunchofAtoms t1_jctnndu wrote
Reply to [P] searchGPT - a bing-like LLM-based Grounded Search Engine (with Demo, github) by michaelthwan_ai
Looks dope