Recent comments in /f/MachineLearning
VelveteenAmbush t1_jd0j2yv wrote
Reply to comment by Carrasco_Santo in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
Assuming that the best corporate models don't have further improvements in architecture and methodology that haven't been shared publicly...
Radiant_Rhino t1_jd0gncq wrote
Could you please share the server invitation link with me? 🙏🏻
And is version 65b publicly available on Discord?
rolexpo t1_jd0fvle wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
You'll have better luck waiting for Intel
currentscurrents t1_jd0f76v wrote
Reply to comment by Educational-Net303 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
I mean of course not, nobody would make such a game right now because there are no >24GB cards to run it on.
Siltala t1_jd0e1b3 wrote
Reply to [P] TherapistGPT by SmackMyPitchHup
This is a privacy nightmare
pointer_to_null t1_jd0bv74 wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
This is definitely true. Theoretically you can page stuff in/out of VRAM to run larger models, but you won't be getting much benefit over CPU compute with all that thrashing.
wywywywy t1_jd09nb5 wrote
HinaCh4n t1_jd09h1q wrote
Reply to comment by pixiegirl417 in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
Got a link to the leaderboard?
PavKon t1_jd09am3 wrote
Reply to comment by SmackMyPitchHup in [P] TherapistGPT by SmackMyPitchHup
Your early version on Github seem to be using OAIs Davinci model. Also weirdly it also includes an avatar of "Birmingham Counseling for Women and Girls". What a bizarre project you have there.
Straight-Comb-6956 t1_jd08cq1 wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
LLaMa/Alpaca work just fine on CPU with llama.cpp/alpaca.cpp. Not very snappy(1-15 tokens/s depending on model size), but fast enough for me.
SomeLongWindedIdiot t1_jd07i7z wrote
Reply to [D] Simple Questions Thread by AutoModerator
Why is AI safety not a major topic of discussion here and in similar communities?
I apologize if the non-technical nature of my question is inappropriate for the sub, but as you’ll see from my comment I think this is very important.
I have been studying AI more and more over the past months (for perspective on my level that consists of Andrew Ng’s Deep Learning course, Kaggle competitions and simple projects, reading a few landmark papers and digging into transformers) The more I learn, the more I am both concerned and hopeful. It seems all but certain to me that AI will completely change life as we know it in the next few decades, quite possibly the next few years if the current pace of progression continues. It could change life to something much, much better or much, much worse based on who develops it and how safely they do it.
To me safety is far and away to most important subfield in AI now, but is one of the least discussed. Even if you think there is a low chance of AI going haywire on its own, in my admittedly very non-expert view it’s obvious that we should be also concerned about the judgment and motives of the people developing and controlling the most powerful AIs, and the risks of such powerful tools being accessible to everyone. At the very least I would want discussion on actionable things we can all do as individuals.
I feel a strong sense of duty to do what I can, even if that’s not much. I want to donate a percentage of my salary to funding AI safety, and I am looking whether I can effectively contribute with work to any AI safety organizations. I have a few of my own ideas along these lines; does anyone have any suggestions? I think we should also discuss ways to shift the incentives of major AI organizations. Maybe there isn’t a ton we can do (although there are a LOT of people looking, there is room for a major movement), but it’s certainly not zero.
djmaxm t1_jd05tgt wrote
Reply to [D] Simple Questions Thread by AutoModerator
I have a 4090 with 32GB of system RAM, but I am unable to run the 30B model because it exhausts the system memory and crashes. Is this expected? Do I need a bunch more RAM? Or am I doing something dumb and running the wrong model. I don't understand how the torrent model, the huggingface model, and the .pt file relate to each other...
Educational-Net303 t1_jd05hmc wrote
Reply to comment by I_will_delete_myself in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Are we still taking consumer grade hardware or specialized GPU made for a niche crowd?
I_will_delete_myself t1_jd05atn wrote
Reply to comment by Educational-Net303 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Now try that on 2-4 monitors. You would be surprised how premium gamers like their hardware. It’s like checking out sports cars but for nerds like me.
pixiegirl417 OP t1_jd0571w wrote
Reply to comment by Carrasco_Santo in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
People can contribute to human reinforcement training here: https://open-assistant.io/. There's a leaderboard :)
Educational-Net303 t1_jd051kh wrote
Reply to comment by I_will_delete_myself in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Cyberpunk on max with psycho takes ~16gb max. It's gonna be a few years before we actually see games demanding more than 24.
pixiegirl417 OP t1_jd04sxc wrote
Reply to comment by timedacorn369 in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
That's right!! Model is here: https://huggingface.co/OpenAssistant/oasst-sft-1-pythia-12b
I_will_delete_myself t1_jd04mia wrote
Reply to comment by Educational-Net303 in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
people are demanding more and more interactivity in their video games (look at the trend of open worlds). It’s only gonna get bigger.
Educational-Net303 t1_jd03se1 wrote
Reply to comment by I_will_delete_myself in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
What game is limited by vram? I haven't heard of any game running over 24gb unless it's Skyrim with a bunch of 8k mods
LetMeGuessYourAlts t1_jd02jkq wrote
Reply to comment by gybemeister in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
Used availability is better on the 3090 as well. I got one for $740 on eBay. Little dust on the heatsinks but at half price it was a steal.
ertgbnm t1_jd028k5 wrote
Reply to [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
I heard 30B isn't very good. Anyone with experience disagree?
nolimyn t1_jd01nm3 wrote
Reply to comment by kross00 in [D] Best ChatBot that can be run locally? by rustymonster2000
the LoRA is like a modular refinement of the base language model, in this case it's the part that makes it feel like a chatbot / assistant, and makes it follow instructions.
you can see the same concept over at civitai.com, filter by LoRAs. Something like a LoRA for one character can be run on different checkpoints that focus on photorealism or anime, etc.
wind_dude t1_jd012ru wrote
I'm not big into image generation, but... some thoughts...
- SSIM - I believe the issue here has to due with the quality of the img captions. Perhaps merging captions on images
- could try training boolean classifiers for both images and captions, `is_junk`, and than using that model to remove junk from the training data.
Thiago_Von_Duck t1_jd00vqq wrote
Reply to comment by timedacorn369 in [P] OpenAssistant is now live on reddit (Open Source ChatGPT alternative) by pixiegirl417
That would you mean self-hosted
whyvitamins t1_jd0j5zq wrote
Reply to comment by currentscurrents in [Project] Alpaca-30B: Facebook's 30b parameter LLaMa fine-tuned on the Alpaca dataset by imgonnarelph
> hope that AMD gets their act together on AI support
walking around picking up coins from the ground to buy a 3090 should be faster honestly