Recent comments in /f/MachineLearning
TehDing t1_jdz3anl wrote
Reply to [P] two copies of gpt-3.5 (one playing as the oracle, and another as the guesser) performs poorly on the game of 20 Questions (68/1823). by evanthebouncy
Similarly, sucks at Wordle
quantic-dream t1_jdz2gq9 wrote
Reply to comment by MysteryInc152 in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
I am noob in ML, but as I understand GPT generates 1 word at a time, and may it be that, for ex, 1 particular word, somewhere in the middle was with confidence 0.1, the highest GPT can get, and everything after this word become a hallucination?
Professional-Gap-243 t1_jdz27jp wrote
Reply to [D] FOMO on the rapid pace of LLMs by 00001746
The way I think about this is like I think about OS. Yes you can build your own OS from scratch, but more often than not you just use windows or Linux. And if you need something custom it is often sufficient to setup your own Linux distro.
To me LLMs are in a similar situation. It doesn't really make a sense to build your own LLM from scratch most of the time just like it wouldn't to build your own OS. This doesn't mean that there is no space for building new LLMs tho.
GPT is in this example like windows (closed, controlled by a corporation) and I think the ML community now needs to focus on building open source alternative that could stand toe to toe with it.
Otherwise the space becomes monopolistic/oligopolistic with large corps running the show (just like before Linux came around).
Rioghasarig t1_jdz24za wrote
Reply to comment by astrange in [D]GPT-4 might be able to tell you if it hallucinated by Cool_Abbreviations_9
People were using the base model when it first came out and some people are still using it today. The game AI Dungeon is still runs on what is essentially a transformer trained on next token prediction. So it would be accurate to say "It's just (attempts to) outputs the next most probable word" .
Nezarah t1_jdz1zqc wrote
For specifically personal use and research? And not commercial? LlaMA is a good place to start, and/or Alpaca 7B. Small scale (can run on most hardware locally), can be Lora trained and fine-tuned. Also has High token limits (I think it’s 2000 or so?).
Can have outputs comparable to GPT3 which can be further enhanced with Pre-Context training.
Can add branching functionality through the Langchain library.
starfries t1_jdz0q2b wrote
Reply to comment by sdmat in [D] FOMO on the rapid pace of LLMs by 00001746
That's not what I meant, so no offense taken.
MammothJust4541 t1_jdz0nxj wrote
Reply to [D] Simple Questions Thread by AutoModerator
If I wanted to make a system that takes an image and transforms it into the style of another image what sort of ML model would I want to use?
sdmat t1_jdz0h51 wrote
Reply to comment by starfries in [D] FOMO on the rapid pace of LLMs by 00001746
I mean no personal offense, but it's strange to see someone generalizing from an extreme outlier in a machine learning sub.
starfries t1_jdz0f1p wrote
Reply to comment by ReasonablyBadass in [D] FOMO on the rapid pace of LLMs by 00001746
Huh, I could have sworn it was a lot older.
ReasonablyBadass t1_jdyzvv1 wrote
Reply to comment by starfries in [D] FOMO on the rapid pace of LLMs by 00001746
That dude was a researcher before he wrote that though
[deleted] t1_jdyzun9 wrote
Reply to comment by UndeadMusterd in [R] Created a Discord server with LLaMA 13B by ortegaalfredo
[removed]
[deleted] t1_jdyzu0l wrote
Reply to comment by H0PEN1K in [R] Created a Discord server with LLaMA 13B by ortegaalfredo
[removed]
WildlifePhysics t1_jdyzon6 wrote
Reply to [D] FOMO on the rapid pace of LLMs by 00001746
There are a lot of ways to get involved in advancing research beyond generating your own foundational models. Do what others don't.
sdmat t1_jdyzb37 wrote
Reply to comment by cheddacheese148 in [D] FOMO on the rapid pace of LLMs by 00001746
Not a movie, but it's definitely SF:
> "Far Centaurus" (1944) by A. E. van Vogt: This classic science fiction story tells the tale of a group of colonists who embark on a centuries-long voyage to the distant star system Centaurus. Upon arrival, they discover that Earth has developed faster-than-light travel during their journey, and a thriving human civilization already exists in Centaurus. > > "The Songs of Distant Earth" (1986) by Arthur C. Clarke: The novel features the crew of a slower-than-light colony ship, Magellan, who arrive at their destination planet Thalassa, only to discover that faster-than-light ships have already colonized other planets in the meantime. The story explores the consequences of different levels of technology and adaptation for the human settlers. > > "Tau Zero" (1970) by Poul Anderson: In this novel, a group of colonists aboard the starship Leonora Christine set out to explore a distant star system. During their journey, they encounter a series of technical malfunctions that cause their ship to accelerate uncontrollably. As a result, they experience time dilation, and the rest of the universe rapidly advances around them. They must navigate their own obsolescence and search for a new home as other expeditions overtake them.
Being able to find anything with a few vague words about content is one of my favourite GPT4 capabilities!
starfries t1_jdyz458 wrote
Reply to comment by sdmat in [D] FOMO on the rapid pace of LLMs by 00001746
No, I mean you don't need anything special or to follow a conventional path.
Flag_Red t1_jdyz0vk wrote
Reply to comment by ZestyData in [P] 🎉 Announcing Auto-Analyst: An open-source AI tool for data analytics! 🎉 by aadityaubhat
Alternatively, we could encourage people to make interesting stuff and share it with the community.
ReasonablyBadass t1_jdyyyo3 wrote
Reply to [P] two copies of gpt-3.5 (one playing as the oracle, and another as the guesser) performs poorly on the game of 20 Questions (68/1823). by evanthebouncy
Interesting. I can't look at the raw data right now: was memory the problem? Did it ignore clues it got? Or was it more conceptual, did it not figure out properties of objects it asked for?
Could you quickly list the terms it did get right?
eamonious t1_jdyyv6n wrote
Reply to [P] two copies of gpt-3.5 (one playing as the oracle, and another as the guesser) performs poorly on the game of 20 Questions (68/1823). by evanthebouncy
Another thing I’ve found that GPT still really struggles with is ranking a list of words accurately by difficulty. I tried many different prompt styles and couldn’t get a result that even approached satisfactory.
sdmat t1_jdyyqwe wrote
Reply to comment by starfries in [D] FOMO on the rapid pace of LLMs by 00001746
Does it? How many other fanfic writer -> well known researcher trajectories come to mind?
Username2upTo20chars t1_jdyypnk wrote
Reply to [D] Will prompting the LLM to review it's own answer be any helpful to reduce chances of hallucinations? I tested couple of tricky questions and it seems it might work. by tamilupk
Okay, I get the response:
>Response Draft: An animal that starts with the first letter of the capital of France is an Eel.Draft Reflection: The answer draft is correct. Paris is the capital of France and the first letter is "P." However, the answer could be more concise by simply stating "Eel" as the answer.
Actual Response: The answer is Eel. The capital of France is Paris and the first letter is "P". Therefore, an animal that starts with the first letter of the capital of France is an Eel.
Given the following instructions:
>For this interaction, we are going to use the following structure.
>
>User (me): [I will ask a topic or question]
>
>You will provide an Response Draft: [Brief or simplified answer to the topic or question]
>
>Then you will undergo Draft Reflection: [You will provide a critique or review of the answer draft, highlighting the limitations, inaccuracies or areas that need improvement, correction or expansion, while providing guidance on how to address these issues in the revised response. Important, think step by step, while reviewing or criticizing the hypothetical response.]
>
>Then you will provide an Actual Response: [The natural and contextually appropriate answer to the topic or question, as generated by the advanced language model, which incorporates the suggestions and improvements from the draft reflection for a more comprehensive and accurate response. This also can include step-by-step reasoning.]You will in general act like the worlds best experts of the respective domain or field of the question.Do you understand?
-_-
Okay, this Update to the instructions gives a correct response:
Actual Response: [The corrected response draft given the draft reflection. This also can include step-by-step reasoning.]You will in general act like the worlds best experts of the respective domain or field of the question.Do you understand?"
Spziokles t1_jdyyies wrote
Reply to comment by WarAndGeese in [D] FOMO on the rapid pace of LLMs by 00001746
Came to say this. Compare yourself with someone who enters the field in two years, or two months. Heck, we all witness what difference even two weeks currently make.
Will they find a job? Will they have a hard time? If your worries are true, then it should be even harder for them. Which means, you have an advantage having this head start.
I guess we can also safely expect the demand for all skill levels around ML to increase, the more it impacts our societies and economies. Yes, we might need less people for a single task, but the amount of tasks will grow more. I do not worry for either new and old folks.
Vrulth t1_jdyy8uy wrote
Reply to [D] Is French the most widely used language in ML circles after English? If not, what are some useful (natural) languages in the field of machine learning? by Subject_Ad_9680
I'm a frenchman and I laugh, thanks !
abnormal_human t1_jdyxteq wrote
Model weights are not currently considered to be copyrightable, and there is no DMCA/RIAA/MPAA machinery providing additional consequences for "pirating" them. At least for the moment, it's not a big risk to use LLaMA/Alpaca models for commercial use so long as you have not made an agreement with Facebook not to do it.
The OpenAI policy is about competing models, and comes from the TOS of using their API. Stanford agreed to that TOS, then released the text (which is again, not copyrightable). Random people downloading that data set aren't party to that agreement or bound by it.
I'm sure that Google, Facebook, Amazon, Netflix, etc will be cautious here, but for a random smaller org, this is a risk/benefit tradeoff, not an absolute.
A person who takes a torrented LLaMA and finetunes it using the Stanford data set didn't necessarily engage in any contracts prohibiting that.
The original leaker of LLaMA weights broke the rules. That's about it. Tsk tsk.
gembancud t1_jdyxpsg wrote
On a more constructive note, using agents or plugins to interface to an actual repository of information is probably the most efficient way to interleave information retrieve and fact checking.
radi-cho t1_jdz40zp wrote
Reply to [P] two copies of gpt-3.5 (one playing as the oracle, and another as the guesser) performs poorly on the game of 20 Questions (68/1823). by evanthebouncy
Last week I released a CLI that can do this at scale: https://github.com/radi-cho/datasetGPT. Will use personal funds to generate somewhat big task oriented dataset later today with gpt-3.5 or gpt-4. Will open source it along a way for people to contribute their own datasets so we can collect bigger ones. Would be helpful both for analysis of how LLMs work and for fine tuning downstream models (Alpaca-like).