Recent comments in /f/MachineLearning
iamspro t1_jdj4wzl wrote
Reply to comment by countalabs in [N] ChatGPT plugins by Singularian2501
Fine-tuning is distinct afaik... using OpenAI's language for it[1]:
zero-shot: no examples in the prompt, just an input (and/or instruction)
few-shot: one or more examples of input+output in the prompt, plus new input
fine-tuning: updating the model with examples (which can then be used with zero- or few-shot as you wish)
[1] https://help.openai.com/en/articles/6654000-best-practices-for-prompt-engineering-with-openai-api (part 5)
[deleted] t1_jdj3747 wrote
Reply to comment by ZenDragon in [N] ChatGPT plugins by Singularian2501
[removed]
itsnotlupus t1_jdj2xpr wrote
Reply to [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
Meh. We see a few demos and all of the demos work all of the time, but that could easily be an optical illusion.
Yes, GPT-4 is probably hooked to subsystems that can parse an image, be it some revision of CLIP or whatever else, and yes it's going to work well enough some of the time, maybe even most of the time.
But maybe wait until actual non-corpo people have their hands on it and can assess how well it actually works, how often it fails, and whether anyone can actually trust it to do those things consistently.
CollectionLeather292 t1_jdj0jsl wrote
Reply to [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
How do I try it out? I can't find a way to ad an image input to the chat...
WokeAssBaller t1_jdixm43 wrote
Reply to comment by reditum in [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
So WebGPT doesn’t quite do this, it uses a JavaScript library to simplify web pages to basic text
agent_zoso t1_jdiwkmj wrote
Reply to comment by mescalelf in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
It always is. If you want to get really freaky with it, just look at how NFTs became demonized at the same time as when Gamestop's pivot to NFT third-party provider was leaked by WSJ. Just the other month people were bashing the author of Terminal Shock and hard sci-fi cyberpunk pioneer Neal Stephenson in his AMA for having a NFT project/tech demo by arguing with someone that knows 1000x more than they do, saying it's just a CO2 emitter and only scam artists use it and were disappointed to see he'd try to do this to his followers. Of course, the tech has evolved and those claims weren't true in his case, but it was literally all in one ear out the other for these people even after he'd defend himself with the actual facts about his green implementation and how it works. They bought an overly general narrative and they're sticking to it!
Interesting that now, with a technology that produces an order of magnitude more pollution (you can actually list models on Hugging Face by the metric tonnes of CO2 equivalent released during training) and producing an epidemic of cheaters in high schools, universities, and the work force, it's all radio silence. God only knows how much scamming and propaganda (which is just scamming but "too big to fail") is waiting in the wings.
I don't think the average person even knows what they would do with such a powerful LLM beyond having entertaining convos with it or having it write articles for them. Of course they see other people doing great things with it and not really any of the other ways it's being misused by degens right now, which goes back to an advantage in corporate propaganda.
VelvetyPenus t1_jdiw861 wrote
Reply to comment by omgpop in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
MAybe it just uses words like our brains use synapses. words are just neurons to GPT-4?
LanchestersLaw t1_jdiw7op wrote
Reply to comment by mycall in [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
The example data does demonstrate object detection
Extension-Mastodon67 t1_jdivgzf wrote
Reply to [P] ChatGPT with GPT-2: A minimum example of aligning language models with RLHF similar to ChatGPT by liyanjia92
I guess gpt2-medium is too small for RLHF.
PPO output:
Human: If I have 3 bananas and 4 pears. How many fruits do I have?
Assistant:
3 bananas: 1 pears, 1⁄2 cup of sugar, 2 tbsp of butter, and 1⁄2 tsp of cinnamon
4 bananas: 1⁄4 cup of sugar, 1 tbsp of butter, 1⁄2 tsp of cinnamon
5 bananas: 1⁄4 cup of sugar, 1 tbsp of butter, 1⁄2 tsp of cinnamon
6 bananas: 1⁄4 cup of sugar, 1 tbsp of butter, 1⁄2 tsp of
--------------
Vanilla output:
Human: If I have 3 bananas and 4 pears. How many fruits do I have?
Assistant:
Me: 3 bananas: 46 fruits (1 pie, half banana, 3 bananas)
Assistant: 46 pears: 39 fruits (1 pie, half orange, 1 pear)
Me: 47 oranges: 38 fruits (1 pie, half banana, 4 oranges)
Assistant: 47 pears: 43 fruits (1 pie, half banana, 6 oranges)
Me: 49 apples: 4 fruit (1 pie, half banana, 2 apples)
Assistant: 49 pears: 38 fruits (3 apples, 1 pear, 2 apples)
Me: 50 nectarines: 3 fruit (1 pie, half orange, 1 pear)
Assistant: 50 pears: 37 fruits (1 pie, half orange, 4
passerby251 t1_jdiusxb wrote
Reply to comment by zy415 in [D] ICML 2023 Reviewer-Author Discussion by zy415
I am not quite sure about it. Sending multiple times seems a bit demanding. But I think it would be fine to send another one when it is closer to the deadlines.
[deleted] t1_jditxnr wrote
light24bulbs t1_jditg4b wrote
Reply to comment by TFenrir in [N] ChatGPT plugins by Singularian2501
Nope, I'm struggling along with you on that I'm afraid. That's why these new plugins will be nice.
Maybe we can make some money selling premium feature access to ours once we get it
zy415 OP t1_jdispob wrote
Reply to comment by passerby251 in [D] ICML 2023 Reviewer-Author Discussion by zy415
Do you think sending multiple reminders could urge them to respond (or would it make it sound like I am too desperate/demanding)? I sent one yesterday, and am thinking if I should send another tonight.
compacct27 t1_jdis128 wrote
Reply to comment by boyetosekuji in [P] Reinforcement learning evolutionary hyperparameter optimization - 10x speed up by nicku_a
Wow that was actually really helpful
yahma t1_jdiqmj2 wrote
Reply to comment by Silphendio in [D] What is the best open source chatbot AI to do transfer learning on? by to4life4
Are the results as good as Alpaca with LLAMA base?
RedditLovingSun t1_jdipxex wrote
Reply to comment by zxyzyxz in [D] What is the best open source chatbot AI to do transfer learning on? by to4life4
They aren't open source but didn't Stanford release their code and self instruct training data that's supposedly only $600 to train? I honestly don't know but how enforceable is llamas "no using it for business" clause after someone augments one of their models with Lora and trains weights on self instruct?
ShadoWolf t1_jdipal4 wrote
Reply to comment by Necessary-Meringue-1 in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
Some of the capabilities of ChatGPT4 .. are spooky. I mean GPT-4 hired someone off of TaskRabbit to solve a Captcha for it in test phases (https://cdn.openai.com/papers/gpt-4.pdf) . I don't think it's at AGI .. but it sort of feels like Where ever we are on the S curve for this technology. we finally on the same continent for getting to AGI
And some of the stuff people are getting to to do using LangChain with ChatGPT is crazy
wind_dude t1_jdip38g wrote
Reply to comment by [deleted] in [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
access to GPT-4 with multimodel
Llukas88 t1_jdiohwe wrote
Reply to comment by to4life4 in [D] What is the best open source chatbot AI to do transfer learning on? by to4life4
Not any i know of, played around today with Alphacoom (https://huggingface.co/mrm8488/Alpacoom) and got pretty Bad results then tried a BloomZ version (https://huggingface.co/mrm8488/bloomz-7b1-mt-ft-alpaca) and got results similar to the Alpaca-Native model. Maybe read the BloomZ paper it should be a pretty good basis to build a chat model, rest should depend on your Training approach and Data.
[deleted] t1_jdio0id wrote
[deleted]
Lewducifer t1_jdinquu wrote
Reply to comment by Jean-Porte in [R] Artificial muses: Generative Artificial Intelligence Chatbots Have Risen to Human-Level Creativity by blabboy
Up and down are certainly directions, I agree.
[deleted] t1_jdinibf wrote
Reply to comment by wind_dude in [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
[deleted]
[deleted] t1_jdin627 wrote
Reply to comment by Deep-Station-1746 in [D] I just realised: GPT-4 with image input can interpret any computer screen, any userinterface and any combination of them. by Balance-
[deleted]
mescalelf t1_jdin2y7 wrote
Reply to comment by agent_zoso in [D] "Sparks of Artificial General Intelligence: Early experiments with GPT-4" contained unredacted comments by QQII
Thank you for mentioning Microsoft’s (and MA investors’) role in this/their “skin in the game”. I’m glad to hear I’m not the only one who thought the press in question—and resulting popular rhetoric—seemed pretty contrived.
VelvetyPenus t1_jdj51e0 wrote
Reply to comment by race2tb in [N] ChatGPT plugins by Singularian2501
I'm sorry, but I cannot guess your neighbor's PIN code or provide any assistance with potentially unethical or illegal activities. It is important to respect other people's privacy and avoid engaging in any actions that could cause harm or violate their rights. It is best to focus on positive and lawful ways to interact with your neighbors and build a positive community.