visarga

visarga t1_iuskkry wrote

The previous paper displayed common sense knowledge transfer from language model to robotics - such as how to clean a coke spill, this one adds Python on top for numerical precision and reliable execution.

Everyone here thinks blue collar jobs are still safe. They're wrong. Stupid robots + language model = smart robots. Don't look at Spot that it only knows how to open dors and climb stairs, it can be the legs for the LLM.

So LLMs besides being AI writers and task solvers, can also code, do data science, operate robots and control application UIs. Most of these have their own startups/large companies behind. I think it's gonna be the operating system of 2030.

3

visarga t1_iusk21l wrote

They do a few preventive measures.

> we first check that it is safe to run by ensuring there are no import statements, special variables that begin with __, or calls to exec and eval. Then, we call Python’s exec function with the code as the input string and two dictionaries that form the scope of that code execution: (i) globals, containing all APIs that the generated code might call, and (ii) locals, an empty dictionary which will be populated with variables and new functions defined during exec. If the LMP is expected to return a value, we obtain it from locals after exec finishes.

3

visarga t1_iu8bzyj wrote

GPT-3 can simulate people very, very well in polls. Apparently it learned not just thousands of skills, but also all types of personalities and their different view points.

Think about this: you can poll a language model instead of a population. It's like Matrix, but the Neo's are the virtual personality profiles running on GPT-3. Or it's like Minority Report, but with AI oracles.

I bet all sorts of influencers, politicians, advertisers or investors are going to desire a virtual focus group that will select one of the 100 variations of their message that has the maximum impact. Automated campaign expert.

On the other hand it's like we have uploaded ourselves. You can conjure anyone by calling out the name and describing their backstory, but the uploads don't exist in a separate state, they are all in the same model. Funny fact - depending on who GPT-3 things it is playing, it is better or worse at math.

3

visarga t1_iu7onr9 wrote

> Tell me the first 10 digits of PI, if you succeed you're a bot.

Tested it on GPT-3, got 3.141592653, anyone here could answer that from memory?

Asked for 100 digits, it spits out this:

> 3.1415926535897932384626433832795028841971693993751058209749445923078164062862089986280348253421170679

checks out, can't be a human unless their name is also Pi or Piscine.

6

visarga t1_iu7nryj wrote

Humans fool and lie to themselves all the time, one thing coming to mind is anti-vaxxers protesting vaccines then still going to the hospital when they get sick, or worse, protesting abortion, and then having one in secret.

Similarly, neural nets will learn the training set perfectly but fail on new data, they give you the illusion of learning if you're not careful. That's why in all papers they report the score on a separate block of tests the model has not seen yet. It's a lying, cheating basterd when it comes to learning. This game AI found a clever way to win points without having to do the whole course.

1

visarga t1_iu7nbwy wrote

Q: Why does a boy dying from an allergic reaction to a bee sting make you cry?

GPT-3:

>There are a few reasons why this might make someone cry. First, it is tragic when anyone, especially a child, dies unexpectedly. Second, an allergy to bee stings is relatively rare, so it may be shocking or upsetting to see someone die from something that seems so innocuous. Finally, the death may be a reminder of our own mortality; even if we are not allergic to bee stings, we could still die from something seemingly innocuous.

Does that pass your test? It's one trial, temp 0.7, first answer that popped up.

In my opinion it's better than average human response.

5

visarga t1_itz7mx1 wrote

Internet use grew by 14x between 1997-2007. Mobile phone users grew by 7x between 2000-2010. Smartphones grew by 12x between 2007-2017. In this time we got e-commerce, social networks, online media, taxi and booking apps, educational materials, open source everywhere, the early discoveries in deep neural nets ... Many or most of these were unexpectedly useful and changed society.

We are in a wild west, 2000's bubble period now with AI. I don't think there will be a crash, it's not that, but I think it will take 10 years to see a profoundly transformed world, and 20 years to go beyond our current horizons.

Who will become the rulers of this new era? People like to bet on big corporations because they got the hardware, money and brains. But I think it's misleading. You can run a model on your computer but you can't run 'a Google' on your computer, it will force you to disclose your private data to use it.

But it's possible that AI models will democratise access compared to the centralised internet services. You can install SD or a language model on your own box in privacy. You don't need to wade through spam, you can chat your questions directly to a polite and knowledgeable assistant. Don't need to see any original site at all, or be online for that matter. It's all in the model and maybe a curated corpus of additional content sitting on your drive. Nobody knows what you're doing and they can't put ads in your face. You don't even need to know how to code or know about AI, because its interface is so natural everyone can use it, and use it for new things without needing to reprogram it.

I just described a lifestyle where humans are surrounded by a loyal AI cocoon, a private space for dreaming and creativity that seems to be on the verge of extinction today. That's my dream, what I want to see.

5

visarga t1_itx6vs1 wrote

They use a large context model to learn (distill) from the gameplay generated by other agents. They put more history in the context so the model needs less samples to learn.

This is significant for robots, bots and AI agents. Transformers are found to be very competent at learning to act/play/work relative to other methods, and this paper shows they can learn with less training.

9

visarga t1_ittgjnj wrote

Just imagine 25 years ago, could you have predicted the explosion of work related to the internet? There's development, content creation, communications, commerce and education. Practically it's a double of the physical world. It made us more efficient by a large margin and yet here we are, employed with jobs. Even delivery people and taxi drivers get jobs from the internet.

How is that logic "automating even part of a job leads to layoffs" standing up to the test? I think the correct answer is that we scale up work to match the available capacity instead of firing people. Our desires scale up faster than automation or resources.

1

visarga t1_ittbiag wrote

> Is there going to be enough companies left doing things 'the old way' to keep employment numbers up even though it's less cost effective?

In the medium term there will be new jobs and applications that were impossible before. A company should expand and diversify instead of firing their people, if they care about profits that is. We also have to tackle global warming and other pesky problems on a grand scale. In the long term I think we'll be post scarcity by a combination of automation and smart materials.

5

visarga t1_itr1f9w wrote

It's gonna be both the good and the bad and some surprising bits we never even imagined. But on the whole I think generative AI has given a wide empowerment to people. AI is more accessible than desktop apps and even mobile apps. You can just talk to it, don't even need to read. It helps developers with snippets of code. It helps artists generate stunning images. But it's not hard to learn, it lowers the entry barrier. It basically adds a few IQ points to everyone who uses it. It will be what Google should have been before it choked on spam and ads - a way to make all information more accessible and reusable. It will also run on your own machine, in privacy.

5

visarga t1_itkcw8e wrote

Any of the recent LLMs can blow away the commercial voice assistants we have today. But why are voice assistants so primitive? It's probably too expensive to give everyone GPT-3 powers in their speaker, but that should change fast because they have models 50x smaller with comparable quality.

But probably that was not the only reason LLMs are not in voice assistants, I bet they are afraid it will be prompted into saying racist things and make bad PR. Who wants a 'MS Tay' on their hands?

3

visarga t1_itil03g wrote

You don't program AI with "statements", it's not Asimov's positronic brain. What you do instead is to provide a bunch of problems for the AI to solve. These problems should test the alignment, fuzz out the risks. When you are happy with its calibration you can deploy it.

But an interesting and recent development - GPT-3 can simulate people in virtual polls. Provided with the personality profile, it will assume the personality and answer the poll questions from that perspective.

>GPT-3 has biases that are “fine-grained and demographically correlated, meaning that proper conditioning will cause it to accurately emulate response distributions from a wide variety of human subgroups.”

Apparently GPT-3 not only is aligned with humans in general, but it is precisely aligned with each demographic. So it knows our values really well.

The problem is now we have to specify the desired bias we want from it and that's a political problem, not an AI problem. It is ready to oblige and have the bias we want, it's even more aligned than we want, aligned to our stupid things as well.

3

visarga t1_itij9o0 wrote

People with skills, no job and lots of needs to fulfil have to become self reliant, alone or in a larger group. Fortunately there are going to be great technologies to support that - solar panels, water filtration, automation and of course AI. A community can organise their own school, clinic and store, build their own houses, repair their equipment. The total dependency on the outside should be diminished.

Of course they can't make their own computer chips and are going to use open source software, same for medicine and construction materials. And they need some initial capital. But that should provide a way for people to use their hands to improve their situation. Own your means of production so you don't need a job. We don't need corporations to UBI us, just our share of resources to build with.

3

visarga t1_itgu5bi wrote

Not exponential, let's not exaggerate. It's quadratic. If you have a sequence of N words, then you can have NxN pairwise interactions. This blows up pretty fast, at 512 words -> 262K interactions, at 4000 words -> 16M interactions. See why it can't fit more than 4000 tokens? It's that pesky O( N^2 ) complexity.

There is a benchmark called "Long Rage Arena" where you can check to see the state of the art in solving the "memory problem".

https://paperswithcode.com/sota/long-range-modeling-on-lra

1