a few weeks ago I finally came around to buy a few Nvidia stocks because hey I also want to get something out of the AI boom these stocks have been dropping ever since why oh right I don't believe in God yeah so it doesn't look good for AI at the moment as investors are panicking and stocks are dropping but in this video I want to make a case that this bubble which is currently bursting is not that of AI per se it's that of the specific type of AI called large language models I'm sure AI
enthusiasm will resume once people get it into their head that there's more to Ai and my stocks will recover the best known problem with large language models is what's become known as hallucinations they sometimes confidently ramble along and produce nonsense you think we all learned this lesson in 2022 but then there was the lawyer who used chat GPT to cook up a defense and ended up citing cases that simply didn't exist whoops large language models have become better at avoiding some obvious pitfalls for example if you ask Chad GPT for book recommendations it'll Now list
books that actually exist it's still often refers to made up papers and reports though and mid Journey now for the most part puts five fingers on each hand so much so that if you explicitly ask for a hand with six fingers it'll still have five fingers you can do this by tying some output closely to the training set that might make the problem appear solvable but it's not that simple because hallucinations are just one symptom of a much bigger problem which is that for a large language model correct output is in a quantifiable Sense close
to wrong output a very illustrative example comes from Colin Fraser who's been using modified versions of kinder Garden riddles Like the Wolf goat and cabbage problem in this riddle the farmer has to get all three in a boat across the river but the boat will only carry one item in addition to the farmer left unattended the wolf will eat the goat and the Goat will eat the cabbage the solution is that the farmer has to take one of the items back on a trip if you ask a large language model this question but leave out
the information that the wolf the goat and the Goat the Cabbage then it'll still give the same answer which now makes no sense I like this example because it's obvious what's going wrong by way of word content the altered riddle is similar to The Riddle That the models have been trained on so they extrapolate from what they know and spit out an answer that's close to the answer for the original riddle but as with hallucinations these answers are close in a sense that we don't care about yes they use similar words but the content is
wrong it's like in some sense a hand with six fingers as close to one with five fingers but it's still wrong the issue is that we have a different metric for good output than the one that the models use when I say we I mean humans just in case there is some misunderstandings and these different metrics for what's good are problem you can't fix by just training a model with more and more input it's fundamentally missing information about what makes its output good for us the solution is to teach AI logic and to use symbolic
language similar to what most math software uses if you combine that with a neuron Network it's called neuros symbolic this can fix a lot of problems with large language models and some of those approaches already exist for example I mentioned already in January that Deep Mind Made remarkable progress with using AI for mathematical proofs just last month they reported that their maths AI now reached the level of a silver medalist in the maths Olympics not only does it solve the problems it also provides a proof that humans can understand well more or less the relevant
point isn't that AI can solve Math's Olympics problems because let's be honest who really cares the relevant point is that this AI can pass the problem and can formulate logically correct answers that a human can understand apply this logical rigor to verbal arguments and boom a lot of problems with large language models will disappear imagine an AI that could win every internet argument Reddit would become a ghost tone overnight what I just told you is neither new nor particularly original it's been pointed out for decades by many computer scientists including Gary Marcus and J Kerr
and others I just want to say I think they're right you can't just train large language models on more and more text and images and hope that it'll begin to understand what's going on and I think no one really expected that that said it's more difficult than lumping symbolic reasoning on top of the already existing models basically because of what wienstein called Ling IC confusion it's that no two people use a word to mean exactly the same and once you've lumped together text from billions of different people logical relations between these words become washed out
if there ever were any to begin with I mean it's not like people are all that good with logic so I'm afraid that the already trained large language models will have to be retrained ultimately the problem with large language models is that the world is not made of words at the deepest level we know of the world is mathematics if you want to build an intelligent AI you need to start with maths and with models about physical reality and then put words on top of that what this all means is that companies which have poured
a lot of money into large language models might never recover those expenses the winners will eventually be those who build an AI on logical reasoning and models of the real world like deep mind what you see here is a recent example in which they created a virtual mouse in a virtual environment that's moving with its own neural network modeled after a real Mouse brain this I think is how you'll get to really intelligent AI next up virtual cats chasing virtual mice across your computer screen during important Zoom course Deep Mind was acquired in 2014 by
Google so I haven't yet lost faith in my Google stocks the brief summary is that all these people working on AI need to think less about words and more about physics just wait until people start using AI to manage their stocks it'll be great artificial intelligence is really everywhere these days if you want to learn more about how neural networks and large language models work I recommend you check out the courses on brilliant.org all courses on brilliant have interactive visualizations and come with follow-up questions I found it to be very effective to learn something new
it really gives you feeling for what's going on and helps you build general problem solving skills they cover a large variety of topics in science computer science and maths from General scientific thinking to Dedicated courses on differential equations or large language models and they're adding new courses each month it's a fast and easy way to learn and you can do it whenever and wherever you have the time sounds good I hope it does you can try brilant yourself for free if you use my link brilliant.org Saina that way you'll get to try out everything brilliant
has to offer for full 30 days and you'll get 20% off the annual premium subscription so go and give it a try I'm sure you will won't regret it thanks for watching see you tomorrow