apparently open AI is on the verge of releasing their brand new model and now we have a ton of additional information about it including the fact that open AI may actually have shown this new technology to the federal government so we're going to go over all of that right now all right so we have two main articles from the information that were posted over the last couple days first open AI shows strawberry so we've talked a lot about strawberry talked a lot about qar I'll link those videos in the description below and I will drop
a link to the full article in the description below open AI shows strawberry AI to the feds and uses it to develop Orion and Orion is new to me and apparently it is their next Frontier Model now here's the thing if you were ever doubting Jimmy apples the infamous open AI leaker look at this tweet from all the way back at November 24th 2023 let's conquer the cosmos mood curious Jimmy so with that did he actually know and then as of today yeah been waiting since last year for this mood patients Jimmy and he's referring
to the information article about Orion so maybe he knew about it so let's read a little bit about the article so this article by Stephanie palazolo starts with in case you were wondering why Sam Alman cryptically posted a picture of strawberries earlier this month the answer almost certainly has to do with strawberry a mysterious technical breakthrough that could help open AI models complete complex tasks such as math problems that conversational AI has traditionally struggled with so just a quick catchup qar strawberry it's all referencing the same thing slower thinking models models that can actually plan
ahead think through problems multi-step reasoning and actually do much better at math and logic and reasoning these models don't just return the first token they predict they actually go off and think about it and actually do long-term thinking in mid July Reuters reported on the existence of strawberry and this morning we published this piece with even more details and we will go over that in a moment here is the most interesting bit this summer his team Sam alman's team demonstrated the technology to American National Security officials said a person with direct knowledge of those meetings
which haven't been previously reported by demonstrating an unreleased technology to government officials open AI could be setting a new standard for AI developers especially as advanced AI increasingly becomes a national security concern the demonstration could be part of open ai's push to be more transparent with policy makers who could cause the company problems if they feel threatened by its technology and scanning down a bit and maybe also take a shot at meta platforms for releasing openweight AI that China and everyone else can access so you all know how I feel about this open source is
the way to go to think that a single private company is going to be able to protect secrets to protect the weights against an adversary like China infiltrating and getting that IP it's just so unlikely that they will be able to protect it forever right here meta CEO Mark Zuckerberg says it's inevitable that China will get it one way or the other and I agree all it takes is one slip up in security which we're humans We Make Mistakes One slip up and then China will have the weights anyways and it's not just about the
weights you also need the compute to power it you need a lot of other things as well I did a full review of Mark Zuckerberg's thoughts on open source and it overlaps with my own quite a bit so I'll drop that video in the description below let's keep reading about strawberry and why it matters to Orion and again strawberry qar it is different from Orion Oran is reportedly their next Frontier Model that is going to leave everybody else in the dust one of the most important applications of strawberry is to generate high quality training data
for Orion open ai's next Flagship large language model that's in development now why is that important much of the training data on the internet has already been used there's almost none left that is public and easily accessible now it's all behind pay walls and authentication that's why open AI has been doing deals with different Publications and Reddit and all these other companies and that's why Twitter's data set is so important to x. a but Sam Alman a while ago hinted that maybe more and more training data isn't actually necessary there's actually two other approaches that
are possible one that they can actually do a lot more with existing data and the second path which it looks like they're talking about here is actually generating in synthetic data using one model for another model now there's been a lot of doubt that that is a sustainable model because if you're creating data with one model to be used on another model it's basically derivative data a lot of people think that large language models aren't actually creating new knowledge it's simply outputting what it already knows so if it's only able to Output what it already
knows how could it really train a model to be that much better using strawberry could help Orion reduce the number of hallucinations or errors it produces resarch ERS tell me that's because AI models learn from their training data so the more correct examples of complex reasoning they see the better so what could be going on here is they have a set of data and Orion and qstar are doing long-term thinking and multi-step planning and basically thinking through the data to make sure it's accurate and then producing new data that is highly accurate but there's also
a push within open AI to simplify and Shrink strawberry through a process called distillation so it can be used in a chat-based product before Orion is released so that tells me that the technology behind strawberry might just be so slow it's unusable in a consumer setting and that would make sense and I'll actually touch more on this in a bit but the gist is strawberry and qar takes time that's a feature not a bug it actually takes a lot of time it thinks through it just like humans would when you ask someone something they don't
immediately spit out the first thing that they think of or at least most people don't instead they take time to think through it they might take notes especially hard questions reasoning logic math these are things things that we take our time with we write down thoughts and we iterate on the result here the information admits we're not sure what a strawberry based product might look like but we can make an educated guess one obvious idea of what strawberry could be an actual product is incorporating strawberries improved reasoning capabilities into Chachi BT and that's the obvious
one however though these answers would likely be more accurate they also might be slower that means that strawberry might be ill suited for applications where users expect immediate responses like open AI search GPT search engine but ideal for less time sensitive use cases like fixing non-critical coding errors in GitHub now another good use case for this is Agents if I have an agent working for me 24 hours a day I really don't need a response immediately I can give it these superpowers powered by qstar powered by strawberry allow it to go do its thing and
then bring me back the best possible response you could imagine a Noto distant future where chat GPT users are able to toggle strawberry on and off depending on how sensitive they request are so that's the first article from the information about Orion that's the first time I've heard of it so hopefully I'm bringing it to you for the first time too now chubby from X who is an amazing follow if you're not already following him what chubby touches on here is that strawberry is slower because rather than just responding immediately with whatever it thinks is
the next correct token it actually takes its time and does what is called system to thinking rather than just system one thinking now let's read more about strawberry so this is another article from from the information open AI raises to launch strawberry reasoning AI to boost chatbot business so it is reported that open AI is looking to raise even more Capital which is insane they've already raised so much but you know what it costs a lot of money to build these models at least for now its researchers are trying to launch a new artificial intelligence
product they believe can reason through tough problems much better than its existing AI here they just touch briefly on qar we've already discussed it quite a bit when given additional time to think the strawberry model can also answer customers questions about more subjective topics such as product marketing strategies to demonstrate strawberry's prowess with language related tasks open AI employees have shown their co-workers how strawberry can for example solve New York Times connections a complex word puzzle so if it were given the puzzle and just asked to Output immediately what it thought system one thinking then
it's not able to solve it nearly as effectively as if it had time to iterate and use maybe tree of thought or Chain of Thought or any these other more advanced capabilities where the model can actually look ahead plan and test things out come back and test other things out as they see things are working or not here the information talks about the open AI business and its sales of llms to corporations and of chat GPT subscriptions have roughly tripled to 283 million in monthly Revenue compared to a year ago that's insane though its monthly
losses are likely higher than that as reported by the information and I've already explained that they are probably still still losing money but they're not going to go bankrupt anytime soon the company is privately valued at 86 billion but open AI prospects rest in part on the eventual launch of a new flagship llm it is currently developing Cod named Orion so we already talked about that now why is this next model so important well open source pretty much caught up with GPT 40 llama 370b llama 345b is a fraction of the price you can run
it locally you can fine-tune it and it's nearly as good or as good as GPT 4 o for the majority of use cases and not only that we have the clad models we have grock 2 we have more grock models coming we have perplexity for search so the competition is heating up really quickly and open AI really needs to launch something incredible to jump ahead because intelligence is being driven down to a cost of nothing here they talk about the same things that we touched on in the last article but let me just read it
again open AI is also using the bigger version of strawberry to generate data for training Orion and set a person with knowledge of the situation that kind of AI generated data is known as synthetic something we've touched on a lot on this channel it means that strawberry could help open AI overcome limitations on obtaining enough highquality data to train new models from Real World data such as text or images pulled from the internet and here apparently open aai is going to be launching agents soon so strawberry could Aid upcoming open aai agents that person said
using strawberry to generate higher quality training data could help open AI reduce the number of Errors its model generate otherwise known as hallucinations now one of the biggest blockers one of the biggest hurdles for artificial intelligence in general to be adopted within Enterprise settings and more critical settings is the fact that it still hallucinates there's a bunch of things that you could do to reduce hallucinations whether that's improving your prompts having multiple agents talk to each other and verify kind of like an agentic system and pulling in information from the internet to verify and of
course doing your own verification but at the end of the day if we want large language models to be fully autonomous and to really run at the scale that we believe it can it really needs to reduce hallucinations to nearly zero so the CEO of agent startup minion Ai and former Chief Architect of GitHub co-pilot says imagine a model without hallucinations a model where you ask it a logic puzzle and it's right on the first try the reason why the model is able to do that is because there is less ambiguity in the training data
so it's guessing less so from Sam Alman at an event back in may we feel like we have enough data for this next model so they are close and I really cannot wait to see the next thing that they bring out I have said this a lot it's really hard for me to imagine that a private AI company is going to have some completely unique research or unique Tech that allows them to make a 50 100% Improvement on what's already out there maybe 10% maybe 20% but a large leap from what is already on the
market today is really hard for me to imagine just because of the way that the scientific community operates all of these papers get published and unless open AI research scientists are the only ones in the world who actually thought of a new idea and it didn't leak out anywhere it's just very unlikely that they're going to have some completely new technology here they talk about why solving math problems could be so lucrative for open AI as a business AI that solves tough math problems could be a potentially lucrative application given that existing AI isn't great
at math heavy Fields such as Aerospace and Structural Engineering now that isn't necessarily true Google's deepmind team actually got I think silver at the math olympiads so it's definitely possible today already that AI can be extremely good at math and right here they reference what I just mentioned Google deep mine said its AI would beat most human participants in the international mathematical Olympiad another major rival anthropic said its latest llm could write more complicated software code than its prior llms could and answer questions about charts and graphs than to Improvement in its reasoning capability so
again competition is there and specifically about anthropic the cloud model is seemingly the favored large language model for coding use cases everybody seems to be using cursor the IDE that is AI native in conjunction with the claw models it continues with to improve models reasoning some startups have been using a cheap hack that involves breaking down a problem into smaller steps though the workarounds are slow and expensive it's kind of weird to call it a cheap hack these are these are discovered prompt techniques and Frameworks to wrap your large language model to make them more
effective and I don't see that as a cheap hack at all and yeah they tend to be slower but what if you just power it with Gro grq and then all of a sudden you have all the benefits of these additional layers of technique to get the best response but you're also getting hundreds of tokens per second so it ends with what ilas saw strawberry has its roots in research it was started Years Ago by ilas suser then open ai's Chief scientist he recently left to start a competing AI lab before he left open AI
researchers Jacob and Simon sedor built on SS's work by developing a new math solving model qar alarming some researchers focused on AI safety and basically the entire super alignment team has left since then so what did Ilia really see and last just to add a pinch of fun to this from New York Magazine the AI guys are driving themselves mad and this is in reference to Strawberry man and qar and the strawberry model Jimmy apples and all of this conjecture and speculation and you know what it's all in good fun it's the same thing that
basically every other industry does when apple is about to release a new phone everybody who follows Apple starts talking about what it could possibly be and talking about the leaks and you know what we're so excited about what's to come why not add a little speculation and try to figure out what might come so this article is pretty funny it just talks about how I rule the world Mo was responded to by Sam Alman and all of his hypey tweets which is great and here's the awesome part they actually included one of my videos about
qar which is flattering it's so cool they also go on to talk about Lily Ashwood who everybody's trying to figure out if she's AI or not we talked about that on the live stream last week so there's definitely a lot going on a lot of fun and as soon as the new open AI model comes out you know I'm going to cover it in depth and I'll give you all the information so be sure to subscribe to my channel and as a reminder I have an awesome newsletter so you can stay up to dat in
the latest AI Trends and news Matthew burman.com check it out if you enjoyed this video please consider giving a like And subscribe and I'll see you in the next one