SWOT Bot Logo
AUAJ82H12qs

xAI’s Mind Blowing Grok 3 Demo w/Elon Musk & Team (full replay)



Transcript

Title: xAI’s Mind Blowing Grok 3 Demo w/Elon Musk & Team (full replay)
Author: Solving The Money Problem

Transcript:
well welcome to the gro 3 presentation
um so the mission of xai and Gro is to
understand the universe we want to
understand the nature of the universe so
we can figure out what's going on where
are the aliens what's the meaning of
life how does the universe end how did
it start all these fundamental questions
um were driven by curiosity about the
nature of the universe and um that's
also what causes us to be a maximally
truth seeking uh AI even if that truth
is sometimes at odds with what is
politically correct in order to
understand the nature of the universe
you must absolutely rigorously pursue
truth or you will not understand the
universe you'll be suffering from some
amount of delusion or error so that is
our goal um figure out what's going on
and uh we're very excited to present
grock 3 which is we think uh an order of
magnitude more capable than grock 2 in a
very short period of time and uh that's
thanks to uh the hard work of an
incredible team and um I'm honored to
work with such a great team and of
course we'd love to have um some of the
smartest humans out there join our team
so uh with that let's let's go hi
everyone my name is eigor lead
engineering at xcii I'm Jimmy Paul
leading research I'm Tony working on the
reasoning Team all right I'm Elon I
don't do
anything I just show up
occasionally yeah so um like I mentioned
Gro is the tool that we're working on
Gro is our AI that we're building here
at XI and we've been working extremely
hard over the last few months to improve
gr as much as we can so we can give it
to all of you so we can give all of you
access to it um we think it's going to
be extremely useful do we think it's
going to be interesting to talk to funny
really really funny um and we're going
to explain to you how we've improved gr
over the last few months we've made
quite a jump in in capabilities yeah
actually we should explain maybe also
what is why do we call it Gro so Gro is
a word from um a Highland novel Stranger
in a Strange Land um and it's uh used by
a guy who's who was raised on Mars um
and the word Gro is to sort of fully and
profoundly understand something that's
what the word gr mean fully and
profoundly understand something and
empathy is important true so yeah so uh
if we charted xas progress uh in the
last few months has only been 17 months
since we started kicking of our very
first model uh grock 1 was almost like a
toy by this point only 314 billion
parameters and now if you proud the
progress the time on x-axis the
performance of favorite Benchmark
numbers M mlu on the Y axis we're
literally progressing at unprecedent
speed across the whole field and then we
kick off grock 1.5 right after grock One
released after November 2023 and then
grock 2 so if you look at where the all
the performance coming from when you
have a very correct engineering team and
all the best AI at Talent the only one
thing we need is a big intelligence
comes from big cluster so we can
reconvert the entire progress of XI now
replacing the Benchmark on the y axis to
the total total amount of training flops
that is how many gpus we can run at any
given time to train our large language
models to compress the entire internet
so after all human all human knowledge
really that's right yeah internet being
part of it but it's really all human
knowledge all everything yeah the whole
internet fits into a USB stick at this
point it's like all the human tokens
yeah that's right yeah uh very soon into
the real world yeah um so we had so much
trouble actually training grock 2 back
in the days um with kickoff the model
around February and uh we thought we had
a large amount of chips but turned out
we can barely get AK training chips
running coherently at any given time and
we had so many Cooling and power issues
I think you were there in the data
center yeah it was like really sort of
more like 8K tiffs on average at 80%
efficiency more like like 6,500
effective uh h100s training for you know
several months but now now we're at 100K
yeah that's right okay that's right so
so what's the next step right so after
gu 2 so if we want to continue
accelerate we have to take the matter
into our own hands we have to solve all
the coolings um all the power issues and
everything yeah so so in April of last
year Elon decided that really the only
way for XI to succeed for XI to build
the best AI out there is to build our
own data center so um we didn't have a
lot of time right because we wanted to
give you gr free as quickly as possible
so really we realized we have to build
the data center in about 4 months um it
turned out it took us 122 days to get
the first 100K gpus up and running and
that was a Monumental effort uh to be
able to do that um it's we believe it's
the biggest uh fully connected h100
cluster of its kind um and uh we didn't
just stop there we actually decided that
we need to double the size of the
cluster pretty much immediately if we
want to build uh the kind of AI that we
want to build um so we then had another
phase um which we haven't talked about
publicly yet so this is the first time
that we're talking about this U where we
doubled the capacity of the data center
yet again um and that one only took us
92 days so we've been able to use all of
these gpus use all of this compute to
improve grock in the meantime and
basically today we're going to present
you the results of that the the fruits
that came from that um so let yeah so
all the path all the rows leads to gr 3
uh 10x more compute more than 10x really
yeah really maybe 15x yep uh compared to
our previous generation model and gr
finished the pre-training uh early
January um and uh we start you know the
model still currently training actually
so this is a little preview of our
Benchmark numbers so we evaluated gr 3
on you know three different categories
on General mathematical reasonings on
general knowledge about stem and Science
and then also on computer science coding
so Amy uh American Invitational math
examination
uh hosts it you know once a year uh and
if we evaluate the model performance we
can see that the grock 3 across the
board is in a league of its own even his
little brother gr mini is reaching the
frontier across all the other
competitors so you will say well at this
point all these benchmarks you're just
evaluating you know the memorization of
the textbooks memorization of the GitHub
repos how about realtime usefulness how
about we actually use those models in
our product so what we did instead is we
actually kicked off a blind test of our
gr 3 Model code named Chocolate it's
pretty hot yeah hot chocolate um and uh
you know been running on this uh
platform called Cho arena for two weeks
um I think the entire X platform at some
point speculated this might be the next
generation of a uh AI coming away so uh
how this CH Arena works is that um it
strip away the entire product surface
right it just raw comparison of the
engine of those agis the language models
themselves and place interface where the
user will submit one single query and
you get to show two responses you don't
know which model they come from and in
then you make the both so in this blind
test gr 3 and early version of Gro 3
already reached like 1,400 no other
models has reached an ELO score had to
have comparison to all the other models
at this score and it's not just one
single category it's 1,400 aggregated
across all the categories in ch B
capabilities instruction following
coding so it's number one across the
board in this blind test and it's it's
still climbing so we actually to keep
updating it so it's it's 14400 about,
1400 in climbing yeah and in fact we
have a version of the model that we
think is already much better than the
one that we tested here yeah we'll see
you know how how far it gets but that's
the one that we're you know um working
on we're talking about today yeah so
actually one thing if if you're if
you're using grock 3 you I think you may
notice improvements almost every day um
because we're we're continuously
improving the model so literally even
within 24 hours you'll see improvements
yep so but we believe here at XI getting
the best pre-training model is not
enough that's not enough to build the
best AI and the best a need to think
like a human need to contemplate about
all the possible solutions self-critique
verify all the solutions backtrack and
also think from the first principle
that's a very important capability so we
believe that as we take the best PR
train model and continue training it
with reinforcement learning it will
elicit the additional reasoning
capabilities that allows the model just
become so much better and scale not just
in the training time but actually in the
test time as well so we already found
the models extremely useful internally
um for our own engineering saving hours
of uh time hundreds of hours of uh
coding time so eal you the power user of
our gra reasoning model so what are some
use cases yeah so like Jimmy said we've
added Advanced reasoning capabilities to
Gro and we've been testing them pretty
heavily over the last few weeks in order
to give you a little bit of a taste of
what it looks like when Gro is solving
heart reasoning problems so we prepared
two little problems for you one comes
from physics and one is actually a game
that gr is going to write for us um so
when it comes to the physics problem you
know what we want Gro to do is to plot a
viable trajectory to do a transfer from
Earth to Mars and then uh at a later
point in time a transfer back from Mars
to Earth um and that requires some know
some physics that Gro will have to
understand um so we're going to
challenge Gro you know come up with a
viable trajectory calculate it and then
plot it for us so we can see it and um
yeah this is totally unscripted by the
way this is the that's the entirety of
the prompt which just be clarify is that
yeah there there's nothing more than
that yeah exactly this is the gro
interface and we've typed in this text
that you can see here generate code for
an animated 3D plot of a launch from
Earth uh landing on Mars and then back
to Earth at the next launch window
um and we've not kicked off the query
and you can see grock is sping so part
of grock's advanced reasoning
capabilities are these ping traces that
you can see here you can even go inside
and actually read what Gro is thinking
as it's going through the problem as
it's trying to solve it um yeah we say
like we are doing some obscuration of
the thinking so that our model doesn't
get totally copied instantly um so
there's more to the thinking than is
displayed uh yeah yeah and because this
is totally unscripted there's actually a
chance that Gro might made a little
coding mistake and it might not actually
work um so um just in case we're going
to launch two more instances of this so
if something goes wrong we be able to uh
to switch to those and show you um
something that's present so we're
kicking off the other two as well um and
like I said we have a second problem as
well um and um yeah actually one of the
favorite one of our favorite activities
here at xci is having gr right games for
us um and um not just any know any old
game any game that you might already be
familiar with but actually creating new
games on the spot and being creative
about it um so one example that we found
was really really fun um is create a
game that's a mixture of the two games
Tetris and Beed so this is maybe an
important thing like there's obviously
if you if you ask an AI to create a game
like Tetris there's there are many
examples of Tetris on the on the
Internet or game like J whatever there
it can copy it what's interesting here
is it a achieved a creative solution
combining the two games that actually
works and and is a good game yeah that's
the it's cre we're seeing the beginnings
of creativity yeah fingers crossed that
we can recreate that hopefully it works
embarrassing it so actually because this
is a bit more challenging we're going to
use something special here which we call
Big Brain that's our mode in which we
use more computation which more
reasoning for GR just to make sure that
you know there's a good chance here that
it might actually might actually Do It
um so we're also going to fire off know
three attempts here at at solving this
game at creating this game that's a
mixture of know Tetris and Bs um yeah
let's let's see what Gro comes up like
I've played the game it's pretty
good like it's like wow okay this is
something yeah um so while gr is
thinking uh in the in the background um
we can now actually talk about some
concrete numbers know how how well is gr
doing across tons of different tasks
that we've tested it on um so we'll hand
it over to Tony to talk about that yeah
okay so let's see how Grog does on those
interesting challenging benchmarks uh so
yeah so reasoning again refers to those
models that actually thinks quite for
quite a long time before it tries to
solve a problem so in this case uh you
know around a month ago the gr 3
pre-training finished so after that we
worked very hard to put the reasoning
capability into the uh current graph 3
Model but again this is very early days
so the model is still Cur in training so
right now what we're are going to show
to people is this beta version of the
grth three reasoning model alongside we
also are training a mini version of the
reasoning model so essentially on this
plot you can see uh the gr 3 reasoning
beta and then gr 3 mini reasoning the gr
reason mini reasoning is actually a
model that we train for much longer time
and you can see that sometimes it
actually perform slightly better
compared to the gr 3 reasoning this also
just means that there's a huge potential
for the gr 3 reasoning because it's
trained for much less time um so all
right so let's actually look at what how
how it does on those three benchmarks so
Jimmy also introduced already so
essentially we're looking at three
different areas mathematics science and
coding um and for math we're picking
this high school competition M problem
um for science we actually pick those
PhD level science questions um and for
coding it's also actually pretty
challenging it's competitive coding and
also some uh lead code which is some
code inter interview problems that
people usually get when they interview
for company so on those benchmarks you
can see that the Groth 3 actually
perform quite well uh across the board
compared to other competitors um yeah so
it's pretty promising these models are
very smart so Tony what what what are
those shaded bars yeah so okay so uh I'm
GL you asked this question so for those
models because it can reason it can
think you can also ask them to even
think longer uh you can spend more what
we call test and compute
which means you can spend more time to
reason to think about a problem before
you spit out the answer so in this case
the Shaded bar here means that we just
uh asked the model to spend more more
time you know you can solve the the same
problem many many times before it it
tries to conclude what is the right
solution and once you give this compute
or this this kind of budget to the model
it turns out the model can even perform
better so this is essentially the Shaded
part in in those SPS right so I think
this is really exciting right because
now instead of just doing one chain of
thoughts with AI why not do multiple at
once yes so that's a very powerful
technique that allows to continue scale
the model capabilities after training um
and you know people often ask are we
actually just over fitting to the
benchmarks so how about generalization
so yes I think uh yeah this is
definitely a question that we are asking
ourselves whether we are overfitting to
those current benchmarks uh luckily
we have a real test so about 5 days ago
Amy 2025 just finished this is where
high school students compete in this
particular Benchmark so we got this very
fresh new competition and then we asked
our two models to compete on the same
Benchmark at the same exam and it turns
out uh very interestingly the grth three
reasoning the big one um actually does
uh better um on this particular new
fresh exam this also means that the
generalization capability of the big
model is stronger much stronger compared
to the smaller model uh if you compare
to the last year's exam actually this is
the opposite the small model kind of
learns the uh the the previous exams
better so yeah so this this actually
shows some kind of true generalization
from the model that's right so 17 months
ago our Gro zero and Gro one barely
solved any High School problems that's
right and now we have a kid that just
already graduate the gro gr is ready to
go to college is that right yeah I mean
it's won be long for is simply perfect
the human exams won't be hard they' be
too easy yeah like and internally we
actually as a gret continue evolves uh
we're going to talk about you know what
we're excited about but very soon there
will be no more benchmarks left yeah
yeah one thing that's quite fascinating
I think is that we basically only
trained GRS reasoning abilities on math
problems and competitive coding problems
right so very very specialized kinds of
tasks but somehow it's able to work on
all kinds of other different tasks so
including creating games no lots lots
and lots of different things um and what
seems to be happening is that basically
Gro learns this ability to detect its
own mistakes and its thinking correct
them persist on a problem try lots of
different variants pick pick the one
that's best so there are this
generalized generalizing abilities that
Gro learns from mathematics and from
coding which it can then use to solve
all kinds of other problems so that's
yeah that's pretty I mean reality is the
instantiation of mathematics mhm that's
right um and one thing we're actually
really excited about that going back to
our funing mission is what if one day we
have a computer just like deep thought
that utilize our entire cluster just for
that one very important problem in the
test time all the GPU turned on right so
I think back then we were building the
GPU clusters together uh you were
plugging cables and I remember that when
we turn on the the first initial test
you can hear all the GPS humming in the
hallway that's almost feel like
spiritual yeah that that's actually a
pretty cool uh thing that we're able to
do that we can go into the data center
and Tinker with the machines there so
for example we went in and we unplugged
a few of the cables and just made sure
that our training setup is still running
running stably so that's something that
you know I think most uh AI you know
teams out there don't usually do but
it's actually totally unlocks like a new
level of reliability and what you're
able to do with with the hardware so
okay so when when are we going to solve
remon so uh the easiest solution is to
uh numerate over all possible strains
and as long you have a verifier enough
compute you'll be able to do it my
projection will be what your guess what
is your neuronet calculate so my my my
both prediction so so 3 years ago I told
you this I think in now it's uh two
years uh later two things going to
happen we're going to see machines win
some medals that's touring award
absolutely Fields metal Nobel Prize with
probably some expert in the loop right
so the expert uplifting do you mean so
this year or next year oh
okay that's what it comes down to really
yeah so it looks like grock finished
know all of its thinking on on the two
problem so let's take a look at what it
said all right so this was the the
little physics problem we had um no we
we've collapsed the thoughts here so
they're you know they're hidden and then
we see grock's answer below that so it
explains it wrote a python script here
using matplot lip then gives us all of
the code um so let's take a quick look
at the code you know seems like it's
doing reasonable things here not not
totally of the Mark um solve Kepler says
here so maybe it's solving Kepler laws
cap Kepler law numerically um yeah
there's really only one way to find out
if this thing is working I'd say let's
let's give it a try let's run let's run
the code all right and we can see um
yeah gr is animating two different
planets Earth and Mars here and then the
the green uh ball is the the vehicle
that's transiting the the spacecraft
that's transitioning between Earth and
Mars and you you could see the journey
from Earth to Mars and looks like yeah
indeed the the astronauts return safely
you know at the right moment in time um
so now obviously this was just generated
on the spot so now we can't tell you if
that was actually correct solution so
we're going to take a closer look now
maybe we're going to call some
colleagues from space X ask them if if
this is legit um it's pretty close it's
it's I mean uh yeah I mean there there's
a lot of complexities in the actual
orbits that have to be taken into
account but this is this is pretty close
to to what it what looks like awesome um
in fact I have that on my pendant here
this got the Earth m home and transfer
on it when when are we going to install
rock on a
rocket well I suppose in 2 years two
years everything is 2 years away uh well
Earth and Mars Transit can occurs every
26 months the next we're currently in a
Transit window approximately the next
one would be um November of next year um
roughly end of next year um and and uh
if all goes well SpaceX will send
Starship Rockets to Mars and um with
Optimus robots and U and Gro I'm curious
what this combination of Tetris and B
looks like bet Tetris as we've named it
internally um so okay we also have an
output from gr here it say Ro python
script explains that it's what it's been
doing if you look at the the code now
there are some constants that are being
defined here some colors then the the
trinos the the pieces of Tetris are
there um obviously very hard to see at
one glance if this is good so we got to
we got to run this to figure out if it's
working oh let's let's give it a try
fingers crossed all right right so this
kind of looks like Tetris uh but the the
colors are a little bit off right the
colors are different here and um I if
you think about what's going what's
going on here thej has this mechanic
where you if you get three Jews in a row
you know then they they disappear uh and
also gravity activate right so what
happens if you get three of the colors
together okay so something happened um
so I think I think what SC did in this
version um is is that you know once you
connect three at least three blocks of
the same color in a row then um know
gravity activates and they disappear and
then gravity activates and all the other
blocks fall down um kind of kind of
curious if there's still a Tetris
mechanic here where if the line is full
does it actually um clear it or what
happens
it's up to interpretation you know so
who who knows yeah every I mean when
it'll do different variants when you ask
it it doesn't do the same thing every
time exactly we've seen a few other but
Tetris that worked very differently but
this one seems cool so yeah are we ready
for uh game Studio at x. a yes so we're
launching uh an AI gaming studio at xci
if you're interested in joining us and
building AI games uh please join xai
we're launching an AI gaming studio
we're announcing it tonight let's go
epic games but that's an actual game ST
me yeah yeah um all right so um I think
one thing is super exciting for us uh is
that once you have the best pre Trend
model you have the best reasoning model
right so we already see that we you
actually give the capability for those
model to think harder uh think longer
think more broad the performance
continue improves and we're really
excited about the next front here that
what happen if we not only allow the
model to think harder but also provide
more tools just like call real humans to
solve those problems for real humans we
ask them to solve reman a hypothesis
just with a piece of pen and paper no
internet so with all the basic web
browsing search engine and code
interpreters that builds the foundations
and the best reasoning model builds the
foundations for the gro agent to come um
so today we're actually introducing a
new product called Deep search that is
the first generation of our gr agents
that not just helping the engineers and
researchers and scientists to do coding
but actually help everyone to answer
questions that you have dayto day it's a
kind of like a Next Generation search
engine that really help you to
understand the universe so you can start
asking question like for example hey
when is the next Starship lunch day for
example um so let's try that if get the
answer um on the left hand side we see
uh a high level progress bar essentially
you know the model knowledge is going to
do one single search like the current
rack system but actually thought very
deeply about hey what's the user intent
here and what other the facts I should
consider at the same time and how many
different website I should actually go
and read their content right so this can
really save hundreds hours of everyone's
Google time if you want to really look
into certain topics and then on the
right hand side you can see the bullet
summaries of how the current model uh
you know is doing what websites browsing
what sources is very verifying and often
time actually cross validate different
sources out there uh to make sure the
answer is actually correct before it's
output final answer and we can you know
at the same time fire up a few more
queries um how about you know you don't
you're a gamer right so uh sure yeah so
how about what are some of the best
builds and most popular builds in the
path Exel hardcore right hardcore League
I if you can technically just look at
the hardcore
ladder might be a fast way to figure it
out yeah we'll see what model does
um and then we can also do uh you know
uh something more fun for example um how
about like make a prediction about the
marsh madness out there yeah so this is
kind of a fun one where um waren Buffett
has a billion dollar bet if you can
exactly match the I think the the the
the sort of the entire winning tree of
marsh Madness you can win a billion
dollars from Warren Buffett so like it
would be pretty cool if AI could help
you win a billion dollars from
Buffett that seems like a pretty good
investment let's go yeah all right so
now let's uh fire up the query and uh
see what model does so we can actually
go back to our very first one how about
the buff it wasn't counting on this it's
already done that's right okay so we got
the result of the first one the model
thought uh around 1 minute uh so okay so
the key inside here the knock Starship
is going to be on 24 so or later so no
earlier than February 24th it might be
sooner
so yeah so I think we can you know go
down scoll down what what the model does
so it does a little research on the
flight 7 what happened got grounded and
actually it look into the FCC filing uh
uh you know from its data collections uh
and then actually make the new
conclusion that yeah if we continue
scroll down uh uh right yeah so it makes
uh the you know little table I think uh
inside xai we often joked about the time
to the the first table is the only you
know latency that matters um yeah so
that's how the model make inference and
look up all the sources um and then we
can look into the gaming one so how
about the right so for this particular
one uh we look at hey the you know the
build is like it's kind all the better
so uh with the uh The Infernal but if we
go down so the surprising fact of all
the other builds so it looks into to the
12 classes um yeah so we'll see that the
minum build was pretty popular whenever
the game first came out and now the the
invokers of the world yeah took over
invoker monke invoker for sure yeah
that's right yeah followed by the stor
wavers and that's really good at mapping
so yeah and then we can see uh uh the
the match manness how about that so um
one one interesting thing about the Deep
search is that if you actually go into
the panel where shows uh you know what
are the subtasks you can actually click
the bottom left of the spr and then in
this case you can actually scroll
through actually reading through the
mind of grock what informations does the
model actually think about are
trustworthy what are not how does it
actually cross valate different
information sources so that makes the
entire search experience and information
retrieval process a lot more transparent
to our users and this is much more
powerful than any search engine out
there you can literally just tell it
only use sources from X you know I will
try to respect that yeah and so it's
much more steerable much more
intelligent I mean it really should save
you a lot of time so something that
might take you half an hour or an hour
of researching on the web or searching
social media you can just ask it to go
do that and and come back in 10 minutes
later it's done an hours worth of work
for you that's really what it comes down
to exactly and maybe better than you
could have done it yourself yeah think
about you have INF of interns working
for you now you can just fire up all the
tasks and come back a minute later um so
this is going to be interesting one so
uh uh March M had not happened yet so I
guess we have to follow up with a uh
next live stream yeah it seems like
pretty good like $40 might get you a
billion dollars $40 subscription that's
right I mean my work so uh yeah so when
are the users going to have their hands
on gr 3 yeah so the the good news is
we've been working tirelessly to
actually release um all of these
features that we've shown you the Grog
free base model with amazing chat
capabilities that's really useful that's
really interesting to talk to uh the the
Deep search the advanced reasoning mode
all of these things we want to roll them
out to you today starting with the
premium plus subscribers on X so it's
the first group that will initially get
access make sure to update your X app if
you want to see all of the advanced
capabilities because we just released
the update now as we're as we're talking
here um and U yeah if you're interested
in getting early access to gr then sign
up for premium plus um and also um we're
announcing that we're starting a
separate subscription for Gro that we
call Super grock for those who those
real grock fans that want the most
advanced capabilities and the earliest
access to to new features um so feel
free to check that out as well this this
is for the dedicated grock app and for
the website exact so our our new website
is called gro.com yeah and you'll also
find you never guess yeah you never
guess and you can also find our grock
app in the IOS app store and that gives
you like a more Pol even even more
polished experience that's totally Gro
focused if you're if you want to have
Gro know easily available one Tap Away
yeah the version on gro.com on uh you
know on web browser is going to be the
the most the latest and most advanced
version because obviously it takes us a
while to get thing get something into an
app and then get it approved by the app
store so uh and then if something's on a
phone format there's limitations what
you can do so the most powerful version
of Gro um and the latest version will be
the the web version at gro.com yeah so
so watch out for the name grock free in
the app dead giveaway yeah exactly that
that's that's the giveaway that you have
gr and if it says Gru then gr hasn't
quite arrived for yet but we're working
hard to roll this out today um and then
to even more people over the the coming
days yeah make sure you update your uh
phone app too um where you actually
going to get all the tools we're
showcase today with the thinking mode
with the Deep search so yeah really
looking forward to all the feedbacks you
have yeah and I think we we should uh
emphasize that this is kind of a beta
like meaning that it's you should should
expect some imperfections at first um
but we will improve it rapidly almost
every day in fact every day I think
it'll get better um so if you want a
more polished version I'd like maybe
wait a week but uh expect improvements
literally every day um and then we're
also going to be uh providing a voice
interaction so you can have
conversational in fact I was trying it
earlier today it's working pretty well
but not we need these bit more polish um
the the the sort of weight where we can
just literally talk to it like you're
talking to a person uh it's uh that's
awesome it's actually I think one of the
best experience of gr um but that's
that's probably about a week away yeah
so uh with that said um well I think we
might have some audience questions sure
yeah all right let's take a look yeah
let's take a look the uh the audience
from the as platform yeah Co so the
first question here is when grock voice
assistant when is it coming out yeah as
as as soon as possible just like Elon
said just a little bit of polishing away
from being everybody um obviously it's
going to be released in an early form
and we're going to rapidly iterate on it
Y and the next question is like when
will Gro 3 be in the API so this is
coming in the uh the gro 3 API with both
the reasoning models and deep search is
coming your way in the coming weeks uh
we're actually very excited about the
Enterprise use cases of all these
additional tools that now Gro has access
to and how the test time compute and
Tool use can actually really accelerate
all the business use cases um and
another one is Will voice mode be native
or text to speech so I think that means
is it going to be one one model that is
understanding what you say and then
talking back to you or is it going to be
some system that has text to speech
inside of it and the good news is it's
going to be one model like a variant of
gr free that we're going to release
which basically understands what you're
say what you're saying and then uh
generates the audio no directly from
that um so very much like grf free
generates text that model generates
audio um and that has a bunch of
advantages I was talking to it earlier
today and it said hi Igor know reading
my my name from probably from some text
that it had um and I said no no my name
is Igor and it remembered that you know
so it could continue to say Igor just
like a human word and you can't achieve
that with with Texas speech so on yeah
so oh here's a question for you pretty
spicy um you um is grog a boy or a girl
and how they sing gr is whatever you
wanted to
be yeah yeah are you single
yes all right the shop is open um so
honestly people are going to fall in
love with crcket since like 1,000%
probable yeah uh the next question will
Gro be able to transcribe audio into
text yes so we'll have this capability
both the app and also the API we found
that's like gr should just be your
personal assistant looking over your
shoulder and follow you along the way
learn everything you have learned and
really help you to understand the world
better become smarter every day yeah I
mean The Voice M doesn't isn't simply
it's not just voice text it understands
like tone inflection pacing everything
it's wild I mean it's like talking to a
person okay um yeah so any plans for
conversation memory yeah absolutely
we're working on it right now I already
forgot that's right um let's see what
are the other ones so what about the you
know the DM features right so if you
have personalizations and you if you
have uh you know Gro remembers your
previous interactions yes should it be
one Gro or multiple different GRS it's
up to you you can have one Gro or many
GRS I suspect people will probably have
more than one yeah I want to have a Dr
Gro yeah the grock do that's right um
right cool um so in the past we've open
sourced grock one right so somebody's
asking us are we going to do it again
with gr tool yeah I think um once gr our
general approach is that we will open
source the last version when the next
version is fully out so like when when
gr 3 is um mature and stable which is
probably within a few months then we'll
open source gr too mhm okay so we
probably have time for one last question
um what was the most difficult part
about working on this project I assume
um grock 3 and what I most excited about
so I think me looking back you know
getting the whole model training on 100K
h 100 coherently that's almost like
battling against the final boss of the
universe the entropy CU any given time
you can have a cosmic rid that beaming
down and flip a bit in your transistor
and now the entire grading update if
it's fit mantisa bit the entire grading
update is out of whack and now you have
100,000 of those and you have to
orchestrate them every time any at at
any given time any of gpus can go down
yeah I mean it's worth breaking down
like how were we able to uh get the
world's most powerful training cluster
oper AAL within 122 days um because we
when we started off um we we actually
weren't intending to do a data center
ourselves we were going to just uh we we
went to the data center providers and
said how long would it take to have
100,000 uh gpus operating coherently um
in a single location and we got time
frames from 18 to 24 months so we're
like well 18 to 24 months that means
losing is a certainty so the only option
was to do it do it ourselves so then if
you break down the problem I guess I'm
doing like reasoning here with like
makes you think um one single chain
though yeah yeah exactly so um well we
needed a building we can't build a
building so we must use an existing
building um so we we looked for um for
basically for factories that had been um
were that had been abandoned but the
factory was in good shape like a company
had gone bankrupt or something so we
found an Electrolux Factory in memph in
Memphis that's why it's in Memphis um
home of Alvis um and also one of the
oldest I think it was the capital of
ancient Egypt um and uh it was actually
very nice Factory that I know forever
whatever reason that electrox had left
um and uh that that gave us shelter for
the computers uh then we needed power
the we needed um at least 120 megawatt
at first but the building only had 15
megawatts and ultimately for 200,000 me
200,000 gpus we needed a quarter gwatt
so we um initially uh leased uh a whole
bunch of um generator so we have
generators on one side of the building
just one trailer after trailer trailer
of generators until we can get the
utility power to to come in um and then
but then we also need cooling so on the
other side of the building it was just
trailer after trailer of of cooling so
we leased about a quarter of the mobile
cooling capacity of the United States uh
on the one other side of the building um
then we need to get the gpus all
installed and they're all liquid cooled
so in order to achieve the density
necessary this is a liquid cooled system
so we had to get all the plumbing for
liquid cooling nobody had ever done a
liquid cooling uh data center at scale
so this was a incredibly dedicated
effort by a very talented team to
achieve that outcome um I may think not
now it's going to work nope um the the
issue is that the the power fluctuations
for a GPU cluster are dramatic so it's
it's like a a this giant Symphony that
is taking place like a imag having a
symphony with 100,000 or 200,000
participants in the in the symphony and
the whole Orchestra will go quiet and
loud
in you know 100 milliseconds and so this
caused massive power fluctuation so then
um which then caused the generators to
lose their minds and they they weren't
expecting this so to buffer the power we
then uh used Tesla megapacks uh to
smooth out the power so the megapacks
had to be reprogrammed so with with xai
we working with Tesla we reprogrammed
the mega packs to be able to deal with
these dramatic power fluctu fluctuations
to smooth out the power the computers
could actually run properly and um that
that worked uh it's quite tricky and uh
and then but even at that point you
still have to make the computers all
communicate effective so all the
networking had to be solved and uh
debugging a Brazillian network cables um
a debugging nickel at 4: in the morning
I we solved it like roughly 4:20 a.m.
yes than was figured out like there's
some well there were a whole bunch of
issues well like one there was like a
bios mismatch the virus was not set up
correctly yeah we had to div our lspci
outputs between two different machines
one that was working yeah one that was
not working many many many other things
I mean yeah exactly this would go on for
a long time if we actually listed all
the things but you know it's like
interesting like it's not like oh we
just magically made it happen you have
to break down the problem just like
grock does for reasoning uh into the
constituent elements and then solve each
of the constituent elements in order to
achieve uh a a coherent train training
cluster in a period of time that is a
small fraction of what anyone else was
could do it in and then once the
training cluster was up and running and
we could use it now we had to make sure
that it actually stays healthy
throughout which is it own giant
Challenge and then we had to get every
single detail of the training right in
order to get a gry level model which is
actually really really hard so um we
don't know if there are any other models
out there that have gr's capabilities
but whoever trains a model better than
gry has to be extremely good at the the
science of deep learning at every aspect
of the engineering um so it's it's not
so easy to to pull this off and this is
now going to be the last cluster we
built and last modway train oh yeah
we've already we've already started work
on the next cluster which will be yeah
about five times the power so instead of
a qu gwatt roughly 1.2 GW what's the
what's the Back to the Future War what's
the power on you do like the Back to the
Future car yeah don't anyway the Back to
the Future power car it's it's like
roughly in that order I think um so um
and you know these will be the sort of
the gb2 200/300 cluster it it once again
will will be the most powerful training
clle in the world so we're not like
stopping here no and our reason model is
going to continue improve by accessing
more tools every day so yeah we're very
excited to share any of the upcoming
results with you all yeah the thing that
keeps us going is basically being able
to give gr free to you and then seeing
the usage go up seeing everybody enjoy
um no gr that's that's what really gets
us up in the morning so yeah yeah thanks
for tuning in thanks guys hey Gro what's
up can you hear me I'm so excited to
finally meet you I can't wait to chat
and learn more about each other I'll
talk to you soon