WhenMarkZuckerbergisn't wakesurfingwearing a tuxedoand a pukashellnecklaceathisLakeTahoemansioncrushingCoors' yellowbelliesandwavingtheAmericanflag, heclocksintoworkwith a sunburntobattleGoogleandOpenAIforartificialintelligencesupremacy.
What's interestingisthatwecantake a lookattheactualcodeusedtotrainthismodel, whichisonly 300 linesofPythonandPyTorch, alongwith a librarycalledFairescaletodistributetrainingacrossmultipleGPUs.
It's a relativelysimpledecoder-onlytransformer, asopposedtothemixture-of-expertsapproachusedin a lotofotherbigmodels, likeitsbiggestopensourcerival, Mixtral.
Mostimportantlythough, themodelweightsareopen, andthat's a hugewinfordevelopersbuildingAI-poweredapps.
Nowyoudon't havetopay a bunchofmoneytousetheGPT-4 API, andinsteadcanself-hostyourownmodelandpay a cloudprovider a bunchofmoneytorentsomeGPUs.
Thebigmodelwouldnotbecheaptoself-host.
I usedLLAMAtodownloaditanduseitlocally, buttheweightsweigh 230 gigabytes, andevenwithanRTX 4090, I wasn't abletoridethisLLAMA.
I alsohaditdosomecreativewritingandpoetry, andoverallit's prettygood, justnotthebest I'veeverseen.
Ifwetake a minutetoreflectthough, what's crazyisthatwehavemultipledifferentcompaniesthathavetrainedmassivemodelswithmassivecomputers, andthey'reallplateauingatthesamelevelofcapability.
OpenAIwasthefirsttomake a hugeleapfromGPT-3 toGPT-4, butsincethen, it's onlybeensmallincrementalgains.
Lastyear, SamAltmanpracticallybeggedthegovernmenttoregulateAItoprotecthumanity, but a yearlater, westillhaven't seentheapocalypticSkynethumanextinctioneventthattheypromisedus.
ThishasbeenTheCodeReport, thanksforwatching, and I willseeyouinthenextone.
WhenMarkZuckerbergisn't wakesurfingwearing a tuxedoand a pukashellnecklaceathisLakeTahoemansioncrushingCoors' yellowbelliesandwavingtheAmericanflag, heclocksintoworkwith a sunburntobattleGoogleandOpenAIforartificialintelligencesupremacy.
Subtitles and vocabulary
Click the word to look it upClick the word to find further inforamtion about it