Some very promising arena results, congrats
@xai
&
@elonmusk
!
Open invitation for a grok-2-large release podcast with us, people would love to hear the technical details 👀
🆕 the one and only
@jxnlco
on:
• Why Pydantic is -still- All You Need
• Why most AI frameworks aren't VC backable
• Why you'd be happier being bootstrapped
• Why you should NOT hire ML Engineers
• the definitive origin of Instructor
(and bonus
🆕 pod: The Yolo Researcher Metagame with
@YiTayML
!
OpenAI (ca. GPT4): ~600 people
Google Gemini: ~950 coauthors
@RekaAILabs
: 20 people
@sama
once speculated on the qualities of "10,000x AI researchers", and more recently
@_jasonwei
described the "Yolo
🆕 The Busy Person's Intro to Finetuning & Open Source AI
with
@winglian
of Axolotl
Covering the SF AI meetup with
@NousResearch
,
@Teknium1
, and all the required knowledge to get started navigating Open Source AI and finetuning them.
(special cohost
Prediction: "In the next 2 years, every LLM will become LMMs" (
@chipro
's term for Large Multimodal Models)
🆕: How to train your own Large Multimodal Model
with
@HugoLaurencon
and
@LeoTronchon
of
@Huggingface
!
We follow how HuggingFace's M4 team
🆕 pod with
@ThomasScialom
of
@AIatMeta
!
Llama 2, 3 & 4: Synthetic Data, RLHF, Agents on the path to Open Source AGI
shoutouts:
- Why
@ylecun
's Galactica Instruct would have solved
@giffmana
's Citations Generator
- Beyond Chinchilla-Optimal: 100x
🆕 The Accidental AI Canvas
The full story of how
@steveruizok
took
@gdb
's GPT4V Demo to its logical conclusion, and accidentally invented the future of multimodal prompting with
@tldraw
.
Now on YouTube + wherever fine podcasts are sold.
🆕 The Winds of AI Winter
The vibes have shifted.
On
@leopoldasch
vs Sequoia, Goldman Sachs,
@benedictevans
,
@cpaik
's End of Software and why the AI Engineer rises above it all.
The future is here, but it is not evenly distributed.
🆕 Building AI for The People
Never has so much been shipped for so many by so few.
@jeremyphoward
is back on the pod! sharing the founding journey of
@AnswerAI
, predicting the
@OpenAI
governance crisis, hiring 1000x researchers and developers like
🆕 pod: How to train a Million Context LLM!
@ylecun
says we should publish, or perish. We asked
@markatgradient
to spill ALL the beans on how his team extended Llama-3 to 1M+ context with ~perfect
@GregKamradt
NIAH evals!
Also covering:
- A Brief
Are GPU clouds intentionally losing money on last month's Mixtral inference price war?
Probably. Depends on batching and aggressiveness of accounting. As
@NaveenGRao
puts it - a lot of people are about to learn the meaning of "TCO"...
🆕 The Four Wars of the AI Stack
Our Dec 2023 recap also includes a framework for looking at the key business battlegrounds of all of 2023:
In Data:
with OpenAI announcing a partnership with Axel Springer (see also its deal with the AP and its Data
🆕 pod: with
@BEBischof
of
@_hex_tech
!
On putting AI Magic into Notebooks, and how RAG is actually a recommendation systems problem.
Also an
@AAAzzam
hot take: LLMOps is more like an "iron mine" than a "gold rush"!
"You shouldn't even be hiring MLEs"
@jxnlco
on how companies are unnecessarily hiring MLEs when they should be looking to the under-utilized but motivated engineers that have a desire to pursue AI Engineering.
🆕How to Hire AI Engineers
a rare guest post (and bonus pod) from
@james_elicit
and
@_adamwiggins_
! Covering:
- Defining the Hiring Process
- Defensive AI Engineering as a chaotic medium
- Tech Choices for Defensive AI Engineering
- How do you Interview for Defensive AI
🆕 pod on Transformer Math 101!
with
@QuentinAnthon15
of
@AiEleuther
everything you ever wanted to know about the nitty gritty of training and making the gpu go brrrrrrrrrrrr
The one thing more valuable than GPUs is knowing how to use them (aka make them go brrrr 🔥)
@QuentinAnthon15
of
@AiEleuther
, co-author of "Transformers Math 101", joined
@swyx
and I on
@latentspacepod
🎙️
Take your notepads out before pressing play! We
🆕 Our first NSFW AI episode, with
@itsjessehi
!
Covering:
- the Waifu vs Fan Platform market
- 2-5x'ing the income of OnlyFans creators
- Building and Finetuning with DSPy
- Model graded evals
- Guarding against prompt injections
- Memory for
OpenAI's RLHF evaluation metrics show us that RLHF can improve some metrics, but it doesn't always mean people are going to be happy with the changes - with
@natolambert
🆕 The AI-First Graphics Editor
We talked with
@suhail
on:
- his journey from Mixpanel to Mighty
- How a conversation with
@model_mechanic
inspired
@playground_ai
- raising $40m to reimagine the graphics editor
- how to self-teach AI with
🆕 NeurIPS 2023 Recap — Top Startups!
part 2 of our NeurIPS coverage is the biggest ensemble of founders and company leaders we’ve ever featured:
•
@jefrankle
- Chief Scientist, MosaicML
•
@lqiao
- CEO, Fireworks AI
•
@amanrsanger
- CEO, Anysphere
🆕 pod: RLHF 201
Our deep dive into Reinforcement Learning with Human Feedback, with
@natolambert
of
@allen_ai
+
@interconnectsai
!
Covering:
- History of RL and its applications
- Intellectual history of RLHF dating to Aristotle (!)
- RLHF for
🆕 Why you should write your own LLM benchmarks
w/ Nicholas Carlini of
@GoogleDeepMind
Covering his greatest hits:
- How I Use AI
- My benchmark for large language models
- Extracting Training Data from Large Language Models (RIP
@openai
logprobs)
Full episode below!
Congrats
@xai
and
@elonmusk
on the open source release 🫡
Grok-1 cheat sheet 📝
- MoE architecture for 314B total params
- 8 experts with 2 active
- Apache 2.0 license
- Trained on JAX and Rust
- Finished training October 2023
- Base model, no task-specific fine-tune
- No
Thanks to
@eugeneyan
for starting the LS Discord
#paperclub
and
@youngphlo_
our fancy new logo!
If you want a welcoming, unrecorded space to dive in to LLM papers every week, we're still open to new members on the LS Discord/Paper club. join in!
🆕 State of the Art: Training >70B LLMs
We are excited to deep dive into
@imbue_ai
's incredible new releases with
@joshalbrecht
(CTO of Imbue)
AND with the best return GUEST COHOST possible for the job:
@jefrankle
(Chief AI Scientist of
@DbrxMosaicAI
Early this year, we trained a 70B model optimized for reasoning and coding. This model roughly matches LLAMA 3 70B despite being trained on 7x less data.
Today, we’re releasing a toolkit to help others do the same, including:
• 11 sanitized and extended NLP reasoning
One Year of Latent Space 🔭
Friday is
@latentspacepod
birthday, so I wrote out some thoughts on our first year of hosting it with
@swyx
. Grateful for all the support, and a special s/o to
@OfficialLoganK
for helping us kick it off as the first guest 🚀
🆕 pod alert: Special
@aidotengineer
Preview Edition!
We're so excited to flip the tables with
@RazRazcle
and help him launch his new podcast, High Agency. A full interview on the 1 year anniversary of the Rise of the AI Engineer, and special preview of
Q: How can 1-bit LLMs match 16-bit LLaMAs? 2-3x faster inference and 20-40x more energy efficient and slightly better evals!!
Great work by RJ for the
@latentspacepod
paper club - a deep dive into the 1-bit LLMs paper! We had a LOT of fun going thru the details this week and
🆕 pod: Heralds of the AI Content Flippening
with Youssef Rizk of
@wondercraft_ai
!
AI content is good and getting better. When will it get SO good that our primary media diet is AI generated?
🇸🇬 Singapore AI Engineers!
Our second meetup is taking place at
@StripeDev
HQ on 6 Feb:
Join
@ivanleomk
and
@swyx
!
For those in other countries, feel free to organize your own and tell us on discord!
🆕 Cloud Intelligence at the speed of 5000 tok/s
with
@vipulved
(CEO) and Ce Zhang (CTO) of
@TogetherCompute
!
A deep dive into the full stack research lab + GPU cloud behind RedPajama, Mamba, StripedHyena, and why openness is core to Together's
🆕 We Are Running Out of Low-Background Tokens
Was 2022 the last year we had "clean" data uncontaminated by ChatGPT?
(h/t
@jeffreyhuber
for the concept!)
also:
-
@martin_casado
's Voice Cloning Red Alert
- that
@pika_labs
video
- incredible LCM demos
🆕 Weekend pod: Jan+Feb recap + 1 Yr of Latent Space!
Our 2023 recap pod was very well received, so here's the next in this new series of research recaps, with takes and guests from:
- OpenAI Sora
- Gemini Pro 1.5 (shoutout
@simonw
's post)
- Groq infra
🆕 pod: Building the Foundation Model Ops Platform
with
@RazRazcle
of Humanloop!
Covering: Three Stages of Evals & Human Feedback, the market for ops platforms, and diving into the most underrated LLM research.
Also asking: Did GPT-4 get dumber?
New weekend pod alert!
this is a long 'un, something in it for everyone, whether you're into
@Grok
or
@GroqInc
, or all about Corporate AI Leadership or Personal AI therapists.
Enjoy and please send feedback!
LLM Fine tuning is here!
San Francisco’s top AI engineers came together to see what’s possible with fine-turning and only 4 hours of hacking.
Here’s an exclusive what we saw at the “Anything But Wrappers” hackathon (🧵):
🆕 pod: The Winds of AI Winter!
The vibes have shifted...
@fanahova
and
@swyx
celebrate 1m downloads and recap the last 3 months in AI!
Discussing the Frontier Labs vibe shift between Claude 3.5, Llama 3.1, Apple Intelligence, and the expansion of
🆕Benchmarks 201 with
@clefourrier
of the
@huggingface
Open LLM Leaderboard!
Why Chat Arenas are **OVERRATED**:
- they bias sycophancy/assertiveness over factuality
- annotators are not representative of humanity
- models are argmaxxing public arena
thanks Leila! yes the philosophy is to just have an ongoing conversation with listeners and honestly a dumping ground for all the stuff that we’re not yet sure about but found noteworthy anyway
the quality of the
@eugeneyan
@yikesawjeez
@YoungPhlo_
@eugenechh
et al paper club
@dwarkesh_sp
I love the *idea* of a subscriber-only discord, but I haven’t yet seen it done well. In general people who post a lot in a large discords usually aren’t very interesting.
The only exception I know is the
@latentspacepod
discord where swyx/Alessio post a lot of good stuff.
Huge congrats to Perplexity! first and most well deserved AI raise of 2024.
check out our NeurIPS convo with Aravind here:
any bets on where Perplexity will end the year?
Excited to announce we've raised 73.6M$ at 520M$ valuation, led by IVP, along with our seed and Series A lead investors NEA, Elad Gil, Nat Friedman. Honored to have NVIDIA, Jeff Bezos,
@tobi
, Databricks,
@naval
,
@rauchg
,
@balajis
partner in this round, among others.
🆕 Truly Serverless Infra for AI Engineers
In 2021
@bernhardsson
wrote a "Software Infra 2.0 wishlist", but quickly decided to take matter into his own hands by building
@modal_labs
.
On today's
@latentspacepod
we explain how VM-based clouds don't work
At Stitch Fix,
@jxnlco
created a $50MM money printer. They built a recommendation engine using embeddings of the comments, reviews, and clothing. This let them recommend items that work in the same outfit.
This is entirely speculative but... Tomorrow's LS paper club with
@picocreator
is going to be extremely lit!
come learn about the state of the art in Speculative Decoding!
What if there are more modalities than meets the eye?
What could be built out of that?
What problems could be solved?
-
@Suhail
with a potential moonshot company and going beyond current modalities with physics
I listen to a ton of AI podcasts and by far the best are where the host is a technical expert rather than a journalist or other generalist. Deeper, smarter, more in tune with their audience
The gold standards are
@latentspacepod
,
@CogRev_Podcast
and
@NoPriorsPod
IMO
🆕 We got that semianalysis guy
@dylan522p
on
@latentspacepod
!
Dishing on:
- what it's like to live the GPU Rich/Poor divide
- Why GPT4 cost $500m but training costs are irrelevant
- non-Nvidia silicon (AMD/TPU/Cerebras/Matx)
- Can we ever move semis
"I want AI to know me really, really well" - Igor
His wild approach? Recording every single conversation to build a massive personal audio dataset
1 year of his life's audio is less than 1 terabyte of data
Cost v Reliability v Performance: Choose Three!
🆕 pod:
w/
@jungofthewon
and
@stuhlmueller
How do you improve prompt engineering and ai agents? The emerging paradigm is sometimes called "Flow engineering", but
@elicitorg
has reinvented the Notebook, in
🆕 pod on Transformer Math 101!
with
@QuentinAnthon15
of
@AiEleuther
everything you ever wanted to know about the nitty gritty of training and making the gpu go brrrrrrrrrrrr
Despite Cursor’s recent insane growth, the current version is just 0.1% of what we have in store.
We’re a small, very strong team and are looking for fantastic SWEs and designers to help shape the future of software development.
Read more here -
And we somehow forgot to post ICLR pod part 2!
First: a good ol' longform convo with
@gneubig
on his work, ft. OpenDevin which has now formed into a company! and our first-ever returning-guest guest-host,
@amanrsanger
!
Announcement:
@rbren_dev
,
@xingyaow_
, and I have formed a company!
Our name is All Hands AI 🙌
And our mission is to build the world’s best AI software development agents, for everyone, in the open.
Here’s why I think this mission is important 🧵
Midjourney web generations (right) feel extremely different than Discord ones (left). It's extremely biased towards making cartoonish / drawing images. Haven't tweaked any setting.
Same prompt, ran 2-3 times each; all web ones were drawings, ~20% of Discord ones were.
"a
signal boosting this important critique of the FMTI BY
@natolambert
,
@BlancheMinerva
,
@aviskowron
et al:
in retrospect any transparency rating that rates gpt4 the third most transparent FM in the world, above
@EMostaque
’s StableLM, is kiiinda sus
The Foundation Model Transparency Index by
@StanfordCRFM
purports to be an assessment of how transparent popular AI models are. Unfortunately its analysis is quite flawed in ways that minimize its usefulness and encourage gamification