Aakanksha Chowdhery Profile
Aakanksha Chowdhery

@achowdhery

7,856
Followers
4,117
Following
14
Media
403
Statuses

Llama @Meta making LLMs and agents… // Previously @GoogleDeepMind :: PaLM, Gemini // @MSFTResearch , @Stanford , @Princeton // views my own and subject to change

Stanford, CA
Joined March 2011
Don't wanna be here? Send us removal request.
Pinned Tweet
@achowdhery
Aakanksha Chowdhery
2 years
#GoogleIO2022 showcases reasoning and multilingual capabilities of PaLM model. Blog: arXiv:
Tweet media one
Tweet media two
14
131
756
@achowdhery
Aakanksha Chowdhery
2 years
Presenting PaLM-E 562B, one-model generalist across robotics, language, and vision-language. It showcases multimodal chain-of-thought reasoning and the ability to reason over multiple images! And positive transfer enables it to work well on robots!!! Check out Danny's thread 👇
Tweet media one
@DannyDriess
Danny Driess
2 years
What happens when we train the largest vision-language model and add in robot experiences? The result is PaLM-E 🌴🤖, a 562-billion parameter, general-purpose, embodied visual-language generalist - across robotics, vision, and language. Website:
32
521
2K
5
54
399
@achowdhery
Aakanksha Chowdhery
2 years
1/ Reflecting back on 2022: we shared our most advanced language model PaLM - a single 540B-parameter dense language model for multiple domains & tasks, trained over two TPUv4 Pods. Research paper: Blog post:
@GoogleAI
Google AI
3 years
Introducing the 540 billion parameter Pathways Language Model. Trained on two Cloud #TPU v4 pods, it achieves state-of-the-art performance on benchmarks and shows exciting capabilities like mathematical reasoning, code writing, and even explaining jokes.
76
1K
4K
9
59
387
@achowdhery
Aakanksha Chowdhery
2 years
Excited to be presenting PaLM and Model Scaling at Stanford’s Advances in Foundation Models Class (CS324) today afternoon!
@Avanika15
Avanika Narayan
2 years
This quarter, Stanford’s Advances in Foundation Models Class (CS 324) will be partnering with the Stanford MLSys Seminar to host a special talk series on foundation models! Our first talk will be @tri_dao . Catch us *TOMORROW* at 3:30 PT:
1
28
122
4
10
343
@achowdhery
Aakanksha Chowdhery
2 years
#palm PaLM presentation at NeurIPS Google booth (incl several works on top of PaLM model this year)!
Tweet media one
4
11
339
@achowdhery
Aakanksha Chowdhery
3 years
Really excited to present the first large-scale use of Pathways system! Joint work with so many of colleagues at Google! @sharan0909 @nfiedel @JeffDean @m_isard @ada_rob @bsaeta .
@GoogleAI
Google AI
3 years
Introducing the 540 billion parameter Pathways Language Model. Trained on two Cloud #TPU v4 pods, it achieves state-of-the-art performance on benchmarks and shows exciting capabilities like mathematical reasoning, code writing, and even explaining jokes.
76
1K
4K
10
38
315
@achowdhery
Aakanksha Chowdhery
2 years
Fun to see PaLM paper in top 5 cited ML papers from 2022!
@madiator
Mahesh Sathiamoorthy
2 years
Top 5 cited papers in ML from 2022 Source:
Tweet media one
2
13
65
2
5
213
@achowdhery
Aakanksha Chowdhery
2 years
#palm But how long did it need to train? Training PaLM 62B to 1.3 trillion tokens results in significant gains as suggested by Chinchilla data scaling. However it does not bridge the gap to PaLM 540B that 5x training FLOP count. See updated results in:
5
27
188
@achowdhery
Aakanksha Chowdhery
10 months
Incredibly fun and interesting panel discussion with Percy Liang ( @percyliang ) and Angela Fan! Thank you so much to Sasha ( @srush_nlp ) for the amazing work at organizing and moderating this panel!
@srush_nlp
Sasha Rush
10 months
Be sure to make it to Hall F today to check out our exciting panel "Beyond Scaling". Thanks to everyone who provided questions.
Tweet media one
5
6
99
3
7
155
@achowdhery
Aakanksha Chowdhery
10 months
Super excited about discussing Gemini and LLM related advances from Google at the Beyond Scaling Panel tomorrow afternoon at NeurIPS, jointly with Sasha Rush ( @srush_nlp ) , Angela Fan, Percy Liang ( @percyliang ), and Jie Tang ( @jietang ).
@NeurIPSConf
NeurIPS Conference
10 months
Two invited talks down, five more to conquer
Tweet media one
0
2
30
2
9
130
@achowdhery
Aakanksha Chowdhery
1 year
Med-PaLM goes multimodal. Check it out!
@vivnat
Vivek Natarajan
1 year
Medicine is inherently multimodal. Thrilled to share Med-PaLM M, the first demonstration of a generalist multimodal biomedical AI system with a stellar team @GoogleAI @GoogleDeepMind @GoogleHealth Paper:
12
108
456
1
8
113
@achowdhery
Aakanksha Chowdhery
2 years
PaLM-SayCan combines the understanding of language models with the real-world capabilities of a helper robot. The accuracy improvements in robotic task execution from PaLM combined with SayCan are impressive. Examples of task-planning:
@hausman_k
Karol Hausman
2 years
1) We updated the underlying LLM to PaLM (), resulting in PaLM-SayCan. This resulted in an interesting trend: Improving the underlying LLM resulted in much higher robotics (!) performance (halving the errors)
Tweet media one
4
26
161
0
12
84
@achowdhery
Aakanksha Chowdhery
2 years
#NeurIPS2022 It was fun presenting PaLM and answering Q&A in the panel at "Has It Trained Yet" ( #HITY ) workshop ()! Thanks to the organizers for a great program @frankstefansch1 @zacharynado @GeorgeEDahl @naman33k @PhilippHennig5 !
Tweet media one
2
7
67
@achowdhery
Aakanksha Chowdhery
2 years
Attending #NeurIPS2022 in person this year (today-Sat)! Looking forward to catching up with many of you! DM me if you would like to meet.
2
2
60
@achowdhery
Aakanksha Chowdhery
10 months
Enthusiastic to see what the community builds with Gemini Pro models just released through Google AI Studio! Technical whitepaper on the model:
@sundarpichai
Sundar Pichai
10 months
Today developers can start building with our first version of Gemini Pro through Google AI Studio at .  Developers have a free quota and access to a full range of features including function calling, embeddings, semantic retrieval, custom knowledge
Tweet media one
344
1K
6K
1
3
58
@achowdhery
Aakanksha Chowdhery
2 years
Excited about these improvements on PaLM model: 1) U-PaLM: finetune with UL2 mixture-of-denoisers 2) Flan-PaLM: finetune on 1.8K tasks phrased as instructions You can even stack these two methods! U-PaLM: Flan-PaLM:
@YiTayML
Yi Tay
2 years
Introducing U-PaLM 540B! @GoogleAI Training PaLM w UL2's mixture-of-denoisers with only 0.1% more compute unlocks: - Much better scaling 📈 - Emergent abilities on BIGBench 😎 - Saving 2x compute (4.4 million TPU hours!) 🔥 - New prompting ability link:
Tweet media one
8
87
509
1
5
55
@achowdhery
Aakanksha Chowdhery
2 years
It was quite a fun event to discuss emerging research and applications for LLMs! Thanks for the invitation to present!
@TreybigDavis
Davis Treybig
2 years
Presentations from a recent event we hosted with @Replit @GoogleAI @huggingface on emerging research and applications of large language models: @amasad @achowdhery @mathemakitten @GoogleAI
4
12
118
0
2
42
@achowdhery
Aakanksha Chowdhery
2 years
#palm Really fun discussing PaLM pushes the frontier on many BIGbench tasks and enables PaLM-SayCan robots in @wbur @Endless_Thread podcast jointly with @jaschasd @ethansdyer @xf1280 @hausman_k (Thanks to @deanwrussell ). Url:
@WBUR
WBUR
2 years
Can robots think? For our series finale, @deanwrussell reports on AI research stretching the limits of machine learning and studying if robotic sentience REALLY matters for our future.
1
3
5
0
8
40
@achowdhery
Aakanksha Chowdhery
2 years
Combining safety+interpretability via affordance grounding with language model PaLM in robotics is really impressive. PaLM-SayCan results show that the system chooses the correct sequence of skills 84% of the time and executes them successfully 74% of the time.
@GoogleAI
Google AI
2 years
Learn how we combined our latest language model, PaLM, with robot learning algorithms to create PaLM-SayCan, a robotics system that uses natural language to complete complex tasks in a real-world environment →
8
132
407
0
3
37
@achowdhery
Aakanksha Chowdhery
10 months
If you're at #NeurIPS2023 come chat with us, the Gemini team! We're at the Google booths tomorrow from 1:30-3:00 to answer your questions on what it's like to work on Gemini.
@JeffDean
Jeff Dean (@🏡)
10 months
I'm excited to head to @NeurIPSConf #NeurIPS2023 this week. We'll be having a couple of "Chat with the Gemini Team" events in the @GoogleDeepMind / @GoogleResearch booth areas on Tuesday and Wednesday from 1:30 to 3:00 PM (New Orleans time). Quite a few Gemini team members will
12
39
502
0
1
30
@achowdhery
Aakanksha Chowdhery
2 years
Tweet media one
Tweet media two
0
1
23
@achowdhery
Aakanksha Chowdhery
2 years
11/ There is just a glimpse of the exciting research with PaLM - the list is too long to summarize here, however I am incredibly grateful to all the amazing collaborators and researchers at @GoogleAI for their contributions and innovations. And super excited for what's next!!!
5
1
21
@achowdhery
Aakanksha Chowdhery
3 years
Paper link:
0
2
17
@achowdhery
Aakanksha Chowdhery
2 years
9/ And we can enable low-latency high-throughput efficient inference at PaLM 540B scale.
@jekbradbury
James Bradbury
2 years
Can multi-100B param language models be served efficiently? We think so! Today we’re announcing the PaLM inference paper and releasing code for low-latency, high-throughput inference of 8B–540B models on TPU v4. Paper: Code: 1/5
Tweet media one
12
139
904
1
1
17
@achowdhery
Aakanksha Chowdhery
2 years
4/ Minerva finetunes PaLM on mathematical content and scientific papers to solve mathematical questions using step-by-step natural language reasoning, establishing new SOTA on STEM benchmarks, MATH and MMLU-STEM
@alewkowycz
alewkowycz
2 years
Very excited to present Minerva🦉: a language model capable of solving mathematical questions using step-by-step natural language reasoning. Combining scale, data and others dramatically improves performance on the STEM benchmarks MATH and MMLU-STEM.
Tweet media one
104
1K
8K
1
1
16
@achowdhery
Aakanksha Chowdhery
2 years
@_arohan_ Thanks for highlighting this Rohan! Implementation is also in open source and runs with T5X/Flax:
0
0
14
@achowdhery
Aakanksha Chowdhery
2 years
Really excited to see this fun and cool application of PaLM to enable robots to execute long abstract instructions from user inputs!
@hausman_k
Karol Hausman
2 years
Have you ever “heard” yourself talk in your head? Turns out it's a useful tool for robots too! Introducing Inner Monologue: feeding continual textual feedback into LLMs allows robots to articulate a grounded “thought process” to execute long, abstract instructions 🧵👇
23
168
895
0
0
15
@achowdhery
Aakanksha Chowdhery
5 years
Thanks to the support of @MarconiSociety ... our students from Celestini Program showcase their work in TF blog post!
@TensorFlow
TensorFlow
5 years
See how VisionAir is using Federated Learning and the TensorFlow Java API to estimate air quality taken from smartphone photos, while keeping user privacy in mind. 🤳🌏 Read the blog →
0
32
115
2
1
15
@achowdhery
Aakanksha Chowdhery
2 years
8/ LLMs can "hallucinate." RARR automatically researches & revises the output to fix hallucinations with help from PaLM.
@kelvin_guu
Kelvin Guu
2 years
New from Google Research! Language models perform amazing feats, but often still "hallucinate" unsupported content. Our model, RARR🐯, automatically researches & revises the output of any LM to fix hallucinations and provide citations for each sentence. 🧵
Tweet media one
12
151
816
1
2
15
@achowdhery
Aakanksha Chowdhery
2 years
10/ We discuss more about PaLM in the #MeetAGoogleResearcher episode:
@GoogleAI
Google AI
2 years
Following up on our introduction of the Pathways Language Model (PaLM), watch the first episode of #MeetAGoogleResearcher , where @drewcalcagno speaks with @achowdhery and @sharan0909 , the researchers who are making robots more helpful with language →
6
57
147
1
2
14
@achowdhery
Aakanksha Chowdhery
2 years
2/ PaLM demonstrates breakthrough capabilities on language, reasoning & code tasks, and can even explain jokes.
@JeffDean
Jeff Dean (@🏡)
3 years
Excited about this @GoogleAI work on "PaLM: Scaling Language Modeling with Pathways" with many authors. Be sure to check out the accompanying 83 page PDF!
Tweet media one
Tweet media two
Tweet media three
Tweet media four
7
109
566
1
1
13
@achowdhery
Aakanksha Chowdhery
2 years
3/ #GoogleIO showcased examples of the reasoning and multilingual capabilities of PaLM.
@Google
Google
2 years
Pathways Language Model (PaLM) is a new advanced AI model that uses a technique called chain of thought prompting to do complex tasks like solve math word problems — and even explain its reasoning process step-by-step. #GoogleIO
56
126
713
1
1
12
@achowdhery
Aakanksha Chowdhery
6 years
⚡️ “Real-time object detection with TensorFlow Lite”
2
4
12
@achowdhery
Aakanksha Chowdhery
2 years
3.2/ Multilingual capabilities of PaLM are surprising and powerful. For example, you can ask novel questions in Bengali, and get surprisingly good answers on both English and Bengali even when it has never seen parallel sentences in both languages.
@JeffDean
Jeff Dean (@🏡)
2 years
Today at #GoogleIO @sundarpichai showed some examples of the capabilities of the PaLM 540B language model. For example, you can prompt the model with: "I will ask a question in Bengali and get English and Bengali answers" And then give it two examples of this behavior. (cont)
Tweet media one
13
83
481
1
1
12
@achowdhery
Aakanksha Chowdhery
2 years
3.1/ Reasoning capabilities from combining chain-of-thought prompting with model scale have several interesting applications.
@GoogleAI
Google AI
2 years
Learn about chain of thought prompting, a method that equips language models to decompose multi-step problems into intermediate steps, enabling models of sufficient scale to solve complex reasoning problems that are not solvable with standard prompting. →
Tweet media one
13
134
639
1
1
10
@achowdhery
Aakanksha Chowdhery
6 years
Congratulations to the Celestini Program India 2018 student team supported by @MarconiSociety ! The Android demo app they built to predict Air Quality in Delhi using #TFLite featured in #TFDevSummit today and in the demos. Link:
0
2
10
@achowdhery
Aakanksha Chowdhery
2 years
@russelljkaplan I wonder if there is a new marketplace for token creation and exchange with content creators and data owners…
0
1
9
@achowdhery
Aakanksha Chowdhery
2 years
7/ Med-PaLM aligns the model to the medical domain to generate safe and helpful answers, achieving 67% in MedQA USMLE improving prior work >17%.
@vivnat
Vivek Natarajan
2 years
Delighted to share our new @GoogleHealth @GoogleAI @Deepmind paper at the intersection of LLMs + health. Our LLMs building on Flan-PaLM reach SOTA on multiple medical question answering datasets including 67.6% on MedQA USMLE (+17% over prior work).
Tweet media one
37
368
2K
1
1
9
@achowdhery
Aakanksha Chowdhery
2 years
@JeffDean Favorite so far: "Bald eagle made of chocolate powder, mango, and whipped cream."
Tweet media one
1
0
9
@achowdhery
Aakanksha Chowdhery
2 years
5/ Tasks that seem simple to humans are actually incredibly complex for helper robots. PaLM-SayCan showcases how a robotics system uses PaLM to interpret natural language to complete complex tasks in a real-world environment.
@Google
Google
2 years
Tasks that seem simple to humans — like cleaning up a spilled drink — are actually incredibly complex for helper robots. That’s why Google Research and Everyday Robots are using language models to improve robot learning.
41
82
443
1
1
8
@achowdhery
Aakanksha Chowdhery
2 years
6/ Flan-PaLM instruction tunes 540B PaLM to follow instructions, establishing a new SOTA on MMLU benchmarks and be helpful in the zero-shot setting with high accuracy.
@hwchung27
Hyung Won Chung
2 years
New paper + models! We extend instruction finetuning by 1. scaling to 540B model 2. scaling to 1.8K finetuning tasks 3. finetuning on chain-of-thought (CoT) data With these, our Flan-PaLM model achieves a new SoTA of 75.2% on MMLU.
Tweet media one
9
64
350
1
1
8
@achowdhery
Aakanksha Chowdhery
2 years
Excited to see the possibilities of model scale of PaLM combined with chain-of-thought prompting.
@JeffDean
Jeff Dean (@🏡)
2 years
Chain of thought promoting. Encouraging language models to "show their work" makes them both more interpretable and more accurate at complex reasoning tasks, solving math problems, etc.
5
19
148
0
0
8
@achowdhery
Aakanksha Chowdhery
4 years
Great discussion on the role of Internet connectivity in the current pandemic: @MarconiSociety
0
1
8
@achowdhery
Aakanksha Chowdhery
4 years
@MarconiSociety Incredibly proud of the work of students in Celestini Program India 2018.
0
3
7
@achowdhery
Aakanksha Chowdhery
5 years
@Mxbonn @petewarden For MobileNet V2, when finegrain_classification_mode is set to False, the model will shrink the last layer small for small multipliers. Please feel free to email me if there are further questions.
0
2
7
@achowdhery
Aakanksha Chowdhery
1 year
@reinerpope @MikeGunter_ Congratulations! Looking forward to the advances and growth from your company!
0
0
6
@achowdhery
Aakanksha Chowdhery
6 years
Real-time pollution check in Delhi on Android app with @TensorFlow Lite and ML Kit! Read blog post for details:
@MarconiSociety
The Marconi Society
6 years
Students at @iitdelhi are predicting and tracking #AirPollution with a new app - thanks to Prof. Brejesh Lall and Marconi Society mentor @achowdhery
0
0
2
0
3
6
@achowdhery
Aakanksha Chowdhery
6 months
@MatXComputing Congratulations!
1
0
5
@achowdhery
Aakanksha Chowdhery
2 years
@character_ai Congratulations!
0
0
5
@achowdhery
Aakanksha Chowdhery
2 years
Tweet media one
1
0
4
@achowdhery
Aakanksha Chowdhery
2 years
@ben_athi Multi query attention is still valuable in batch 1 setting enabling lower memory cost for loading the KV cache
1
0
4
@achowdhery
Aakanksha Chowdhery
2 years
Video snippet:
@Google
Google
2 years
Pathways Language Model (PaLM) is a new advanced AI model that uses a technique called chain of thought prompting to do complex tasks like solve math word problems — and even explain its reasoning process step-by-step. #GoogleIO
56
126
713
0
1
4
@achowdhery
Aakanksha Chowdhery
2 years
@AllennxDD @arankomatsuzaki Allen, PaLM 540B is actually SOTA in STEM category of MMLU. There was a correction from copying GitHub leaderboard. Please see table 6 updated version.
1
0
4
@achowdhery
Aakanksha Chowdhery
2 years
@kaushikpatnaik The downstream results are in line with Chinchilla data scaling laws. The larger models need to be trained longer with more data!
0
0
4
@achowdhery
Aakanksha Chowdhery
6 years
AI with solar cells:
0
2
3
@achowdhery
Aakanksha Chowdhery
7 years
My research at Princeton covered by Open Fog Consortium @OpenFog blog Part I ((link: ) ) and Part II ((link: ) ) on ‘‘Fog Computing on drone networks
0
0
3
@achowdhery
Aakanksha Chowdhery
2 years
0
0
3
@achowdhery
Aakanksha Chowdhery
2 years
Will be at Google Booth in the coffee break at 3:30pm Today.
0
0
2
@achowdhery
Aakanksha Chowdhery
2 years
@igorcosta @JeffDean Pathways system:
2
0
2
@achowdhery
Aakanksha Chowdhery
6 years
Real-time pet detection on Mobile with @Tensorflow Lite!
@TensorFlow
TensorFlow
6 years
Now you can train your Object Detection models on Cloud TPUs! Learn how in this end-to-end walkthrough. Bonus: we'll run the trained model on a phone using TensorFlow Lite and detect pets. Read the post here →
5
155
433
0
0
2
@achowdhery
Aakanksha Chowdhery
2 years
@RandomlyWalking @Google Amazing impact! Congratulations!
0
0
1
@achowdhery
Aakanksha Chowdhery
2 years
@hardmaru @StabilityAI Congratulations David!
0
0
1
@achowdhery
Aakanksha Chowdhery
6 years
@OhThisDataGuy Please file a GitHub issue with exact reproducible instructions if it is a model that we have already provided support for.
1
0
1
@achowdhery
Aakanksha Chowdhery
2 years
@TaliaRinger Happy Birthday!
1
0
1
@achowdhery
Aakanksha Chowdhery
2 years
@_arohan_ Credit to Jacob Devlin for the jokes here!
0
0
1
@achowdhery
Aakanksha Chowdhery
4 months
@lateinteraction Congrats Omar!
0
0
1