Zi Wang, Ph.D. Profile Banner
Zi Wang, Ph.D. Profile
Zi Wang, Ph.D.

@ziwphd

1,346
Followers
180
Following
18
Media
80
Statuses

Research Scientist @ Google DeepMind. CS PhD @ MIT CSAIL. Opinions my own.

Cambridge, MA
Joined September 2013
Don't wanna be here? Send us removal request.
@ziwphd
Zi Wang, Ph.D.
2 years
Interested in Bayesian optimization? Check out Google BayesOpt Speaker Series! We’ve had a fascinating series of talks given by speakers including @arkrause , Yoshua Bengio, @peter_i_frazier , Kirthevasan Kandasamy, @PhilippHennig5 .. Stay tuned for more!
1
62
342
@ziwphd
Zi Wang, Ph.D.
6 years
Still don’t believe reinforcement learning is an appropriate model for human-level intelligence. We are taught delayed gratitude all our lives while RL agents use discounted reward?! Like wut?! I’ll be so depressed if I’m an RL agent.
11
14
180
@ziwphd
Zi Wang, Ph.D.
2 years
Alex Terenin's new Google TechTalk on Gaussian processes is a must-watch! @avt_im He offers some really cool perspectives on this statistical tool. I learned a lot from it, and I'm sure you will too!
3
28
162
@ziwphd
Zi Wang, Ph.D.
10 months
Our team at @GoogleDeepMind is hiring a student researcher based in Cambridge MA in 2024 to work on uncertainty-aware decision making for LLMs and/or text-to-image models. Check out the full info and how to apply:
1
17
153
@ziwphd
Zi Wang, Ph.D.
6 years
@Tkaraletsos @roydanroy I had an even funnier review from Nips before: what’s that little square in the end of the proof
2
1
49
@ziwphd
Zi Wang, Ph.D.
5 years
Tweet media one
Tweet media two
3
0
45
@ziwphd
Zi Wang, Ph.D.
6 years
Very fancy lunch at MIT ML retreat. Grad students can’t believe they are having cup cakes with the cup made of actual chocolate. Formal sit down lunch rather than sandwich box. Mind blown.
Tweet media one
Tweet media two
Tweet media three
0
2
47
@ziwphd
Zi Wang, Ph.D.
2 years
Jasper @latentjasper talking about the ongoing journey towards BIG Gaussian processes! A team effort with @hoonkp , Ben Adlam, @shreyaspadhy and @zacharynado . Join us at NeurIPS GP workshop
Tweet media one
0
6
43
@ziwphd
Zi Wang, Ph.D.
10 months
Introducing Gaussian Process Probes (GPP) for probing and measuring uncertainty about concepts represented by AI models. #NeurIPS poster #1504 this Tue at 5pm! Paper: Joint work with @alex_y_ku , @jasonbaldridge , Tom Griffiths @cocosci_lab and @_beenkim
Tweet media one
3
8
32
@ziwphd
Zi Wang, Ph.D.
2 years
Interpretable BayesOpt solution via sparsity! Nice talk by @su_lin_liu at NeurIPS GP workshop!
Tweet media one
0
1
26
@ziwphd
Zi Wang, Ph.D.
2 years
Where should Bayesian priors come from and how can we trust them for decision making? tldr is pre-training to align with expert beliefs! Kudos to @GeorgeEDahl , @kswersk , Chansoo Lee, @zacharynado , @jmgilmer , @latentjasper , @ZoubinGhahrama1 for the collab!
@GoogleAI
Google AI
2 years
Hyper Bayesian optimization (HyperBO) is a highly customizable interface that pre-trains a Gaussian process model and automatically defines model parameters, making Bayesian optimization easier to use while outperforming traditional methods. Learn more →
10
132
643
0
2
14
@ziwphd
Zi Wang, Ph.D.
10 months
Using LLMs for a new domain-specific language? We developed Grammar Prompting to teach them rules&structures, boosting their skills in semantic parsing, planning 🤖, and even molecule generation! Excited to share this work #NeurIPS2023 poster #329 Tuesday at 5pm!
@bailin_28
Bailin Wang
10 months
@Ziwphd and I will present our Grammar Prompting work tomorrow at #NeurIPS2023 Happy to chat more on - LLMs for structured language generation (e.g., programs, molecules and robotic plans) - Structured/formal chain-of-thought for reasoning
Tweet media one
Tweet media two
1
3
20
0
2
13
@ziwphd
Zi Wang, Ph.D.
2 years
Join us at Room 387 for NeurIPS workshop on Gaussian processes! @ZoubinGhahrama1 giving a opening remark now
Tweet media one
0
3
13
@ziwphd
Zi Wang, Ph.D.
2 years
InfoBax! @willieneis on general information theoretic framework for algorithm execution @ NeurIPS GP workshop
Tweet media one
0
3
12
@ziwphd
Zi Wang, Ph.D.
8 months
Ever struggled with transferring knowledge in Bayesian optimization due to varying dimensions? Our TMLR paper solves this with MPHD! New Gaussian process Model Pre-training method works seamlessly across Heterogeneous Domains. #MachineLearning
3
3
20
@ziwphd
Zi Wang, Ph.D.
2 years
Check out to learn more about how to pre-train a Gaussian process!
@kswersk
Kevin Swersky
2 years
This is a really natural framework to improve Bayesian optimization when you have access to related optimization tasks Joint work with @ziwphd , @GeorgeEDahl , Chansoo Lee, @zacharynado , @jmgilmer , @latentjasper , @ZoubinGhahrama1
1
5
27
0
1
11
@ziwphd
Zi Wang, Ph.D.
2 years
Awesome talk by @anmakaro from ETH Zurich on Bayesian Optimization in the Wild: Risk-Averse Decisions and Budget Constraints, featuring best paper award at #AutoML_Conf . Check it out on Google BayesOpt @ YouTube! #bayesopt #automl
0
1
9
@ziwphd
Zi Wang, Ph.D.
11 months
Do we really need a gigantic amount of data to train AI models? We've been exploring this question from several facets and here's our first step to ensure robustness during data downsampling!
@ilijabogunovic
Ilija Bogunovic
11 months
How to downsample data to safeguard against worst-case generalization performance issues (stemming from class imbalance and distributional shifts)? Great work with @bankes_william , G. Hughes and @ziwphd
0
1
7
0
1
7
@ziwphd
Zi Wang, Ph.D.
2 years
Never thought I would ever say this, but look.. transformers can do optimization, quite well 😬 glad to be involved in this project.
@GoogleAI
Google AI
2 years
Today on the blog learn about the OptFormer, one of the first Transformer-based frameworks for hyperparameter tuning, learned from large-scale optimization data using flexible text-based representations →
10
115
457
0
0
6
@ziwphd
Zi Wang, Ph.D.
2 years
Special thanks to our video producer Mark Chow and our organizing/advisory committee @latentjasper , Chansoo Lee, @SagiPerel and @ZoubinGhahrama1 !
0
1
5
@ziwphd
Zi Wang, Ph.D.
8 months
Thank you to those who expressed interest in the position. I've received hundreds of emails and unfortunately won't have the bandwidth to reply. Please make sure to apply through the official channel directly
@ziwphd
Zi Wang, Ph.D.
10 months
Our team at @GoogleDeepMind is hiring a student researcher based in Cambridge MA in 2024 to work on uncertainty-aware decision making for LLMs and/or text-to-image models. Check out the full info and how to apply:
1
17
153
1
0
5
@ziwphd
Zi Wang, Ph.D.
8 months
MPHD enhances Bayesian optimization with domain adaptability. Using a neural net pre-trained in existing domains, MPHD generates tailored Gaussian process models for new domains, leading to smarter optimization decisions.
Tweet media one
0
0
3
@ziwphd
Zi Wang, Ph.D.
2 years
Super cool talk by Frank Hutter at the latest Google BayesOpt Speaker Series: “Deep Learning 2.0: How Bayesian Optimization May Power the Next Generation of DL”. A new era of deep learning? Thank you @FrankRHutter for joining us!
0
0
3
@ziwphd
Zi Wang, Ph.D.
8 months
Congrats to my wonderful co-authors Zhou Fan @evensgn & Nicole Han @nahelocin from @Harvard for the paper acceptance!
1
0
3
@ziwphd
Zi Wang, Ph.D.
8 months
@evensgn @nahelocin @Harvard Our paper reveals a fundamental connection: multi-task learning and single-task learning are equivalent for stationary kernels. This theoretical insight means MPHD models are guaranteed to improve over time, ultimately reflecting the real-world process that produces the data.
Tweet media one
0
2
4
@ziwphd
Zi Wang, Ph.D.
10 months
@alex_y_ku @jasonbaldridge @cocosci_lab @_beenkim GPP judges uncertainty rationally, just like a rational human, e.g. no extreme judgements if no evidence, and getting close to a firm judgment only if episteme is high. We can even prove theoretically that episteme has no upper bound, i.e., there is no limit to knowledge.
Tweet media one
1
0
2
@ziwphd
Zi Wang, Ph.D.
10 months
With episteme, GPP knows when it does not know. This means GPP not only probes, but also detects out-of-distribution data! On 3D Shapes and ImageNet OOD detection tasks, GPP performed consistently well and on par with the best benchmark method for each task.
Tweet media one
0
0
1
@ziwphd
Zi Wang, Ph.D.
10 months
@alex_y_ku @jasonbaldridge @cocosci_lab @_beenkim GPP probes model representation with epistemic and aleatory uncertainty evaluated as episteme (knowledge / confidence), alea (intrinsic randomness) and judged probability. This helps GPP distinguish fuzzy concepts and unknown concepts, an ability linear probes do not possess.
Tweet media one
1
0
1
@ziwphd
Zi Wang, Ph.D.
2 years
Contributed Talk by Renzhi Chen on hyperparameter hardware/software codesign! Join us at NeurIPS GP workshop
Tweet media one
0
1
1