Google at ICML 2024
Google at ICML 2024
Researchers across Google actively explore the field of machine learning (ML), from theory to application. We build ML systems to solve deep scientific and engineering challenges in areas of language, music, visual processing, algorithm development, and more. We aim to build a more collaborative ecosystem with the broader ML research community through open-sourcing tools and datasets, publishing our work, and actively participating in conferences.
Google Research is proud to be a Diamond Sponsor of the 41st International Conference on Machine Learning (ICML 2024), a premier annual conference, which is being held this week in Vienna, Austria. As a leader in ML research, Google Research has a strong presence at this year’s conference with more than 85 accepted papers and active involvement in a number of workshops and tutorials. Google Research is also proud to be a Platinum Sponsor for both the LatinX in AI and Women in Machine Learning workshops. We look forward to sharing some of our extensive ML research and expanding our partnership with the broader ML research community.
Registered for ICML 2024? We hope you’ll visit the Google Research booth to learn more about the exciting work, creativity, and fun that goes into solving a portion of the field’s most interesting challenges. Visit the @GoogleAI X (Twitter) account to find out about Google Research booth activities (e.g., demos and Q&A sessions). See Google DeepMind’s blog to learn about their technical participation at ICML 2024.
Take a look below to learn more about the Google research being presented at ICML 2024 (Google affiliations in bold).
Quick links
Quick links
Board & Organizing Committee
-
Fei Sha
- Senior Area Chair
-
Amin Karbasi
- Communications Chair
-
Katherine Heller
- Program Chair & Board Member
-
Corinna Cortes
- Board Members
Expo talks
-
Sun, July 21 | 4:00PM — 5:00PM, Hall A8
AI for software development at GooglePresenters: Alexander Frömmgen & Maxim Tabachnyk
-
Sun, July 21 | 1:00PM — 2:00PM, Hall A8
Giving your Graph a Voice: Graph Representations and Large Language ModelsPresenters: Bryan Perozzi & Sami Abu-el-haija
Orals
-
Tue, July 23 | 10:30AM — 11:30AM, Hall A8 (Oral 1D Video)
VideoPoet: A Large Language Model for Zero-Shot Video GenerationDan Kondratyuk, Lijun Yu, Xiuye Gu, Jose Lezama, Jonathan Huang, Grant Schindler, Rachel Hornung, Vighnesh Birodkar, Jimmy Yan, Ming-Chang Chiu, Krishna Somandepalli, Hassan Akbari, Yair Alon, Yong Cheng, Josh Dillon, Agrim Gupta, Meera Hahn, Anja Hauth, David Hendon, Alonso Martinez, David Minnen, Mikhail Sirotenko, Kihyuk Sohn, Xuan Yang, Hartwig Adam, Ming-Hsuan Yang, Irfan Essa, Huisheng Wang, David A. Ross, Bryan Seybold, Lu Jiang
-
Tue, July 23 | 4:30PM — 5:30PM, Hall A2 (Oral 2C Privacy)
How Private are DP-SGD Implementations?Lynn Chua, Badih Ghazi, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Amer Sinha, Chiyuan Zhang
-
Tue, July 23 | 4:30PM — 5:30 PM, Hall A2 (Oral 2C Privacy)
Private Truly-Everlasting Robust-PredictionUri Stemmer
-
Wed, July 24 | 4:30PM — 5:30PM, Hall C 1-3 (Oral 4A: Reinforcement Learning 2)
Rate-Optimal Policy Optimization for Linear Markov Decision ProcessesUri Sherman, Alon Cohen, Tomer Koren, Yishay Mansour
Accepted papers
Perturb-and-Project: Differentially Private Similarities and Marginals
Vincent Cohen-Addad, Tommaso d'Orsi, Alessandro Epasto, Vahab Mirrokni, Peilin Zhong
Replicable Learning of Large-Margin Halfspaces
Alkis Kalavasis, Amin Karbasi, Kasper Green Larsen, Grigoris Velegkas, Felix Zhou
Decoding-time Realignment of Language Models
Tianlin Liu, Shangmin Guo, Leonardo Bianco*, Daniele Calandriello, Quentin Berthet, Felipe Llinares-López, Jessica Hoffmann, Lucas Dixon, Michal Valko, Mathieu Blondel
Target Networks and Over-parameterization Stabilize Off-policy Bootstrapping with Function Approximation
Fengdi Che, Chenjun Xiao, Jincheng Mei, Bo Dai, Ramki Gummadi, Oscar A Ramirez*, Christopher K Harris*,
A. Rupam Mahmood, Dale Schuurmans
Dynamic Correlation Clustering in Sublinear Update Time
Vincent Cohen-Addad, Silvio Lattanzi, Andreas Maggiori, Nikos Parotsidis
PriorBoost: An Adaptive Algorithm for Learning from Aggregate Responses
Adel Javanmard, Matthew Fahrbach, Vahab Mirrokni
How Free is Parameter-Free Stochastic Optimization?
Amit Attia, Tomer Koren
Practical Performance Guarantees for Pipelined DNN Inference
Aaron Archer, Matthew Fahrbach, Kuikui Liu, Prakash Prabhu
Regression with Multi-Expert Deferral
Anqi Mao, Mehryar Mohri, Yutao Zhong
Data-Efficient Learning via Clustering-Based Sensitivity Sampling: Foundation Models and Beyond
Kyriakos Axiotis, Vincent Cohen-Addad, Monika Henzinger, Sammy Jerome, Vahab Mirrokni, David Saulpic, David Woodruff, Michael Wunder
Isometric Representation Learning for Disentangled Latent Space of Diffusion Models
Jaehoon Hahm, Junho Lee, Sunghyun Kim, Joonseok Lee
Learning from Students: Applying t-Distributions to Explore Accurate and Efficient Formats for LLMs
Jordan Dotzel, Yuzong Chen, Bahaa Kotb, Sushma Prasad, Gang Wu, Sheng Li, Mohamed S. Abdelfattah, Zhiru Zhang
LEVI: Generalizable Fine-tuning via Layer-wise Ensemble of Different Views
Yuji Roh, Qingyun Liu, Huan Gui, Zhe Yuan, Yujin Tang, Steven Euijong Whang, Liang Liu, Shuchao Bi,
Lichan Hong, Ed H. Chi, Zhe Zhao
Out of the Ordinary: Spectrally Adapting Regression for Covariate Shift
Benjamin Eyre, Elliot Creager, David Madras, Vardan Papyan, Richard Zemel
Privacy-Preserving Instructions for Aligning Large Language Models
Da Yu*, Peter Kairouz, Sewoong Oh, Zheng Xu
Representation Surgery: Theory and Practice of Affine Steering
Shashwat Singh, Shauli Ravfogel*, Jonathan Herzig, Roee Aharoni, Ryan Cotterell, Ponnurangam Kumaraguru
A Statistical Framework for Data-dependent Retrieval-Augmented Models
Soumya Basu, Ankit Singh Rawat, Manzil Zaheer
Two Heads are Actually Better than One: Towards Better Adversarial Robustness via Transduction and Rejection
Nils Palumbo, Yang Guo, Xi Wu, Jiefeng Chen, Yingyu Liang, Somesh Jha
Bayesian Regret Minimization in Offline Bandits
Marek Petrik, Guy Tennenholtz, Mohammad Ghavamzadeh
Break the Sequential Dependency of LLM Inference Using Lookahead Decoding
Yichao Fu, Peter Bailis, Ion Stoica, Hao Zhang
Do Large Code Models Understand Programming Concepts? Counterfactual Analysis for Code Predicates
Ashish Hooda*, Mihai Christodorescu, Miltiadis Allamanis, Aaron Wilson, Kassem Fawaz, Somesh Jha
DySLIM: Dynamics Stable Learning by Invariant Measure for Chaotic Systems
Yair Schiff, Zhong Yi Wan, Jeffrey B. Parker, Stephan Hoyer, Volodymyr Kuleshov, Fei Sha, Leonardo Zepeda-Núñez
A Field Guide for Pacing Budget and ROS Constraints
Santiago R. Balseiro, Kshipra Bhawalkar, Zhe Feng, Haihao Lu, Vahab Mirrokni, Balasubramanian Sivan, Di Wang
How Private is DP-SGD?
Lynn Chua, Badih Ghazi, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Amer Sinha, Chiyuan Zhang
Improved Differentially Private and Lazy Online Convex Optimization: Lower Regret without Smoothness Requirements
Naman Agarwal, Satyen Kale, Karan Singh, Abhradeep Guha Thakurta
LayerMerge: Neural Network Depth Compression through Layer Pruning and Merging
Jinuk Kim, Marwa El Halabi, Mingi Ji, Hyun Oh Song
Learning and Forgetting Unsafe Examples in Large Language Models
Jiachen Zhao, Zhun Deng, David Madras, James Zou, Mengye Ren
A Near-Linear Time Approximation Algorithm for Beyond-Worst-Case Graph Clustering
Vincent Cohen-Addad, Tommaso d'Orsi, Aida Mousavifar
The Non-linear F-Design and Applications to Interactive Learning
Alekh Agarwal, Jian Qian, Alexander Rakhlin, Tong Zhang
Pi-DUAL: Using Privileged Information to Distinguish Clean from Noisy Labels
Ke Wang, Guillermo Ortiz-Jimenez, Rodolphe Jenatton, Mark Collier, Efi Kokiopoulou, Pascal Frossard
Position: Cracking the Code of Cascading Disparity Towards Marginalized Communities
Golnoosh Farnadi, Mohammad Havaei, Negar Rostamzadeh
Unmasking Vulnerabilities: Cardinality Sketches Under Adaptive Inputs
Sara Ahmadian, Edith Cohen
What is Dataset Distillation Learning?
William Yang, Ye Zhu, Zhiwei Deng, Olga Russakovsky
Can Looped Transformers Learn to Implement Multi-step Gradient Descent for In-context Learning?
Khashayar Gatmiry, Nikunj Saunshi, Sashank J. Reddi, Stefanie Jegelka, Sanjiv Kumar
Cell2Sentence: Teaching Large Language Models the Language of Biology
Daniel Levine, Syed A Rizvi, Sacha Lévy, Nazreen Pallikkavaliyaveetil, David Zhang, Xingyu Chen,
Sina Ghadermarzi, Ruiming Wu, Zihe Zheng, Ivan Vrkic, Anna Zhong, Daphne Raskin, Insu Han, Antonio Henrique de Oliveira Fonseca, Josue Ortega Caro, Amin Karbasi, Rahul Madhav Dhodapkar, David van Dijk
Consistent Submodular Maximization
Paul Duetting, Federico Fusco, Silvio Lattanzi, Ashkan Norouzi-Fard, Morteza Zaddimoghadam
Controlled Decoding from Language Models
Sidharth Mudgal, Jong Lee, Harish Ganapathy, YaGuang Li, Tao Wang*, Yanping Huang, Zhifeng Chen, Heng-Tze Cheng, Michael Collins, Trevor Strohman, Jilin Chen, Alex Beutel*, Ahmad Beirami
Differentially Private Domain Adaptation with Theoretical Guarantees
Raef Bassily, Corinna Cortes, Anqi Mao, Mehryar Mohri
Eluder-Based Regret for Stochastic Contextual MDPs
Orin Levy, Asaf Cassel, Alon Cohen, Yishay Mansour
A Minimaximalist Approach to Reinforcement Learning from Human Feedback
Gokul Swamy*, Christoph Dann, Rahul Kidambi, Zhiwei Steven Wu, Alekh Agarwal
Multi-View Stochastic Block Models
Vincent Cohen-Addad, Tommaso d'Orsi, Silvio Lattanzi, Rajai Nasser
Near-Optimal Regret in Linear MDPs with Aggregate Bandit Feedback
Asaf Cassel, Haipeng Luo, Aviv Rosenberg, Dmitry Sotnikov
Patchscopes: A Unifying Framework for Inspecting Hidden Representations of Language Models (see blog post)
Asma Ghandeharioun, Avi Caciularu, Adam Pearce, Lucas Dixon, Mor Geva
Robust Inverse Graphics via Probabilistic Inference
Tuan Anh Le, Pavel Sountsov, Matthew Douglas Hoffman, Ben Lee, Brian Patton, Rif A. Saurous
Score identity Distillation: Exponentially Fast Distillation of Pretrained Diffusion Models for One-Step Generation
Mingyuan Zhou, Huangjie Zheng, Zhendong Wang, Mingzhang Yin, Hai Huang
Tandem Transformers for Inference Efficient LLMs
Aishwarya P S, Pranav Ajit Nair, Yashas Samaga B L, Toby James Boyd, Sanjiv Kumar, Prateek Jain, Praneeth Netrapalli
Transforming and Combining Rewards for Aligning Large Language Models
Zihao Wang, Chirag Nagpal, Jonathan Berant, Jacob Eisenstein, Alexander D'Amour, Sanmi Koyejo,
Victor Veitch
USTAD: Unified Single-Model Training Achieving Diverse Scores for Information Retrieval
Seungyeon Kim, Ankit Singh Rawat, Manzil Zaheer, Wittawat Jitkrittum, Veeranjaneyulu Sadhanala, Sadeep Jayasumana, Aditya Krishna Menon, Rob Fergus, Sanjiv Kumar
Adaptive Accompaniment with ReaLchords
Yusong Wu, Tim Cooijmans, Kyle Kastner, Adam Roberts, Ian Simon, Alexander Scarlatos, Chris Donahue, Cassie Tarakajian, Shayegan Omidshafiei*, Aaron Courville, Pablo Samuel Castro, Natasha Jaques, Cheng-Zhi Anna Huang
A Decoder-Only Foundation Model for Time-Series Forecasting (see blog post)
Abhimanyu Das, Weihao Kong, Rajat Sen, Yichen Zhou
Deep Fusion: Efficient Network Training via Pre-trained Initializations
Hanna Mazzawi, Javier Gonzalvo, Michael Wunder, Sammy Jerome, Benoit Dherin
Extracting Training Data from Document-Based VQA Models
Francesco Pinto, Nathalie, Rauschmayr, Florian Tramer, Philip Torr, Federico Tombari
FrameQuant: Flexible Low-Bit Quantization for Transformers
Harshavardhan Adepu, Zhanpeng Zeng, Li Zhang, Vikas Singh
H-Consistency Guarantees for Regression
Anqi Mao, Mehryar Mohri, Yutao Zhong
Implicit Bias of Policy Gradient in Linear Quadratic Control: Extrapolation to Unseen Initial States
Noam Razin, Yotam Alexander, Edo Cohen-Karlik, Raja Giryes, Amir Globerson, Nadav Cohen
Interpretability Illusions in the Generalization of Simplified Models
Dan Friedman*, Andrew Kyle Lampinen, Lucas Dixon, Danqi Chen, Asma Ghandeharioun
Large Language Models Can Automatically Engineer Features for Few-Shot Tabular Learning
Sungwon Han*, Jinsung Yoon, Sercan O Arik, Tomas Pfister
MC-GTA: Metric-Constrained Model-Based Clustering Using Goodness-of-Fit Tests with Autocorrelations
Zhangyu Wang, Gengchen Mai, Krzysztof Janowicz, Ni Lao
Mean Estimation in the Add-Remove Model of Differential Privacy
Alex Kulesza, Ananda Suresh, Yuyan Wang
More Benefits of Being Distributional: Second-Order Bounds for Reinforcement Learning
Kaiwen Wang, Owen Oertell, Alekh Agarwal, Nathan Kallus, Wen Sun
Online Learning with Bounded Recall
Jon Schneider, Kiran Vodrahalli
Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Lu Yin, You Wu, Zhenyu Zhang, Cheng-Yu Hsieh, Yaqing Wang, Yiling Jia, Gen Li, Ajay Kumar Jaiswal, Mykola Pechenizkiy, Yi Liang, Michael Bendersky, Zhangyang Wang, Shiwei Liu
Promises and Pitfalls of Generative Masked Language Modeling: Theoretical Framework and Practical Guidelines
Yuchen Li, Alexandre Kirchmeyer, Aashay Mehta, Yilong Qin, Boris Dadachev, Kishore Papineni, Sanjiv Kumar, Andrej Risteski
SCoRe: Submodular Combinatorial Representation Learning
Anay Majee, Suraj Kothawade, Krishnateja Killamsetty, Rishabh K Iyer
Simplicity Bias via Global Convergence of Sharpness Minimization
Khashayar Gatmiry, Zhiyuan Li, Sashank J. Reddi, Stefanie Jegelka
Auto-Linear Phenomenon in Subsurface Imaging
Yinan Feng, Yinpeng Chen, Peng Jin, Shihang Feng, Youzuo Lin
FRAPPÉ: A Group Fairness Framework for Post-Processing Everything
Alexandru Tifrea*, Preethi Lahoti, Ben Packer, Yoni Halpern, Ahmad Beirami, Flavien Prost
Individualized Privacy Accounting via Subsampling with Applications in Combinatorial Optimization
Badih Ghazi, Pritish Kamath, Ravi Kumar, Pasin Manurangsi, Adam Sealfon
Online Speculative Decoding
Xiaoxuan Liu, Lanxiang Hu, Peter Bailis, Alvin Cheung, Zhijie Deng, Ion Stoica, Hao Zhang
The Pitfalls of Next-Token Prediction
Gregor Bachmann, Vaishnavh Nagarajan
PolySketchFormer: Fast Transformers via Sketching Polynomial Kernels
Praneeth Kacham, Vahab Mirrokni, Peilin Zhong
Position: Social Environment Design Should be Further Developed for AI-based Policy-Making
Edwin Zhang, Sadie Zhao, Tonghan Wang, Safwan Hossain, Henry Gasztowtt, Stephan Zheng, David C. Parkes, Milind Tambe, Yiling Chen
Prompt-Tuning Latent Diffusion Models for Inverse Problems
Hyungjin Chung, Jong Chul Ye, Peyman Milanfar, Mauricio Delbracio
VideoPrism: A Foundational Visual Encoder for Video Understanding (see blog post)
Long Zhao, Nitesh Bharadwaj Gundavarapu, Liangzhe Yuan, Hao Zhou, Shen Yan, Jennifer J. Sun, Luke Friedman, Rui Qian, Tobias Weyand, Yue Zhao*, Rachel Hornung, Florian Schroff, Ming-Hsuan Yang, David A Ross, Huisheng Wang, Hartwig Adam, Mikhail Sirotenko, Ting Liu, Boqing Gong
RLAIF vs. RLHF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Harrison Lee, Samrat Phatale, Hassan Mansoor, Thomas Mesnard, Johan Ferret, Kellie Ren Lu, Colton Bishop, Ethan Hall, Victor Carbune, Abhinav Rastogi, Sushant Prakash
From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformers
Muhammed Emrullah Ildiz, Yixiao HUANG, Yingcong Li, Ankit Singh Rawat, Samet Oymak
Generalized Neural Collapse for a Large Number of Classes
Jiachen Jiang, Jinxin Zhou, Peng Wang, Qing Qu, Dustin G. Mixon, Chong You, Zhihui Zhu
High-Dimensional Geometric Streaming for Nearly Low Rank Data
Hossein Esfandiari, Praneeth Kacham, Vahab Mirrokni, David Woodruff, Peilin Zhong
Improved Communication-Privacy Trade-Offs in L2 Mean Estimation Under Streaming Differential Privacy
Wei-Ning Chen, Berivan Isik, Peter Kairouz, Albert No, Sewoong Oh, Zheng Xu
On Discrete Prompt Optimization for Diffusion Models
Ruochen Wang, Ting Liu, Cho-Jui Hsieh, Boqing Gong
OSSCAR: One-Shot Structured Pruning in Vision and Language Models with Combinatorial Optimization
Xiang Meng, Shibal Ibrahim, Kayhan Behdin, Hussein Hazimeh, Natalia Ponomareva, Rahul Mazumder
Weisfeiler-Leman at the Margin: When More Expressivity Matters
Billy Joe Franks, Christopher Morris, Ameya Velingker, Floris Geerts
Workshops & tutorials
-
Mon, July 22 | 3:30PM — 5:30PM, Hall A8
Graph Learning: Principles, Challenges, and Open DirectionsAdrián Arnaiz-Rodríguez, Ameya Velingker
-
Fri, July 26 | 9:00AM — 3:00PM, Lehar 1
AI for Math WorkshopOrganizer: Isabelle Guyon
-
Fri, July 26 | 9:00AM — 6:00PM, Stolz 2
ML for Life and Material Science: From Theory to Industry ApplicationsOrganizer: Preethi Lahoti
-
Fri, July 26 | 9:00AM, Hall A1
Next Generation of AI SafetyOrganizers: Preethi Lahoti, Ahmad Beirami
-
Fri, July 26 | 9:00AM — 5:00PM, Straus 3
Next Generation of Sequence Modeling ArchitecturesSpeakers: Joao Sacramento
-
Fri, July 26 | 9:00AM — 6:00PM, Schubert 4 - 6
Structured Probabilistic Inference and Generative ModelingOrganizers: Ruiqi Gao
-
Sat, July 27 | 9:00AM — 9:00PM, Straus 3
Data-Centric Machine Learning Research (DMLR): Datasets for Foundation ModelsSpeaker: Lucas Beyer
-
Sat, July 27 | 9:00AM, Lehar 4
1st ICML Workshop on In-Context Learning (ICL @ ICML 2024)Speaker: Mehran Kazemi
-
Sat, July 27 | 9:00AM, Lehar 1
Mechanistic InterpretabilityOrganizer: Mor Geva
-
Sat, July 27 | 9:00AM, HallA8
Text, Camera, Action! Frontiers in Controllable Video GenerationSpeaker: Tali Dekel
-
Sat, July 27 | 9:00AM, Straus 2
Theoretical Foundations of Foundation Models (TF2M)Organizer: Ziteng Sun
Demos and Q&A at the Google Research Booth
– Dates and times may be subject to change. Stop by the Google Research booth (#118) for more details.
-
Mon, July 22 | 12:30PM
Engaging in Research at GoogleKatherine Heller
-
Mon, July 22 | 1:00PM
Recruiting Q&AKarishma Buxani Stauder, Connor McCleary, Adrian Opera & Kelsie Pedone
-
Mon, July 22 | 3:00PM
Graph Reasoning with LLMsBryan Perozzi
-
Tues, July 23 | 10:00AM
Large-Scale Foundation Models Applied to BiologyAnkit Gupta (from Ginkgo Bioworks)
-
Tues, July 23 | 1:30PM
Retrieval Augmented Generation (RAG) Quality Control for Google SearchAndrew Gilchrist-Scott, Smriti Pramanick & Brad Stocks
-
Tues, July 23 | 2:00PM
Recruiting Q&AKarishma Buxani Stauder, Connor McCleary, Adrian Opera & Kelsie Pedone
-
Tues, July 23 | 4:00PM
Productionizing Extreme Low-Latency LLMs at Google Search ScaleAndrew Gilchrist-Scott, Smriti Pramanick & Brad Stocks
-
Wed, July 24 | 10:00AM
Deep Retrieval at GoogleFelix Yu
-
Wed, July 24 | 1:30PM
Multi-task Encoders at Google Search ScaleAndrew Gilchrist-Scott, Smriti Pramanick & Brad Stocks
-
Wed, July 24 | 4:00PM
Attention-Based Model Structure OptimizationThomas Fu
* Work done while at Google