Machine Learning at the Flatiron Institute: Brandon Amos

Date


Title: On meta prompt optimization and coding agents

Abstract: Language models (LMs) encode vast amounts of human knowledge in a queryable form and are now widely deployed across domains that previously could not directly benefit from language. Yet major challenges remain: how to effectively extract information (e.g., where and how to prompt the model), and how to ensure alignment with human values and preferences (e.g., personalization, safety, undesirable behaviors). A path forward to solving these is to quantify the objective, perform (test-time) optimization over the prompt space to improve that objective, and ultimately integrate the improvement back into the base model. This talk explores this in two domains:

1) Prompt optimization and adversarial attacks — in AdvPrompter (ICML 2025) we learn a meta-prompting model that rapidly predicts the optimal adversarial suffix for jailbreaking an LM, and in AdvPrefix (NeurIPS 2025), we show variants on the jailbreaking optimization problem objective can result in significantly improved performance, and

2) Numerical code synthesis with agents — in AlgoTune (NeurIPS D&B 2025), we propose a benchmark suite of 154 numerical coding tasks along with a baseline code agent to solve it. The coding agent iteratively optimizes the code (and prompts) for the task’s runtime as the objective.

We’ll draw connections between these two seemingly separate settings and show how meta-learning and agentic test-time computations squeeze the most out of the base model, advancing both capabilities and alignment.

About the Speaker

Brandon Amos is a Research Scientist at Meta Superintelligence Labs in NYC. He holds a PhD in Computer Science from Carnegie Mellon University where he was supported by the NSF Graduate Research Fellowship. Prior to joining Meta, he also worked at Adobe Research, Google DeepMind, and Intel Labs. He has received best reviewer awards at AISTATS, NeurIPS, ICML, and ICLR, and best paper awards at the ICML Theoretical Foundations Workshop and ACM MMSys. His research focuses on foundational topics spanning machine learning, optimization, reinforcement learning, and control, with the goal of building safe intelligent systems that understand and interact with our world. Recently he has been focusing on language modeling and diffusion/flow-based generative modeling. Major themes of his research involve attacking language models to improve safety and alignment, improving RL and control systems, applied optimal transport and flows, amortization and meta-learning between tasks, and integrating structural information and domain knowledge into AI systems through differentiable optimization.

Advancing Research in Basic Science and MathematicsSubscribe to Flatiron Institute announcements and other foundation updates

privacy consent banner

Privacy preference

We use cookies to provide you with the best online experience. By clicking "Accept All," you help us understand how our site is used and enhance its performance. You can change your choice at any time here. To learn more, please visit our Privacy Policy.