FAGEN
ICML 2026 Workshop

Failure Modes in Agentic AI

Reproducible Triggers, Trace Diagnostics, and Verified Fixes

News

Workshop X account is live

Follow @FAGENWorkshop for deadline reminders, accepted-paper highlights, and program updates.

Follow @FAGENWorkshop

OpenReview submissions are open

The submission portal is now live. Submission deadline May 8 (AOE); notifications by May 15.

Open OpenReview

Workshop accepted at ICML 2026!

About the Workshop

Reliability has been studied in ML for a long time, mostly through robustness benchmarks, adversarial evaluation, and red-teaming on chat-style language models. Foundation-model agents push the question somewhere harder. An agent run goes for hundreds of steps, each step depending on tool calls and memory writes from the steps before it. When the run breaks, it rarely breaks at the obvious moment. A bad assumption at step 3 quietly contaminates step 50, and by step 200 the agent has been wrong for a while without noticing. It might have spent its budget on the wrong subtask. It might be reading from memory it polluted itself. Or it landed on an answer at step 12 and spent the rest of the run defending it.

FAGEN is a place to take these failures seriously. The workshop is organized around four kinds of contributions. Definitions matter: what does "failure" actually mean here, beyond the loose way the term gets thrown around? Reproducible triggers matter at least as much. We want the smallest setup that breaks the agent the same way every time, so other groups can build on the case. Diagnostics should look at the trace itself, not just the final score, because final-score evaluation hides almost everything interesting. And the fixes worth presenting are the ones that admit what they cost in latency, in capability, or in how well they generalize.

Tentative Speakers & Panelists

Yoshua Bengio

Yoshua Bengio

Universite de Montreal, LawZero, and Mila

AI safety, frontier model governance, and deep learning foundations

Bo Li

Bo Li

University of Illinois Urbana-Champaign and Virtue AI

Agent attack surfaces and tool-chain exploits

Iryna Gurevych

Iryna Gurevych

TU Darmstadt and MBZUAI

Long-horizon drift and error localization

Greg Durrett

Greg Durrett

New York University

Verification gaps and grounded checking

Samy Bengio

Samy Bengio

Apple and EPFL

Reasoning brittleness and evaluation artifacts

Nouha Dziri

Nouha Dziri

Allen Institute for AI

Faithfulness failures and feedback grounding

Maarten Sap

Maarten Sap

Carnegie Mellon University and AI2

Social failure modes and human-agent risk

Yu Su

Yu Su

NeoCognition / OSU

Web-agent failures and online evaluation gaps

Rishi Bommasani

Rishi Bommasani

Stanford HAI

Frontier AI governance and deployment implications

Andrea Zanette

Andrea Zanette

Carnegie Mellon University

RL training and feedback failures in language-model agents

Call for Papers

Format

  • 01Please use an ICML template for your submission. Max 8 pages excluding references and appendix.
  • 02This is a non-archival venue. Authors retain rights and can publish their work elsewhere. Dual submission is allowed.
  • 03Accepted papers are eligible for a Best Paper Award.

Submit on OpenReview

Open OpenReview

Topics of Interest

We welcome submissions on:

  • ·Operational definitions with clear failure boundaries.
  • ·Reproducible triggers or minimal reproductions.
  • ·Comparable evaluation protocols or trace-level diagnostics.
  • ·Mitigation, repair, or intervention strategies with explicit evidence about what improves and what does not.

Well-documented negative results are in scope when the analysis is careful and the lesson transfers.

01

Failure Taxonomies and Mechanisms

Operational definitions, triggering preconditions, minimal reproductions, composable failure primitives, and falsifiable mechanistic hypotheses.

02

Closed-loop Evaluation and Trace Diagnostics

Long-horizon evaluation protocols, interpretable process metrics, counterfactual tests, and logging tools that expose failures beyond terminal success.

03

Training and Systems Interventions

Mitigations, recovery strategies, tool and memory interface improvements, reward and budget design, and repair mechanisms with verifiable trade-offs.

Key Dates

Submission deadline

2026-05-08

AOE · May 8, 2026 (AOE)

Notification date2026-05-15AOE · May 15, 2026 (AOE)

Workshop Schedule (Tentative)

08:00 - 08:10

Opening remarks

08:10 - 08:50

Keynote 1

TBD

08:50 - 09:30

Keynote 2

TBD

09:30 - 10:10

Contributed spotlights

Five to eight short talks from accepted contributions

10:10 - 10:40

Coffee break

10:40 - 11:20

Keynote 3

TBD

11:20 - 12:00

Keynote 4

TBD

12:00 - 14:00

Lunch and posters

14:00 - 14:40

Keynote 5

TBD

14:40 - 15:20

Keynote 6

TBD

15:20 - 15:50

Coffee break

15:50 - 16:50

Panel discussion

Panelists TBD

16:50 - 17:00

Closing remarks and awards

Organizing Committee

Zihan Wang

Zihan Wang

Northwestern University

Canyu Chen

Canyu Chen

Northwestern University

David Acuna

David Acuna

NVIDIA Research

Jaehun Jung

Jaehun Jung

NVIDIA Research and University of Washington

Niloofar Mireshghallah

Niloofar Mireshghallah

humans& and Carnegie Mellon University

Yejin Choi

Yejin Choi

Stanford University

Dawn Song

Dawn Song

University of California, Berkeley

Manling Li

Manling Li

Northwestern University and Amazon Scholar

Sponsors

We sincerely thank the support from our sponsors:

O2 logoAbaka logoEight Sleep logo

Contact

Reach out for submissions, sponsorship, speaker logistics, or collaboration.