Principled Design for Trustworthy AI - Interpretability, Robustness, and Safety across Modalities

ICLR 2026 Workshop

๐Ÿ“ International Conference on Learning Representations (ICLR 2026)
๐Ÿ“… Date: Sunday April 26 or Monday April 27 ยท ๐ŸŒ Location: Rio de Janeiro, Brazil


Overview

Modern AI systems, particularly large language models, vision-language models, and deep vision networks, are increasingly deployed in high-stakes settings such as healthcare, autonomous driving, and legal decisions. Yet, their lack of transparency, fragility to distributional shifts between train/test environments, and representation misalignment in emerging tasks and data/feature modalities raise serious concerns about their trustworthiness.

This workshop focuses on developing trustworthy AI systems by principled design: models that are interpretable, robust, and aligned across the full lifecycle โ€“ from training and evaluation to inference-time behavior and deployment. We aim to unify efforts across modalities (language, vision, audio, and time series) and across technical areas of trustworthiness spanning interpretability, robustness, uncertainty, and safety.


Call for Papers

We invite submissions on topics including (but not limited to):

Reviews are double-blind and the accepted papers are non-archival. Accepted papers will be presented as posters and/or short talks.

Submission Instruction

Note that for Openreivew submission, new profiles created without an institutional email will go through a moderation process that can take up to two weeks. New profiles created with an institutional email will be activated automatically.


Important Dates

Event Date
Submission deadline Feb 2, 2026
Notification to authors Feb 28, 2026
Camera-ready deadline Mar 6, 2026
Workshop date April 26 or 27, 2026

(All deadlines are AoE.)


Invited Speakers

Yan Liu
Yan Liu
USC, Full Professor
 Mihaela van der Schaar
Mihaela van der Schaar
U Cambridge, Full Professor
Nanyun (Violet) Peng
Nanyun (Violet) Peng
UCLA, Associate Professor
Hamed Hassani
Hamed Hassani
UPenn, Associate Professor
Martin Wattenberg
Martin Wattenberg
Harvard, Professor
Fernanda Viegas
Fernanda Viegas
Harvard, Professor

Organizers

Lily Weng
Lily Weng
UC San Diego
Nghia Hoang
Nghia Hoang
Washington State U
Tengfei Ma
Tengfei Ma
Stony Brook U
Jake Snell
Jake Snell
Princeton
Francesco Croce
Francesco Croce
Aalto U
Chandan Singh
Chandan Singh
Microsoft Research
Subarna Tripathi
Subarna Tripathi
Intel
Lam Nguyen
Lam Nguyen
IBM Research

Contact

๐Ÿ“ง Lily Weng (lweng@ucsd.edu), Nghia Hoang (trongnghia.hoang@wsu.edu)