Important Dates

Timeline (Tentative)

Workshop Schedule
Event Start time End time
Opening Remarks 9:00 9:15
Invited Talk #1: Prof. Bo Li 9:15 9:45
Invited Talk #2: Prof. Chaowei Xiao 9:45 10:15
Contributed Talk #1 10:15 10:30
Coffee Break 10:30 10:45
Invited Talk #3: Prof. Ziwei Liu 10:45 11:15
Invited talk #4: Prof. Florian Tramèr 11:15 11:45
Contributed Talk #2 11:45 12:00
Lunch (12:00-13:30)
Invited Talk #5: Dr. Nouha Dziri 13:30 14:00
Invited Talk #6: Prof. Yaodong Yang 14:00 14:30
Invited Talk #7: Prof. Aditi Raghunathan 14:30 15:00
Poster Session 15:00 16:00
Challenge Session 16:00 16:30
Poster Session #2 16:30 17:00

Call for Papers

Vision-language agents, embodied or digital systems powered by multimodal foundation models, are rapidly emerging as a central paradigm for intelligent perception, decision-making, and human-AI interaction. These agents integrate perception (vision), cognition (language and reasoning), and action (planning and control) within a unified framework, thereby bridging the gap between visual recognition and autonomous behavior. However, beyond traditional pixel-level perturbations, new attack surfaces arise from adversarial prompts, instruction injections, and jailbreak manipulations, which can disrupt reasoning chains, mislead perception, or induce harmful actions. To foster the development of safe, robust, and trustworthy vision-language agents for real-world applications, we invite submissions on both theoretical and practical aspects of adversarial machine learning, with a specific focus on the safety of vision-language agents. We welcome research contributions related to the following (but not limited to) topics:
  • Attack and defense on vision-language agents
  • Datasets and benchmarks that could evaluate vision-language agents
  • Adversarial / Jailbreak attacks on vision-language agents
  • Improving the robustness of agents or deep learning systems
  • Interpreting and understanding model robustness, especially agentic AI
  • Adversarial attacks for social good
  • Alignment of vision-language agents
Format: Submissions papers (.pdf format) must use the CVPR 2026 Author Kit for LaTeX/Word Zip file and be anonymized and follow CVPR 2026 author instructions. The workshop considers two types of submissions: (1) Long Paper: Papers are limited to 8 pages excluding references; (2) Extended Abstract: Papers are limited to 4 pages including references. Accepted papers have the option to be included in the CVF and IEEE Xplore Proceedings.

Submission Site: https://openreview.net/group?id=thecvf.com/CVPR/2026/Workshop/Advml
Submission Due (both Paper and Supplementary Material): March 5, 2026, 11:59 PM (UTC±0)


Sponsors

logo-img
logo-img
logo-img
logo-img


logo-img