Online · Pacific Time (PT)

Agentic AI Frontier Seminar

A seminar series on Agentic AI: models, tools, memory, multi-agent systems, online learning, and safety, featuring leading researchers and industry experts.

Join us! Register here to get seminar updates.

Recordings (with speaker consent) will be posted on our YouTube channel.

Incoming Seminar

Online · 2026-02-06 · 09:00–10:00 PT

Talk Title: Enabling Human-centric and Culturally Aware Safety of AI Agents

Assistant Professor · Maarten Sap · CMU

Talk Abstract: AI safety has made substantial strides, yet still struggles to keep up with increasingly agentic AI use cases, and often overly focuses on technical solutions rather than human centered ones. In this talk, I'll outline some recent works towards making AI safety more human-centric and culturally aware. First, I'll introduce HAICosystem and OpenAgentSafety, two new interactive benchmarks for evaluating LLM agents in multi-turn and tool-using interactions via simulations, which shows that agents still have safety issues due to tool use that were not previously known. Then, focusing on users, I'll outline a recent study on how LLM agents should or should not refuse queries, showing that user perceptions, trust, and willingness to use LLMs are strongly affected by their refusal strategies, and that many current LLMs use least-preferred refusal strategies. Finally, I'll cover an oft-overlooked aspect of safety, namely, cultural safety. Introducing MC-Signs, a new benchmark to measure the cultural safety of LLMs, VLMs, and T2I systems with respect to culturally offensive non-verbal communication (e.g., hand gestures), showing strong western-centric biases of all AI systems. I'll conclude with some future directions towards better cultural and human-centric safety.

Bio: Maarten Sap is an assistant professor in Carnegie Mellon University's Language Technologies Department (CMU LTI), and a courtesy appointment in the Human-Computer Interaction institute (HCII). He is also a part-time research scientist and AI safety lead at the Allen Institute for AI. His research focuses on (1) measuring and improving AI systems' social and interactional intelligence, (2) assessing and combatting social inequality, safety risks, and socio-cultural biases in human- or AI-generated language, and (3) building narrative language technologies for prosocial outcomes. He has presented his work in top-tier NLP and AI conferences, receiving paper awards or nominations at NeurIPS 2025, NAACL 2025, EMNLP 2023, ACL 2023, FAccT 2023, WeCNLP 2020, and ACL 2019. He was named a 2025 Packard Fellow and a recipient of the 2025 Okawa Research Award. His research has been covered in the press, including the New York Times, Forbes, Fortune, Vox, and more.

Focus Areas

Foundation Models & Core Capabilities

Agent Infrastructure & Tooling

Learning, Adaptation & Feedback

Multi-Agent Systems & Social Intelligence

Evaluation, Safety & Alignment

Applications & Vertical Use Cases

Interface & Interaction Design

Governance, Ethics & Ecosystem Building

Organizing Committee

Photo of Ming Jin

Ming Jin

Virginia Tech

He is an assistant professor in the Bradley Department of Electrical and Computer Engineering at Virginia Tech. He works on trustworthy AI, safe reinforcement learning, foundation models, with applications for cybersecurity, power systems, recommender systems, and CPS.

Photo of Shangding Gu

Shangding Gu

UC Berkeley

He is a postdoctoral researcher in the Department of EECS at UC Berkeley. He works on AI safety, reinforcement learning, and robot learning.

Photo of Yali Du

Yali Du

KCL

She is an associate professor in AI at King’s College London. She works on reinforcement learning and multi-agent cooperation, with topics such as generalization, zero-shot coordination, evaluation of human and AI players, and social agency (e.g., human-involved learning, safety, and ethics).