Yue Jiang, University of Utah
The Synthetic User: Simulating Viewing Behaviors through Multimodal AI.
Understanding how users visually process information is critical for effective interface design, yet gathering eye-tracking data remains resource-intensive. The rise of multimodal generative AI presents a compelling alternative: the creation of “synthetic users” that can simulate human viewing behaviors with high fidelity. This talk shows the capabilities of state-of-the-art generative models to predict visual attention on Graphical User Interfaces (GUIs). Unlike traditional saliency maps, these multimodal agents account for specific user goals and GUI layouts, generating sequential fixation patterns that mirror human viewing strategies. The talk will cover the technical nuances of training these models on large-scale eye-tracking datasets, the evaluation metrics required to assess generative scanpaths, and the current limitations regarding individual differences.
Yue Jiang is an Assistant Professor at the University of Utah, where she leads research at the intersection of Human-Computer Interaction and Computer Vision. Her work focuses on Computational User Interface Understanding, with specific expertise in applying multimodal generative AI to model human viewing behavior and generate adaptive, accessible UIs. Yue’s research has been published in top venues including CHI, UIST, and CVPR. She initiated and served as the lead organizer of five Computational UI workshops at CHI (2022–2026) and co-organized the Mobile Eye-based Interaction Workshop at ETRA. She was the Accessibility Chair for CHI 2023 and CHI 2024 and will be the Workshop Chair for IUI 2027. Her contributions have been recognized with the Meta Research PhD Fellowship, the Google Europe Scholarship, and the Nokia Scholarship. She was also selected as a Heidelberg Laureate Forum Young Researcher and a UCSD Data Science Rising Star. Previously, Yue has worked or interned at Intel, Adobe Research, Apple, the Max Planck Institute, and Carnegie Mellon University.
Zoya Bylinskii, Adobe Firefly
Perceptual Insights for GenAI Evaluation
In just a few years, visual generative AI models have advanced from struggling to produce the right body morphology to generating portraits indistinguishable from professional headshots. Contributing to this rapid progress are advances in model evaluation—the large-scale measurement of model quality, realism, and harm. Evaluation methods draw from both crowdsourced annotation and perceptual science toolboxes: evaluators may be asked to detect visual artifacts, assess realism, classify content, compare sets, or assign quality ratings. In perceptual science terms, they are effectively performing visual search, categorization, ensemble and memory tasks among others. In this talk, I’ll share insights from developing the evaluation strategy for Adobe Firefly and explore how evaluation methods have been evolving alongside model capabilities. I’ll argue that the deep body of knowledge on human perception and experimental design remains underutilized—and that grounding GenAI evaluation in perceptual science can yield richer, more reliable measures of model quality and usefulness.
Zoya Bylinskii leads the Scientific Evaluation Team for Adobe Firefly's suite of generative imaging and video models. Previously, she was a Senior Research Scientist in the Imaging & Languages Lab at Adobe Research. She was an Executive Board Member of The Readability Consortium, an affiliate at MIT and the Harvard Institute of Applied Computational Science. Zoya received her Computer Science Ph.D. from MIT in 2018, and an Hon. B.Sc. in Computer Science and Statistics from the University of Toronto in 2012. Zoya was a 2023 Adobe Founder's Award Recipient, 2018 EECS Rising Star, 2016 Adobe Research Fellow, 2014-2016 NSERC Postgraduate Scholar, 2013 Julie Payette Research Scholar, and 2011 Anita Borg Scholar. Zoya's expertise lies at the interface of human perception & cognition, machine learning, and human-computer interaction - with applications to photography, graphic design, readability, and generative AI.
Vaibhav Unhelkar, Rice University
Training Humans to Work with Robots
We are steadily approaching a future where humans work with robotic assistants, teammates, and even tutors. While significant effort is being dedicated to training robots to work with humans, much less attention has been given to the human side of this partnership. For instance, we see increasingly capable AI models and algorithms that enable robots to infer human intent, plan interactions, and communicate with them. However, human users often lack understanding of how robots will react in novel situations and receive minimal training to work with robots. To ensure the safe and effective use of robots, I argue that training human users is equally critical. This talk will present human-centered computing methods for generating and delivering such training. First, I will discuss Explainable AI methods that can help generate the training content, enabling humans to develop a Theory of Mind regarding the robot. Next, I will discuss mechanisms for delivering this training interactively using physiological sensors (such as eye trackers) and interactive simulation. The talk will conclude with an overview of open problems in making this training more effective through personalization.
Vaibhav Unhelkar is an Assistant Professor at Rice University, where he leads the Human-Centered AI & Robotics group. His work spans the development of robotic assistants, intelligent tutors, and decision-support systems aimed at enhancing human performance in domains ranging from healthcare to disaster response. Underpinning these systems are Unhelkar's contributions to imitation learning and explainable AI, designed to train robots, humans, and human-AI teams. He earned his Ph.D. in Autonomous Systems from MIT in 2020 and completed his undergraduate studies at IIT Bombay in 2012. Unhelkar is actively engaged in the AI and robotics research communities, serving on the editorial board of robotics and AI conferences. His research and service have been recognized with awards from the International Foundation for Autonomous Agents and Multi-Agent Systems (AAMAS), among others.