News & Posts

Publication Report 2025 - Eye Tracking and Multimodal Analysis

General Trend
Eye tracking is increasingly used as a functional component within intelligent systems. In vehicles, robots, and interactive environments, gaze data is combined with additional signals—such as motion, context, and physiology—to support models of attention, intention, and risk. Rather than serving solely as a measurement tool, eye tracking is becoming part of operational perception pipelines in real-world systems.

In 2025, more than 50 peer-reviewed studies were published by research laboratories worldwide using Ergoneers technology. These publications span transportation, human–machine interaction, perception research, and multimodal AI. This report reviews the year’s output and highlights five studies that exemplify current methodological and applied trends

Here are the 5 publications that stood out to us and why we believe they deserve a closer look:

1. A Foundational for next-generation in-car AI and human–robot collaboration.

Enhancing Gaze Prediction in Multi-Party Conversations via Speaker-Aware Multimodal Adaptation by MC Lee, Z Deng

ACM CHI 2024/25 – Multimodal Attention & Real-World Interaction


Why this study stands out:
This paper represents the state of the art in multimodal human sensing—combining eye tracking with other signals (e.g., head pose, motion, context, sometimes physiology) to infer user intent and cognitive state in naturalistic environments.

Impact trajectory:

  • Directly applicable to driver monitoring systems (DMS) in vehicles.
  • Enables AI models to reason about attention, intention, and risk in real time.
  • Sets the methodological standard for fusing gaze with other streams—exactly what future autonomous and semi-autonomous systems require.

External Link: https://dl.acm.org/doi/abs/10.1145/3716553.3750755

 

Multi Modal study using motion, eye tracking, video observation and skin viscousity
Source: https://dl.acm.org/doi/full/10.1145/3716553.3750755

2. A tight alignment with latest automotive OEM needs and regulatory safety frameworks.

Collision Risk Perception Models Using Physiological and Eye-Tracking Signals by H. Lee, O. Lim, A. Singh and S. Samuel

IEEE – Eye Tracking in Applied Transport / Human–Machine Contexts

Why it matters:
IEEE transport and HMI papers are often engineering-grade rather than purely theoretical. This work situates gaze in operational environments where vehicles, interfaces, or safety systems rather than constrained lab tasks are tested in a realistic setup.

Impact trajectory:

  • Bridges perception research with deployable systems.
  • Supports real-time gaze inference for:
    • Hazard detection
    • Interface adaptation
    • Driver distraction mitigation

External Link: https://ieeexplore.ieee.org/abstract/document/11106510/

eye tracking and multi modal study for road safety
Source: Collision Risk Perception Models Using Physiological and Eye-Tracking Signals

3. Elevating eye tracking from a measurement tool into a cognitive signal for intelligent systems.

Driving Simulator Evaluation of Long Persistent Self-Luminous Pavement Markings’ Visual Guidance By Xiaolong YangChuanqi XianXiaowei FengYingting Cao, Chunhong PengXiang Liu

Springer 2025 – Visual Attention in Real-World Perception

Why it matters:
This is part of a new generation of perception research focused on ecological validity: how vision and gaze operate in the wild rather than in screen-based experiments.

Impact trajectory:

  • Advances models of how humans allocate attention in dynamic environments.
  • Provides theoretical grounding for:
    • Predictive driver models
    • Human-aware robotics
    • AI systems that infer “what matters now’
    • Validates evaluation of new road marking systems in the field

External Link: https://link.springer.com/article/10.1007/s42947-025-00679-1

 

Source: https://link.springer.com/

4. Engineering a backbone for real-world multimodal AI.

CogMamba: Multi-Task Driver Cognitive Load and Physiological Non-Contact Estimation with Multimodal Facial Features by Yicheng Xie & Bin Guo

Why it matters:
This work emphasizes hardware integration, synchronization, and real-time pipelines.

Impact trajectory:

  • Demonstrates how gaze can be fused with:
    • IMUs
    • Cameras
    • Physiological sensors
  • Critical for:
    • In-vehicle monitoring
    • Wearable safety systems
    • Embodied AI and robotics

External Link: https://www.mdpi.com/1424-8220/25/18/5620

CogMamba: Internal structure of three types of feature embedding.

5. Connecting eye tracking to population-level safety outcomes.

Inflating system expectations prior to SAE level 3 automated vehicle use: effects on monitoring behavior, resumption of control, and attitudes toward driving automation by Dustin J. Souders, Shubham Agrawal, Irina Benedyk, Yuntao Guo, Yujie Li, Srinivas Peeta 

Transportation Research (TRIP, 2025) (doi: 10.1016/j.trip.2025.101725)

Why it matters:
This study is explicitly rooted in transport safety. TRIP papers (Transportation Research Interdisciplinary Perspectives) influence policy, infrastructure design, and industry standards.

Impact trajectory:

  • The Study uses gaze and behavioral measures to understand:
    • Distraction
    • Risk perception
    • Decision latency in traffic contexts
  • Feeds directly into:
    • Road-safety interventions
    • Driver assistance systems
    • Latest urban mobility design

External Link: https://www.sciencedirect.com/science/article/pii/S259019822500404X

eye tracking in driving simulator
Source: https://www.sciencedirect.com/science/article/pii/S259019822500404X#f0005

Summary

These five resources are particularly noteworthy for their practical and forward-looking applications. They operate in open, dynamic real-world environments; integrate gaze within multimodal cognitive frameworks; are designed for seamless AI- deployment, and offer direct insights for advancing automotive, transport, and robotic technologies.

Seen from a wider angle, the most significant study was Enhancing Gaze Prediction in Multi-Party Conversations via Speaker-Aware Multimodal Adaptation by Meng-Chen LeeZhigang Deng

For us, this study underlines the turning point in how eye tracking is understood. Today gaze tracking is fused with body motion, head pose, and context to form a living model of intent. The system does not ask what is visible, but what matters now. This mirrors how humans move: sparsely, purposefully, and in time. Attention becomes event-based rather than continuous. Perception becomes prediction rather than reaction. Vision becomes a control signal rather than an input stream. The paper shows how machines can learn when to look, not only where. It transforms gaze into a computational resource. This enables AI systems that infer risk, readiness, and intention. For vehicles, robots, and embodied agents, this is foundational. It is the moment where “quiet eye” becomes an algorithm. Human expertise is no longer studied—it is used as a blueprint to engineer. This is where we are headed with Prophea.X. Our new research suite, which naturally supports multi-modal data acquisition and inbound event analysis in one cohesive framework—saving hours of time, diminishing signal loss, while increasing precision. Overall we aim to empower researchers to be able to focus on their creative tasks rather than fiddling between codecs and formats.

Congratulations to Meng-Chen Lee, Zhigang Deng and all other researchers in the field for their outstanding publications which yearly contribute to a more human-centred future in technology and life science. 

Access all studies published in 2025 & head over to the Ergoneers Publication Hub !

Human behavior analysis is extremely complex. looking at a cloud diagram of data streams, Prophea.X simplifies data handling