Shiyan Yang Profile

Yang, Shiyan 杨世炎

Scientist, Melbourne, Australia

About Me

The goal of my career is to empower machines and AI to better understand and support humans. Specifically, my research focuses on integrating human factors—behavioral, cognitive, and physiological—into machine learning and AI systems to enhance user safety, performance, and overall experience.

This vision inspired me to join Seeing Machines (2017-2024), where I contributed to the innovation of driver monitoring systems, a groundbreaking technology at the time aimed at saving lives on the road. By combining human factors expertise with machine learning, I advanced the detection of driver distraction, drowsiness, and the transition of control in complex driving scenarios.

My research has been instrumental in shaping the technical roadmaps, algorithm development, and industry guidelines for automotive-grade driver monitoring systems. It has also garnered recognition through publications in top-tier journals and prestigious international awards, such as Patricia F. Waller Award.

My Story of Integrating Human Factors into Machine Learning for Driver State Monitoring

Over the past decade, driver monitoring systems have evolved from lab prototypes to automotive-grade products that save lives on the road. I was fortunate to contribute to this breakthrough, integrating human factors into machine learning for real-world impact.

Texas A&M University (2010-2016): Where My Journey Began

My journey started with my doctoral research on human factors and cognitive systems, where I explored novel metrics to enhance human performance in complex environments. Under Thomas Ferris’ supervision, I proposed “cognitive efficiency” - a metric quantifying how much information a person processes per “unit” of cognitive resources. A core chapter of my thesis involved measuring cognitive workload using physiological sensors. To deepen this research, I took graduate courses across multiple areas - computer science, statistics, psychology, and kinesiology - which fueled my curiosity about interdisciplinary connections. In my final year, a paper from Birsen Donmez’s lab, “Smart driver monitoring: When signal processing meets human factors”, opened my eyes to an exciting future: the intersection of human factors and machine learning to create intelligent human sensing.

UC Berkeley (2016-2017): A Brief but Pivotal Chapter

As a postdoctoral scholar, I collaborated with Steven Shladover, a pioneer of intelligent transportation systems, to study driver-automation interaction in truck platooning. It was an unforgettable experience of operating a three-truck platooning on Bay Area freeways. But Berkeley offered even more. During the kickoff meeting of the NSF VeHiCal project, I met Ruzena Bajcsy, a renowned Computer Science professor, who emphasized the importance of human factors for human sensing. Surrounded by several high-profile Berkeley professors, I was, however, too nervous to contribute - yet that moment profoundly shaped my vision of connecting human factors with machine learning.

By 2017, human factors was mostly framed as UI/UX in Silicon Valley, which didn’t fully aligned with my goal. Then, an unexpected opportunity arose on the other side of the world - Seeing Machines, an Australian company, was transitioning from eye tracking to driver monitoring. It needed human factors experts to help computer scientists understand driver state and behavior. After interviewing with Mike Lenné, their Human Factors Lead, and consulting Steve and Tom, I took a leap of faith - moving to Australia to embrace a new adventure.

Seeing Machines (2017-2024): Where Research Met Real-World Impact

Joining Seeing Machines in its ascent, I led the CANdrive Automated vehicle trial (AU$1.35 millions) and later the collaboration with the MIT Advanced Vehicle Technology consortium (founded by Bryan Reimer). I initiated cutting-edge research on distraction, drowsiness, takeover readiness, and gaze-context analysis. These research demonstrated the unique value of human factors in defining ground truth, curating high-quality data, and enhancing feature engineering for driver state detection. Moreover, I explored computation models - from probabilistic machine learning to deep learning to generative AI - to analyze driver behavior and dynamic contexts. The findings offered novel and holistic insights into driver-automation-environment interactions, shaping the future of driver state monitoring.

The Next Chapter

Every journey has its end. At the end of 2024, my time at Seeing Machines came to an end due to the volatility of the automative industry. I knew this also meant the end of my career as an automotive human factors researcher - Australia doesn’t build cars. However, this also means a new start, towards a broader future of empowering AI to support human tasks. I hope my story will inspire others who are passionated about human factors and AI for real-world influence.

Research Projects and Publications

2017-2024, Seeing Machines

  • MIT Advanced Vehicle Technology Consortium

    Yang, S., McKerral, A., Mulhall, M. D., Lenné, M. G., Reimer, B., & Gershon, P. (2023). Takeover Context Matters: Characterising Context of Takeovers in Naturalistic Driving using Super Cruise and Autopilot. In Proceedings of the 15th International Conference on Automotive User Interfaces and Interactive Vehicular Applications, 112-122.

    Yang, S., Lenné, M.G., Reimer, B., Gershon, P. (2022). Modeling Driver-Automation Interaction using A Naturalistic Multimodal Driving Dataset. In Proceedings of the 66th Human Factors & Ergonomics Society International Annual Meeting,66(1), 1462-1466.

  • Advanced Safe Truck Concept (ASTC)

    Yang, S., Kuo, J., Lenné, M. G., Fitzharris, M., Horberry, T., Blay, K., … and Truche, C. (2021). The impacts of temporal variation and individual differences in driver cognitive workload on ECG-based detection. Human Factors, 63(5), 2021.

  • CANdrive Automated Vehicle Trial

    Yang, S., Wilson, K.M., Shiferaw, B., Trey, R., Kuo, J., and Lenné, M.G. (2024). Sensor fusion to connect gaze fixation with dynamic driving context for driver attention management. Transportation Research Part F, (107)578-588.

    Yang, S., Wilson, K.M., Trey, R., Kuo, J., and Lenné, M.G. (2022). Beyond gaze fixation: Modeling peripheral vision in relation to speed, Tesla Autopilot, cognitive load, and age in highway driving. Accident Analysis & Prevention, 177, 106670.

    Yang, S., Wilson, K.M., Trey, R., Kuo, J., and Lenné, M.G. (2022). Evaluating driver features for cognitive distraction detection and validation in manual and Level 2 automated driving. Human Factors, 64(4), 746-759.

    Yang, S., Kuo, J., and Lenné, M.G. (2021). Effects of distraction in on-road level-2 automated driving: Impacts on glance behavior and take-over performance. Human Factors, 63(8), 1485-1497.

    Wilson, K. M., Yang, S., Roady, T., Kuo, J., and Lenné, M. G. (2020). Driver trust and mode confusion in an on-road study of automated vehicle technology. Safety Science, 130, 104845.

2016-2017, PATH, The University of California, Berkeley

  • Connected and Automated Truck Platooning

    Yang, S., Shladover, S.E., Lu, X., Ramezani, H., Kailas, A., and Altan, O.D. (2021). A Bayesian regression analysis of truck drivers' use of cooperative adaptive cruise control (CACC) for platooning on California highways. Journal of Intelligent Transportation Systems, 1-12.

    Yang, S., Shladover, S.E., Lu, X., Ramezani, H., Kailas, A., and Altan, O.D. (2018). A first investigation of truck drivers' preferences and behaviors using a prototype cooperative adaptive cruise control system. Transportation Research Record, 2672(2), 39-48.

2010-2016, Industrial Engineering, Texas A&M University

  • Multimodal Display Design

    Yang, S., & Ferris, T. K. (2019). Supporting Multitracking Performance with Novel Visual, Auditory, and Tactile Displays. IEEE Transactions on Human-Machine Systems, 50(1), 79-88.

  • Cognitive Efficiency in Human-Machine Systems

    Yang, S., & Ferris, T. K. (2018). Cognitive efficiency in human-machine systems: Metrics of display effectiveness for supporting multitask performance. Journal of Cognitive Engineering and Decision Making, 12(2), 153-169.

Awards

  • Best Papaer Award, by the Road User Measurement and Evaluation Committee (ACH60), Transportation Research Board (TRB), 2023
  • Best Extended Abstract/Paper, Australasian Road Safety Conference, 2022
  • Patricia F. Waller Award, TRB, 2018
  • Deborah Freund ACS60 Paper Award, by TRB Truck and Bus Safety Committee, 2018
  • North American Automotive Innovation & Startup Competition Award, NACSAE, 2016
  • Alphonse Chapanis Best Student Paper Award (Finalist, top 3 of 37), the Human Factors and Ergonomics Society, 2014

Education

  • PhD, Texas A&M University, 2016