May 2019 - August 2020

A system with a DIY head-mounted webcam which captures the eye and is able to classify 1) corneal surface reflection to know what the person is looking at and 2) physiological parameters that contain information on cognitive load. The goal was to build an open-sourced system with a DIY guide to replicate the low-cost hardware with an easy to use software toolkit for human factors researchers, especially for those researchers without an engineering background, to detect cognitive load and screen time across devices.

I helped build the software toolkit for EyeKnowYou and assisted in the experiment design, data collection, and manuscript preparation. This project was led by Tharindu Kaluarachchi at the Augmented Human Lab. The research was published as an extended abstract in Mobile HCI 2021.

Tools: Java, Javascript