Eye movements in social interaction
In our lab we investigate eye movements during dyadic interaction. The setup for this project was built by Tim Cornelissen
. A short example can be seen below.
In order to map eye-tracking data to the face in the video, we've developed a semi-automatic Area of Interest method based on Voronoi tessellation:
Fixation classification in low quality eye-tracking data
I've been involved in many eye-tracking studies with infants and young children as the participant group. While eye tracking can provide valuable insights into (cognitive) development, eye-tracking data obtained from infants and children are generally of lower quality as compared with eye-tracking data from adults. This is in part due to the fact that infants are difficult to restrain in their movement. I've been involved in two eye-tracking tests in which we compared eye trackers on their robustness to movement (view the first
and the second
here). Moreover, I've developed a fixation-classification algorithm that is specifically built for eye-tracking data of low quality. The software is freely available from GitHub