Top 5 Eye Tracking Fails

I've worked with a lot of eye tracking data, and have reviewed a lot of academic papers that use eye-tracking.

This method is fantastic - I'm not exaggerating when I say that without it, we wouldn't know half the things we know about infant cognition. It also is a great tool for studying non-verbal populations.

BUT, I've learned a few lessons over the years, some of them the hard way. I thought I'd share them to help others who use, or are thinking of using, eye tracking.

I present: Top 5 Eye Tracking Fails and How to Avoid Them .

1. You haven't assessed the quality of your data.

This is especially important when you are working with a population with noisier data (I'm lookin at you, babies), but is critical no matter what. Without knowing whether the data are high quality, you don't know if you can trust it.

Being able to trust your data is critical, and everything on this list is really just a variant of that.

Here are some things I always, always (always) investigate:

  • What is the % of samples with high-quality data? Most eye-trackers provide an estimate of data quality for each sample. I make sure that most of my samples have accurate recordings. If a trial has too many samples with poor quality, I'll chuck it.

  • How noisy are my data?A quick and dirty way to do this is to calculate the root-mean-square-error of each fixation. Ideally, you don't want this to be high - if it is, it means the “best guess” of where the infant is looking is not smooth, and looks really “jumpy” which is no bueno.

  • How much data do I even have? It's surprisingly easy to not realize that you are making inferences from trials with 90% missing data. Make sure you know how much missing data you have.

Which leads me to my next point ….

2. You haven't established how much data is enough.

This one is pretty common. Missing data is to be expected. With babies, you're going to have a lot of missing data. But it's critical that you decide ahead of time how much missing data is too much missing data.

Why decide ahead of time? To avoid the temptation to lower/raise your threshold for what counts as too much data to get a statistically significant effect. No bueno.

We're trying to make inferences about a person's mental state based on where they are looking on a computer screen. Don't make this even more sketchy by making this inference based on 2 seconds of data.

3. Calibration drift!

Before collecting eye tracking data, you need to “calibrate” a person to the eye-tracker. This is Step 0. Bad calibration=bad data.

Lots of studies make sure that the calibration is good at the beginning, press “go” and move on.
Sadly, calibration drift is a thing.
I highly recommend that you validate the calibration throughout the experiment. Yep - just once isn't good enough. Again, this is a bigger issue in kids than adults but is an issue for everyone.

4. Waning attention.

Man, some eye tracking tasks are BORING. If you pay an undergrad $10 to stare at a black and white display of faces for 45 minutes, they're going to zone out. Always check that your effects aren't mainly driven by time (e.g. increasing boredom).

Imagine looking at these bad boys for 45 minutes.

5. Quality is correlated with your outcome or varies by group.

Let's say you're comparing 5-year-olds to 10-year-olds, and find that 10-year-olds are way better at a visual search task. Time to publish, right? Not so fast! You need to make sure that your results aren't driven by differences in data quality. Otherwise, you think you're seeing a cool developmental trend, but you're really just looking at how eye tracking data quality improves with age.

Sometimes difference in data quality between groups is unavoidable - and that's actually ok! Just include quality covariates in your statistical model so that you're controlling for them.

That's all for now. Leave a comment if I've overlooked something critical!

comments powered by Disqus

Related