The eyetools package is designed to provide a consistent set of functions that helps researchers perform straightforward, yet powerful, steps in the analysis of eye-data. The suite of functions will allow the user to go from relatively unprocessed data to trial level summaries that are ready for statistical analysis.
You can install eyetools using the following code:
A quick note before starting: the following tutorial is written using only the dependencies contained within eyetools, so that means that if you are planning on following this tutorial through, you only need to install eyetools.
Eyetools has built in example data that can be used while getting to grips with the package. For the purpose of this tutorial, we will start with a small dataset that contains binocular eye data from two participants (and six trials) from a simple contingency learning task (the data are from Beesley, Nguyen, Pearson, & Le Pelley, 2015)1. In this task there are two stimuli that appear simultaneously on each trial (to the left and right in the top half of the screen). Participants look at these cues and then make a decision by selecting an “outcome response” button.
Let’s load in this binocular data and explore the format. As we can see, the dataset is formed of 31,041 rows and 7 variables
## [1] 31041     7To get a basic idea of the contents of the data, we can look at the first 10 observations in the data
We can see that our seven variables contain the participant
identifier (pID), the timepoint of the trial
(time), the left eye x and y coordinates
(left_x and left_y), the right eye coordinates
(right_x and right_y), as well as a trial
identifier (trial).
By default eyetools assumes that the resolution of eye data is in pixels (and a default screen size of 1920x1080), however the functions should work with any units (with no guarantee as it is not tested with units other than pixels).
## # A tibble: 10 × 7
##    pID    time left_x left_y right_x right_y trial
##    <chr> <dbl>  <dbl>  <dbl>   <dbl>   <dbl> <dbl>
##  1 118       0   909.   826.   1003.    808.     1
##  2 118       3   912.   829.   1001.    812.     1
##  3 118       7   912.   826.   1010.    813.     1
##  4 118      10   908.   824.   1006.    807.     1
##  5 118      13   912.   824.   1005.    805.     1
##  6 118      17   912.   826.   1000.    802.     1
##  7 118      20   910.   826.    997.    806.     1
##  8 118      23   912.   825.   1005.    806.     1
##  9 118      27   911.   819.   1006.    808.     1
## 10 118      30   914.   821.   1013.    804.     1eyetools functions can accept either single-participant data or
multi-participant data. By default, most functions assume that data is
of a single participant unless a participant identifier
column called participant_col is present,in which case the
data is handled as multi-participant. In cases where participants are
identified with a different variable, functions accept a parameter of
participant_col that takes a character string of the
identifying column. In situations where participant_col is not declared,
and when duplicated non-consecutive trials are detected (that implies
multi-participant data being input as single-participant), the function
will error. This is a good point to then check your data structure,
ensure that it is labelled correctly and ordered by trial and time.
We need to combine the left and right eye x and y coordinates to get
a single pair of [x,y] coordinates for each timepoint. The eyetools
function combine_eyes() can do this for us. The
method parameter gives the option to either “average” the
two eyes, or we can use the “best_eye”. For “average”, the result is
based on the average of the two eyes for each sample, or for samples
where there is data from only a single eye, that eye is used. For
“best_eye”, a summary of the proportion of missing samples is computed,
and the eye with the fewest missing samples is used. Here we use the
default parameter “average”.
combine_eyes() is one of the few core eyetools functions
that doesn’t handle multi-participant/single-participant data
differently. Primarily because of the function’s construction and its
intended useage, it does not need to do so.
The above code returns a flattened list of all participants data, and if we take a look at just one participant, we are returned with a dataframe that has x and y variables in place of the left_* and right_* variables. This is the data format needed by many of the other eyetools functions: time, x, y, and trial. The ordering of the variables should not matter, however most of the functions will impose an ordering in how the information is returned.
##   pID time trial        x        y
## 1 118    0     1 955.8583 816.5646
## 2 118    3     1 956.5178 820.6221
## 3 118    7     1 960.7383 819.7616
## 4 118   10     1 956.9727 815.3331
## 5 118   13     1 958.6214 814.0815
## 6 118   17     1 956.0035 814.1564The next stage of the process would be to remove missing data within
continuous streams of eye data which are likely to be caused by
blinking. We can do this using the interpolate() function.
The maxgap parameter specifies the maximum time frame of consecutive NAs
to fill. Any longer gaps will be left unchanged. This is set as default
to 150, the typical duration of a blink. The method parameter can either
be “approx” (default) or “spline” which are both calls to
zoo::na.approx and zoo::na.spline
respectively.
Note that as the participant identifier column is not “participant_col” (as eyetools expects as default), it needs to be specified in the function call.
You can also request a report of the differences in NA values present before and after the interpolation as well
interpolate_report <- interpolate(data, maxgap = 150, method = "approx", report = TRUE)
interpolate_report[[2]]##   pID missing_perc_before missing_perc_after
## 1 118         0.001157156        0.001157156
## 2 119         0.000000000        0.000000000An additional step that can be beneficial is to pass the eye data
through the smoother() function. This removes particularly
jerky transitions between samples and is critical for the analysis of
velocities in the eye-movements. For now, let’s store the smoothed data
in a new object. We can also ask for a plot of the data so that we can
visually inspect it to see how well it fits the data.
set.seed(0410) #set seed to show same participant and trials in both chunks
data_smooth <- smoother(data,
                        span = .1, # default setting. This controls the degree of smoothing
                        plot = TRUE) # whether to plot or not, FALSE as default## Showing trials: 4, 6 for participant 119The plot above shows the difference between the raw and smoothed data for a randomly selected participant and two random trials (simply for visualisation purposes and to keep the amount of plotting to a minimum).
With the default smoothing setting, we can see that the smoothed data
does not track the actual data as closely as it could. The lower the
value of span, the closer the smoothed data represents the
raw data. A visual inspection of the plot suggests that a span of .02 is
a good value for this data example. It is important that the fixations
and saccades are matched closely to ensure data quality. Oversmooth and
you end up with significant data loss, undersmooth and all the jerky eye
movement is preserved rendering the use of smoother()
meaningless, so a good inspection and testing of values can be
useful.
set.seed(0410) #set seed to show same participant and trials in both chunks
data_smooth <- smoother(data, 
                        span = .02,
                        plot = TRUE)## Showing trials: 4, 6 for participant 119Many psychology experiments will position stimuli on the screen in a counterbalanced fashion. For example, in the example data we are using, there are two stimuli, with one of these appearing on the left and one on the right. In our design, one of the cue stimuli is a “target” and one is a “distractor”, and the experiment counterbalances whether these are positioned on the left or right across trials.
Eyetools has a built in function which allows us to transform the x
(or y) values of the stimuli to take into account a counterbalancing
variable: conditional_transform(). This function currently
allows for a single-dimensional flip across either the horizontal or
vertical midline. It can be used on raw data or fixation data. It
requires the spatial coordinates (x, y) and a specification of the
counterbalancing variable. The result is a normalised set of data, in
which the x (and/or y) position is consistent across counterbalanced
conditions (e.g., in our example, we can transform the data so that the
target cue is always on the left). This transformation is especially
useful for future visualisations and calculation of time on areas of
interest. Note that conditional_transform() is another
function that does not discriminate between multi-participant and
single-participant data and so no participant_col parameter is
required.
The keen-eyed will notice that the present data does not contain a
variable to specify the counterbalanced positions. This is contained in
a separate dataset that holds the behavioural data, including the
response times, the outcome, accuracy, and cue_order which
tells us whether the target cue was on the left (coded as 1) or on the
right (coded as 2).
## # A tibble: 6 × 8
##   pID   trial P_cue NP_cue cue_order correct_out accuracy     RT
##   <chr> <int> <dbl>  <dbl>     <dbl>       <dbl>    <dbl>  <dbl>
## 1 118       1     2      5         2           2        0 13465 
## 2 118       2     2      6         2           2        0  7796.
## 3 118       3     1      6         2           1        1  5267.
## 4 118       4     1      5         2           1        1  9911.
## 5 118       5     2      5         2           2        1  4424.
## 6 118       6     2      6         2           2        1  5224.First we need to combine the two datasets based upon the participant
identifier. Once the data has been joined we can use
conditional_transform() to transform the x coordinates
across the midline.
data <- merge(data_smooth, data_behavioural) # merges with the common variables pNum and trial
data <- conditional_transform(data, 
                              flip = "x", #flip across x midline
                              cond_column = "cue_order", #this column holds the counterbalance information
                              cond_values = "2",#which values in cond_column to flip
                              message = FALSE) #suppress message that would repeat "Flipping across x midline" In this next stage, we can start to explore the functions available
for determining fixations within the data. The two main functions here
are fixation_dispersion() and fixation_VTI().
Alongside these, is the option to compare_algorithms()
which produces a small number of metrics and plots to help visualise the
two fixation algorithms. We will first demonstrate and explain the two
algorithms before demonstrating compare_algorithms() as
this relies on the two fixation algorithms.
fixation_dispersion() detects fixations by assessing the
dispersion of the eye position using a method similar to that
proposed by Salvucci and Goldberg (1996)2. This evaluates the
maximum dispersion (distance) between x/y coordinates across a window of
data, and looks for sufficient periods in which this maximum dispersion
is below the specified dispersion tolerance. NAs are considered breaks
in the data and are not permitted within a valid fixation period. By
default, it runs the interpolation algorithm and this can be switched
off using the relevant parameter.
data_fixations_disp <- fixation_dispersion(data,
                                           min_dur = 150, # Minimum duration (in milliseconds) of period over which fixations are assessed
                                           disp_tol = 100, # Maximum tolerance (in pixels) for the dispersion of values allowed over fixation period
                                           NA_tol = 0.25, # the proportion of NAs tolerated within any window of samples evaluated as a fixation
                                           progress = FALSE) # whether to display a progress bar or notThe resultant data output from fixation_dispersion()
presents data by trial and fixation. It gives the start and end time for
these fixations along with their duration and the x,y coordinates for
the entry of the fixation.
##   pID trial fix_n start  end duration    x   y prop_NA min_dur disp_tol
## 1 118     1     1     0  173      173  959 810       0     150      100
## 2 118     1     2   197  393      196  961 592       0     150      100
## 3 118     1     3   397  643      246  958 492       0     150      100
## 4 118     1     4   806 1063      257 1370 839       0     150      100
## 5 118     1     5  1236 1386      150  997 545       0     150      100
## 6 118     1     6  1390 1700      310  969 477       0     150      100The fixation_VTI() function operates differently to
fixation_dispersion(). It determines fixations by assessing
the velocity of eye-movements, using a method that is similar
to that proposed by Salvucci & Goldberg (1996). This applies the
algorithm used in VTI_saccade() (detailed below) and
removes the identified saccades before assessing whether separated
fixations are outside of the dispersion tolerance. If they are outside
of this tolerance, the fixation is treated as a new fixation regardless
of the length of saccade separating them. Compared to
fixation_dispersion(), fixation_VTI() is more
conservative in determining a fixation as smaller saccades are
discounted and the resulting data is treated as a continued fixation
(assuming it is within the pixel tolerance set by disp_tol).
In simple terms, fixation_VTI() calculates the saccades
within the data and identifies fixations as (essentially) non-saccade
periods. To avoid eye gaze drift, it applies a dispersion tolerance
parameter as well to ensure that fixations can be appropriately
localised to an x,y coordinate pair. One current limitation to
fixation_VTI() that is not present in
fixation_dispersion() is the need for data to be complete
with no NAs present, otherwise it cannot compute the saccades.
The fixation_VTI() works best on unsmoothed data (with
default settings), as the smoothing process alters the velocity of the
eye movement. When working with smoothed data, lowering the default
threshold parameter is recommended as the “jerky” saccadic starts are
less sudden and so the entry point of a saccade is sooner.
data_fixations_VTI <- fixation_VTI(data,
                                   threshold = 80, #smoothed data, so use a lower threshold
                                   min_dur = 150, # Minimum duration (in milliseconds) of period over which fixations are assessed
                                   min_dur_sac = 20, # Minimum duration (in milliseconds) for saccades to be determined
                                   disp_tol = 100, # Maximum tolerance (in pixels) for the dispersion of values allowed over fixation period
                                   smooth = FALSE,
                                   progress = FALSE) # whether to display a progress bar or not, when running multiple participants ##   pID trialNumber fix_n start  end duration         x        y min_dur disp_tol
## 1 118           1     1     0  723      723  958.9797 598.9772     150      100
## 2 118           1     2   810 1160      350 1379.8363 837.3056     150      100
## 3 118           1     3  1250 1696      446  977.0969 497.9093     150      100
## 4 118           1     4  1703 2156      453  958.3357 179.0800     150      100
## 5 118           1     5  2270 2670      400  377.8732 831.8053     150      100
## 6 118           1     6  2780 3116      336  966.0428 397.5553     150      100This is also a sensible point to briefly highlight the underlying
saccade detection process. This can be accessed directly using
saccade_VTI(). This uses the velocity threshold algorithm
from Salvucci & Goldberg (1996) to determine saccadic eye movements.
It calculates the length of a saccade based on the velocity of the eye
being above a certain threshold.
##   pID trial sac_n start   end duration  origin_x origin_y terminal_x terminal_y
## 1 118     1     1  2170  2260       90  882.6673 239.5014   407.7372   799.6982
## 2 118     1     2  2690  2730       40  500.7881 705.4664   750.8695   493.8955
## 3 118     1     3  3659  3733       74  930.7806 195.9004   522.4383   697.4420
## 4 118     1     4  4213  4256       43  463.0039 721.5722   725.4624   495.3391
## 5 118     1     5 10429 10469       40 1046.7707 226.6985  1202.6984   489.9631
## 6 118     1     6 10475 10502       27 1221.5729 523.9163  1313.1192   697.7792
##   mean_velocity peak_velocity
## 1      207.0041      279.4394
## 2      206.6130      273.4517
## 3      220.6035      330.3718
## 4      201.3317      267.9775
## 5      194.7090      250.8300
## 6      183.0823      228.7045As mentioned above, a supplementary function exists to compare the
two fixation algorithms, the imaginatively named
compare_algorithms(). To demonstrate this, we apply it
against a reduced dataset of just one participant. It takes a
combination of the parameters from both the fixation algorithms, and by
default prints a summary table that is also stored in the returned list.
It is recommended to store this in an object as the output can be quite
long depending on the number of trials.
#some functions are best with single-participant data
data_119 <- data[data$pID == 119,]
comparison <- compare_algorithms(data_119,
                                 plot_fixations = TRUE,
                                 print_summary = TRUE,
                                 sample_rate = NULL,
                                 threshold = 80, #lowering the default threshold produces a better result when using smoothed data
                                 min_dur = 150,
                                 min_dur_sac = 20,
                                 disp_tol = 100,
                                 NA_tol = 0.25,
                                 smooth = FALSE)##     algorithm trial  percent fix_n    corr.r        corr.p   corr.t
## 1         vti     1 81.95538    16 0.7422902  0.000000e+00 61.13824
## 2  dispersion     1 82.02100    25 0.7422902  0.000000e+00 61.13824
## 3         vti     2 86.18421    12 0.7489586 1.297124e-273 44.03837
## 4  dispersion     2 86.90789    17 0.7489586 1.297124e-273 44.03837
## 5         vti     3 85.16072    11 0.6554354 6.261147e-279 41.33860
## 6  dispersion     3 84.94055    16 0.6554354 6.261147e-279 41.33860
## 7         vti     4 88.21797    14 0.7177763  0.000000e+00 53.70491
## 8  dispersion     4 86.19293    20 0.7177763  0.000000e+00 53.70491
## 9         vti     5 87.61361    15 0.7111490  0.000000e+00 62.75578
## 10 dispersion     5 81.87484    26 0.7111490  0.000000e+00 62.75578
## 11        vti     6 91.29406    12 0.6730488  0.000000e+00 56.09738
## 12 dispersion     6 85.87585    27 0.6730488  0.000000e+00 56.09738Once we have collected our fixation data (we will proceed using the
fixations_disp dataset), we can start looking at Areas of
Interest (AOIs) and then plots of the fixations.
For the AOI_ “family” of functions, we need to specify
where our AOIs were presented on the screen. This will enable us to
determine when a participant enters or exits these areas.
# set areas of interest
AOI_areas <- data.frame(matrix(nrow = 3, ncol = 4))
colnames(AOI_areas) <- c("x", "y", "width_radius", "height")
AOI_areas[1,] <- c(460, 840, 400, 300) # Left cue
AOI_areas[2,] <- c(1460, 840, 400, 300) # Right cue
AOI_areas[3,] <- c(960, 270, 300, 500) # outcomes
AOI_areas##      x   y width_radius height
## 1  460 840          400    300
## 2 1460 840          400    300
## 3  960 270          300    500AOI_time() analyses the total time on defined AOI
regions across trials. Works with fixation and raw data as the input
(must use one or the other, not both). This gives a cumulative total
time spent for each trial.
data_AOI_time <- AOI_time(data = data_fixations_disp, 
                          data_type = "fix",
                          AOIs = AOI_areas)
head(data_AOI_time, 10)##    pID trial   AOI time
## 1  118     1 AOI_1 1599
## 2  118     2 AOI_1 1095
## 3  118     3 AOI_1  730
## 4  118     4 AOI_1 3438
## 5  118     5 AOI_1 1090
## 6  118     6 AOI_1  879
## 7  118     1 AOI_2 3401
## 8  118     2 AOI_2 1310
## 9  118     3 AOI_2  570
## 10 118     4 AOI_2 2378The returned data show the time in milliseconds on each area of
interest, per trial. It is also possible to specify names for the
different areas of interest. Or you can request that the function
returns the time spent in AOIs as a proportion of overall time in the
trial, which requires an input vector that has the values, luckily this
is something contained in the HCL_behavioural obect.
AOI_time(data = data_fixations_disp,
         data_type = "fix",
         AOIs = AOI_areas, 
         as_prop = TRUE, 
         trial_time = HCL_behavioural$RT) #vector of trial times##    pID trial   AOI       time
## 1  118     1 AOI_1 0.11875232
## 2  118     1 AOI_2 0.25258076
## 3  118     1 AOI_3 0.39873747
## 4  118     2 AOI_1 0.14046205
## 5  118     2 AOI_2 0.16804136
## 6  118     2 AOI_3 0.45165925
## 7  118     3 AOI_1 0.13860409
## 8  118     3 AOI_2 0.10822511
## 9  118     3 AOI_3 0.55992253
## 10 118     4 AOI_1 0.34688380
## 11 118     4 AOI_2 0.23993300
## 12 118     4 AOI_3 0.23811686
## 13 118     5 AOI_1 0.24637223
## 14 118     5 AOI_2 0.05560327
## 15 118     5 AOI_3 0.47986077
## 16 118     6 AOI_1 0.16825543
## 17 118     6 AOI_2 0.14987941
## 18 118     6 AOI_3 0.45097814
## 19 119     1 AOI_1 0.15903946
## 20 119     1 AOI_2 0.32910147
## 21 119     1 AOI_3 0.22163173
## 22 119     2 AOI_1 0.16408962
## 23 119     2 AOI_2 0.19781477
## 24 119     2 AOI_3 0.39602398
## 25 119     3 AOI_1 0.25199097
## 26 119     3 AOI_2 0.28091446
## 27 119     3 AOI_3 0.29214047
## 28 119     4 AOI_1 0.21153443
## 29 119     4 AOI_2 0.44527168
## 30 119     4 AOI_3 0.13763545
## 31 119     5 AOI_1 0.24571651
## 32 119     5 AOI_2 0.34556075
## 33 119     5 AOI_3 0.21207165
## 34 119     6 AOI_1 0.22065320
## 35 119     6 AOI_2 0.25228779
## 36 119     6 AOI_3 0.35263490As mentioned, it also works with raw data too:
##    pID trial   AOI time
## 1  118     1 AOI_1 1743
## 2  118     2 AOI_1 1353
## 3  118     3 AOI_1  793
## 4  118     4 AOI_1 3776
## 5  118     5 AOI_1 1150
## 6  118     6 AOI_1  946
## 7  118     1 AOI_2 3676
## 8  118     2 AOI_2 1476
## 9  118     3 AOI_2  593
## 10 118     4 AOI_2 2546
## 11 118     5 AOI_2  340
## 12 118     6 AOI_2  817
## 13 118     1 AOI_3 6026
## 14 118     2 AOI_3 4226
## 15 118     3 AOI_3 3070
## 16 118     4 AOI_3 2876
## 17 118     5 AOI_3 2356
## 18 118     6 AOI_3 2763
## 19 119     1 AOI_1 1759
## 20 119     2 AOI_1  943
## 21 119     3 AOI_1 1993
## 22 119     4 AOI_1 2156
## 23 119     5 AOI_1 3673
## 24 119     6 AOI_1 3089
## 25 119     1 AOI_2 3792
## 26 119     2 AOI_2 1047
## 27 119     3 AOI_2 2306
## 28 119     4 AOI_2 4309
## 29 119     5 AOI_2 4932
## 30 119     6 AOI_2 3439
## 31 119     1 AOI_3 2616
## 32 119     2 AOI_3 2183
## 33 119     3 AOI_3 2596
## 34 119     4 AOI_3 1533
## 35 119     5 AOI_3 3113
## 36 119     6 AOI_3 5036When working with raw data, you can also take advantage of
AOI_time_binned(), which enables data binning based on
time, and calculating time spent in AOIs as a result. Using
bin_length, you specify a desired length of bin (say 100ms)
and splits data into these bins, with ant remaining data being dropped
from the analysis (as it does not form a complete bin which could skew
analyses). As with AOI_time() you can specify either
absolute or proportional time spent.
binned_time <- AOI_time_binned(data = data_119,
                               AOIs = AOI_areas, 
                               bin_length = 100,
                               max_time = 2000,
                               as_prop = TRUE)
head(binned_time)##   pID trial bin_n AOI_1 AOI_2 AOI_3
## 1 119     1     1     0     0  0.00
## 2 119     1     2     0     0  0.00
## 3 119     1     3     0     0  0.00
## 4 119     1     4     0     0  0.00
## 5 119     1     5     0     0  0.77
## 6 119     1     6     0     0  1.00The AOI_seq() function analyses the sequence of entries
into defined AOI regions across trials. This works with fixation
data.
data_AOI_sequence <- AOI_seq(data_fixations_disp,
                             AOI_areas,         
                             AOI_names = NULL)   
head(data_AOI_sequence)##   pID trial AOI start  end duration entry_n
## 1 118     1   3   397  643      246       1
## 2 118     1   2   806 1063      257       2
## 3 118     1   3  1390 2166      776       3
## 4 118     1   1  2256 2666      410       4
## 5 118     1   3  2766 3646      880       5
## 6 118     1   1  3749 4199      450       6The returned data provide a list of the entries into the AOIs, across each trial. By default the data is returned in long format, with one row per entry.
Finally, eyetools contains plot_* functions,
plot_seq() plot_spatial(), and
plot_AOI_growth(). These functions are not designed to
accommodate multi-participant data and work best with single trials.
plot_seq() is a tool for visualising the timecourse of
raw data over a single trial. If data from multiple trials are present,
then a single trial will be sampled at random. Alternatively, the
trial_number can be specified. Data can be plotted across
the whole trial, or can be split into bins to present distinct plots for
each time window.
The most simple use is to just pass it raw data:
But the parameters of plot_seq() can help exploration be
more informative. We can also add a background image and/or the AOIs we
have defined:
plot_seq(data, pID_values = 119, trial_values = 1, bg_image = "data/HCL_sample_image.png") # add background imageYou also have the option to split the time into bins to reduce the amount of data plotted
plot_spatial() is a tool for visualising raw eye-data,
processed fixations, and saccades. Fixations can be labeled in the order
they were made. You can also overlay areas of interest (AOIs) and
customise the resolution. It takes separate parameters for fixations,
raw data, and saccades so they can be plotted simultaneously. You can
also specify the trial_number to plot.
Or as mentioned, simultaneously:
plot_spatial(raw_data = data_119,
             fix_data = fixation_dispersion(data_119),
             sac_data = saccade_VTI(data_119),
             pID_values = 119, 
             trial_values = 6)The function plot_AOI_growth() helps visualise how
attention is directed across the development of a trial. It presents a
line graph of the change in proportion of time spent in each AOI region
across the trial time.
#standard plot with absolute time
plot_AOI_growth(data = data, 
                pID_values = 119, 
                trial_values = 1, 
                AOIs = AOI_areas, 
                type = "abs")#standard plot with proportional time
plot_AOI_growth(data = data, 
                pID_values = 119, 
                trial_values = 1, 
                AOIs = AOI_areas, 
                type = "prop")#only keep predictive and non-predictive cues rather than the target AOI
plot_AOI_growth(data = data, pID_values = 119, trial_values = 1, 
                AOIs = AOI_areas, 
                type = "prop", 
                AOI_names = c("Predictive", "Non Predictive", NA))Beesley, T., Nguyen, K. P., Pearson, D., & Le Pelley, M. E. (2015). Uncertainty and predictiveness determine attention to cues during human associative learning. Quarterly Journal of Experimental Psychology, 68(11), 2175-2199.↩︎
Salvucci, D. D., & Goldberg, J. H. (2000). Identifying fixations and saccades in eye-tracking protocols. Proceedings of the Symposium on Eye Tracking Research & Applications - ETRA ’00, 71–78.↩︎