Evaluation of Driver Distraction with Changes in Gaze Direction Based on a Vestibulo-Ocular Reflex Model

With the aim of improving parameter identification and, eventually, evaluating driver distraction with changes in gaze direction, we applied a genetic algorithm (GA) method to identify parameters for an existing vestibulo-ocular reflex (VOR) model. By changing the initial inputs to the GA and fixing two parameters pertaining to the horizontal direction, we achieved improved parameter identification with a lower mean-square error. The influence of driver distraction on eye movement with changes in gaze direction was evaluated from the difference between the predicted and observed VOR in the vertical axis. When a driver was given an additional mental workload, the mean-square error between the measured and simulated values was bigger than that in the absence of the mental workload. This confirmed the relationship between driver distraction and eye movement in the vertical direction. We hope that this method can be applied in evaluating driver distraction.


Introduction 1.The Vestibulo-Ocular Reflex Model and Its Application
The vestibular system, which is a sensory mechanism in the inner ear, provides the principal contribution to the sense of balance and spatial orientation.The system consists of two parts: the otoliths and the semicircular canals.The former have two main functions: detection of linear motions of the head, and detection of the head's position relative to the force of gravity.The semicircular canals consist of three canals in each ear: the horizontal, anterior, and posterior canals; the horizontal canal detects angular accelerations of the head, whereas the anterior and posterior canals detect vertical movements.Previous researches have shown that movements of the eye can be estimated from movements of the head [1] [2].
When the head is turned to the left, the eyes move in the opposite direction to stabilize the visual world based on the input to the vestibular organ.This vestibule-ocular reflex (VOR) is a reflex eye movement that can be used to estimate eye movements from movements of the head.VOR models have been proposed by several researchers and have been used in various applications.In this research, we used the VOR model proposed by [1], which reflects interactions between the otoliths and the semicircular canals.In this model, head movement is represented by a linear acceleration and an angular velocity as inputs, and the movement of the eyeball is the output [2] (Figure 1).uations of simulated brake motion.In their research, the passenger's comfort was measured by comparing the observed eye movement with that simulated by using the VOR model.The researchers found a tendency toward a relationship, but the number of subject was insufficient for the purpose of making a decision.
In addition, their parameter identification by using a hybrid genetic algorithm left a gap between simulation and measurement.Obinata's group has also used the VOR model to evaluate mental workload [3] and driver distraction in terms of memory-decision workload [4].They presented a new method for quantifying mental workloads by utilizing the VOR.They did not, however, consider changes in gaze direction.
In the case of seated drivers controlling a steering wheel, the head movement has three axes of translation and rotation in the Cartesian coordinate system.However, Obinata et al. have stated that the horizontal axis has a much narrower range of movements than the vertical axis [3].Consequently, we only considered movements in vertical direction, which have a strong effect on eye movement.

Parameter Identification for the VOR Model
Currently several methods have been proposed for identifying good parameters for a model; these include brute-force search, the Nelder-Mead simplex method, the trust region method, and the Gauss-Newton method [5].Another successful method is the genetic algorithm (GA) method, which is based on Darwinian evolutionary principles: the mechanics of natural genetics (Goldberg, 1989).This method has been applied in a variety of areas, particularly for parameter estimations [6] [7] [8] [9].The GA is a suitable method for optimization of nonlinear problem by determining global optimal solutions through natural selection and genetics.
For VOR parameters, Obinata et al. used a hybrid GA method to identify [10] [11] [12].In these cases, eight parameters from Merfeld and Robinson's studies were examined simultaneously, with Merfeld and Zupan parameters as initial values.However, because of the wide range of each parameter, results after 100 generations were still not consistent with the measured values.The mean-square error increased as the time for one simulation or one experiment increased.Furthermore, the mean-square error also increased with increasing number of changes in gaze direction.
We therefore needed to improve the accuracy of the parameters in our research.To do this, in this study, we applied the technique of Son et al. (2015) to enhance the GA by generating near-optimal initial populations [13] to determine the range of each parameter and by changing the population size and relative elitism [14] [15].In addition, an improved method was applied for parameter identification to permit the evaluation of driver distraction with changes in gaze direction.

Objective
Because the difficult of simulating eye movement with changing the gaze, the evaluation driver cognitive distraction became challenged in the past.Therefore, in this study, by improving the parameter identification method, the eye will be simulated and used to evaluate the driver distraction in case of changing gaze.
To do that, in this research, we had two main objectives: • To improve parameter identification by updating the range of each parameter, based on the results of [15] and to fix two parameters pertaining to the horizontal direction; and • To estimate mental workload on the basis of the difference between the predicted and actual eye movements.
In this research, we used the gap between the observed VOR responses and those predicted by the mathematical model as a measure of performance.

The Vestibulo-Ocular Reflex Model
The main purpose of the VOR model is stabilizes images on retina during head movement by producing an eye velocity that is equal and opposite to head velocity.Figure 2 shows schematically how the VOR works during a horizontal head movement.
According to the results obtained by Zupan and Merfeld [16] and by Angelaki and co-workers [17]  gravitational (g, orientation) and translational (f) components.In the second, gravitational estimates are used to transform head-fixed angular-velocity signals from the semicircular canals (ω) to the inertial velocity, i.e., the space-referenced angular velocity (ωS).
Figure 3 shows the detailed blocks of the VOR model.The semicircular canals [which measure the angular velocity of the head (ω)] and the otolith organs [which measure linear accelerations of the head (α) and gravity (g)] sense the proprioceptive information from the environment; this is the first step in calculating sensory information from the measurements (α oto , α scc ).In the next step, the measurements are compared with the sensory information predicted by the internal model.This model combines four free parameters (k ω , k fω , k f , and k a ).
The parameters values were determined by the feedback of this error.
Because we're only considered with the head and eye movements in the vertical direction, the two parameters for the horizontal gain (p kshor and p krhor ) remained constant.
After calculating the eye movement, the final common path proposed by [2] was applied.In this part, two parameters (k i , k p ) were used, based on the different types of muscle fibers present in muscles of the eye (Figure 4).

• Parameter Identification
In the experiment, a subject was asked to sit on the seat of a driving simulator with six degrees of freedom.The driving simulator had a cylindrical 360˚ screen 6 m in diameter.The simulator was controlled by CarSim (Mechanical Simula-  tion Co.) which can simulate the dynamic behavior of a vehicle (Figure 5).By using MATLAB Simulink (MathWorks) to control CarSim, the seat was moved with a fixed frequency in the vertical plane, as in previous experiments.
We measured eye movements by using SmartEye Pro (Smart Eye AB) with four cameras on the simulator.This equipment is noninvasive, simple to install, and provides data by using a camera recorder.To collect information on movements of the head, we used a Fastrak electromagnetic tracker (Polhemus Inc.).
A total of 15 subjects who drive on a regular basis participated in the experiment to test the parameter-identification method.During the parameter-identification phase, the participant was seated in the car simulator and looked straight ahead for 10 s without changing their gaze.The seat movements were be controlled by CarSim and were identical for each individual.
• Driver distraction Subsequently, we conducted an experiment to examine the effect of mental workload on eye movement.The gaze movement was as shown below (Figure 6).
In this case, CarSim produced identical movements of the car seat in all the An n-back task was used to evaluate distraction.In the n-back task, a series of one-digit numbers is presented verbally to the subject.The subject is then asked to answer by pressing a button on the steering wheel when the current number matches the number given n steps earlier in the sequence.In this study, n = 1 and the interval between two successive numbers was 2 s.
A total of 15 subjects participated (two trials per subject); the subjects had different individual characteristics.

Parameters Identification
Data from 15 subjects who drive every day were collected for parameter identification.The collected data were subjected to a cleaning process and subsequently used in parameter identification, according to the method proposed by Son et al. (2015).The result of parameter identification showed a good performance with an average mean-square error of 3.09E−04; the maximum meansquare error was 6.13E−04 and the minimum was 6.12E−05 (Figure 7).By changing the initial input of the GA and fixing two parameters in final common-path part (p_kshor and p_krhor), the result of parameters identification by using Fastrak and SmartEye Pro in this experiment produced a better performance than motion capture (MAC3D) or the eye tracker (T.K.K.2930a, Takei Scientific Instruments Co., Ltd.) [15].The average of the mean-square errors was reduced from 1.26E−03 to 3.09E−04.The maximum and minimum values of the mean-square error were also lower than the previous ones; in the study by Son et al., the maximum was 4.4E−03 and the minimum was 6.5E−05, whereas in the current study, the maximum mean-square error was 6.13E−04 and the minimum was 6.12E−05.plicable to experiments with SmartEye and Polhemus.

Parameters Identification for the Driver Distraction Experiment
As reported by Obinata et al. and others [3] [4] [15] [17], motion sickness has a strong effect on VOR parameters.To identify the optimal VOR parameters for driver distraction, we used the first ten seconds in our distraction-evaluation experiment for parameter identification, and then we applied the identified parameters throughout the remainder of the experiments.The results of parameter identification for each subject in the driver-distraction experiment showed good performance, with a low mean-square error (Table 1).The range of mean-square errors was 1.59E−04 to 5.96E−04, with an average of 3.23E−04.The time and the frequency responses also matched well (Figure 9).We conducted our experiments as follows.First, the participant was asked to gaze steadily at the target for 10 s to permit parameter identification.We then used the results of the parameter identification to simulate the second part of the experiment, in which the subject changed their case sequentially according to instructions.The results of the final simulation measurements are shown in Figure 10.
As can be seen in Figure 10, for the eye movement in the vertical axis of Subject 2, the new parameters showed a good performance with a mean-square error of 9.25E−04.Conversely, the parameters proposed by Merfeld and Zupan (2002) were poor in terms of both the trend and value, and the mean-square error in this case was very high (7.33),more than 7000 times larger than that achieved with our new set of parameters.Similar results were obtained with the other subjects.After incorporation of the offset (Figure 12), the mean-square value for Subject 2 fell to 4.25E−04.When an n-back task with a 1 s interval was used, a large difference was observed between the case with an MWL and that with no MWL.The mean-square error without an offset increased 1.33 times to 0.0123.With an offset, it increased 2.83 times to 0.0012.Figure 13 shows details of the simulation and results of measurement for this case.

Effect of the N-Back Task on Eye Movement with Changing Gaze Direction
With an MWL, the time and frequency response became mismatched.Consequently, the mean-square error increased compared with that in the absence of an MWL.
We conducted the same experiments for each participant with and without an MWL and we took the average of two trials for each subject; the details of the mean-square error are shown in Table 2.
Next, we performed an analysis of variance (ANOVA) for 30 trials of 15 subjects in the presence and absence of a mental workload, and with and without an offset with the following null hypothesis.
-The results with and without an offset are equal.
-The results in the presence and absence of a mental workload are equal.
-The results of Trial 1 and Trial 2 are equal.
The result of the ANOVA is shown in Table 3.
As expected, the effects of the offset and MWL were statistically significant (F = 142, p < 0.001, and F = 40.9,p < 0.001, respectively).The mean-square error in  the presence of an offset was significantly smaller than that in the absence of an offset, whereas that for no MWL was significantly smaller than that with a MWL.This shows that the method we have developed works well for evaluating MWL.There was no significant difference between the two trials and no interaction between the trial for MWL + offset and that with MWL only (p = 0.424 and 0.52, respectively).This shows that our method can be used to evaluate driver distraction on the basis of the difference in mean-square error between the simulated eye movement and the measured values.The results are shown in more detail in Figure 14.
Our results showed a better performance than those previously obtained by Obinata and co-workers [3] [4] [11] [12].The difference between the case with an MWL and that in its absence was clearer than the previous case, in which the average mean-square error for no MWL was 1.56E−03 and that with an MWL was 3.48E−03 (more than two times).On the other hand, in previous researches, the participants focused one fixed point and the computer controlled the seat movement.In our study, the subject changed their gaze throughout the study, which is much closer to actual behavior when driving.

Summary/Conclusions
These results confirm that our method can be applied in VOR experiments, and can be performed with a noninvasive sensor with a low mean-square error.In the near future, we hope to quantify in real time in a real vehicle.The distraction evaluation needed to combine the VOR model with optokinetic eye-movement response (OKR) models to determine the motions of the eyeball based on head movement and visual-scene movement.
The increase in the mean-square error in the presence of a demanding mental task implies that there is a relationship between the VOR mechanism and driver distraction.This result is consistent with those of previous studies by other re-searchers.On the other hand, our research produced a better performance with a lower mean-square error.With respect to applications pertaining to changes in gaze direction, we hope in the future that our technique will permit the successful application of VOR to the quantification of driver state for better cockpit design and development of driver-assistance systems.
Because of the limitation of subject, in the future, we will continue conducting the effect of driver cognitive distraction on eye movement not only in driving simulator but also actual vehicle to validate this method.

Figure 8 Figure 7 .
Figure8shows the eye movements in the vertical axis for Subject 2. The red line is the measured eye movement and the blue line is the simulation.The results shown in Figure8confirm that the identification produces good matches in time responses and frequency responses.This means that our method is ap-

Figure 8 .
Figure 8. Eye movement in the vertical axis for subject 2.

Figure 11
Figure 11 shows the original measurements and simulation results for Subject 2 in the absence of an additional mental workload (MWL) for a total of 83.8 seconds.The mean-square error for Subject 2 without an offset was 9.2E−03.

Figure 10 .
Figure 10.Eye movement in vertical axis of Subject 2.

Figure 12 .
Figure 12.Vertical eye movement simulation with offset.

Figure 13 .
Figure 13.Vertical eye-movement simulation with a mental workload.

Figure 14 .
Figure 14.Average mean square error with offset for each subject.

Table 1 .
Mean-square error in parameter identification for each subject.

Table 2 .
Mean-square error in the presence and absence of an MWL.