5
Multimodal Classification of Driver Glance Author 1 Institution 1 Address 1 Address 1 Email 1 Author 2 Institution 2 Address 2 Address 2 Email 2 Author 3 Institution 3 Address 3 Address 3 Email 3 ABSTRACT This paper presents a multimodal approach to in-vehicle classification of driver glances. Driver glance is a strong predictor of cognitive load and is a useful input to many ap- plications in the automotive domain. Six descriptive glance regions are defined and a classifier is trained on video record- ings of drivers from a single low-cost camera. Visual features such as head orientation, eye gaze and confidence ratings are extracted by tracking software, then statistical methods are used to perform failure analysis and calibration on the vi- sual features. Non-visual features such as steering wheel angle and indicator position are extracted from a RaceLogic VBOX system. We evaluate our approach on a dataset of 13 drivers, each consisting of 60 seconds at 30 frames per second. We compare our multimodal approach to separate unimodal approaches using both Support Vector Machine (SVM) and Random Forests (RF) classifiers. RF Mean De- crease in Gini Index is used to rank selected features which gives insight into the selected features and improves the clas- sifier performance. We demonstrate that our multimodal approach yields significantly higher results than unimodal approaches. The final model achieves an average F1 score of 68.3% across the six classes. Keywords Multimodal Classification; Glance Classification; Driver Dis- traction; Vehicle CAN data 1. INTRODUCTION Driver distraction has emerged as one of the leading causes of vehicle related accidents [1]. The increasing adoption of digital interactions inside the vehicle threatens to aggravate this issue. Driver distraction can be modelled by consider- ing factors such as orientation, duration and history of the driver’s gaze. This can then be used for notification schedul- ing and other safety related applications. Liang et al. [1] published research on the 100-Car Natu- ralistic Driving Study comparing 24 driver distraction algo- Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full cita- tion on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or re- publish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]. WOODSTOCK ’97 El Paso, Texas USA c 2016 ACM. ISBN 123-4567-24-567/08/06. . . $15.00 DOI: 10.475/123 4 rithms. These algorithms relied heavily on three characteris- tics: glance location, duration and history. Although based on limited manually-labelled data, the research highlights the theoretical importance of differentiating the glance lo- cations as categories rather than raw angles. This is since driving related glances such as side mirrors are less risk in- ducing than glances at in-vehicle systems [1]. The research demonstrates the importance of considering a short window (less than 3 seconds) of off-road glances to predict crash risk. This motivates the need for a qualitative classifier versus a quantitative one. We propose a qualitative 6-class driver glance classifica- tion system. The labels Forward, Down, Left, Right, Left Blind Spot, and Right Blind Spot are chosen as useful de- scriptors of the driver’s glance. A baseline glance classifier is first implemented that classifies the glance direction based on the yaw and pitch of the driver’s head. This baseline model assumes a perfect head tracker and linearly separable regions. However due to poor calibration, visual noise and limitations of the tracker, more predictors are required to determine the driver’s glance accurately. We propose a mul- timodal approach combining visual cues with non-visual car parameters to classify driver glances. Visual features such as head position, orientation, confidence and eye gaze orien- tations are extracted. Statistical methods are implemented to filter, calibrate and extrapolate the visual information. Non-visual car parameters such as steering wheel angle and indicator position are extracted from CAN (Controller Area Network) data collected by a RaceLogic VBOX system. The main contributions of this paper can be summarised as follows: 1. Proposing a multimodal approach for driver glance classification that combines driver’s facial features with car parameters. 2. Demonstrating that our multimodal approach performs significantly better than single modalities. 3. Presenting a ranking of predictive importance of the vi- sual and non-visual features considered, which informs future work in this field. This paper will begin with a discussion of the related work in Section 2. In Section 3, the construction and preprocess- ing of the dataset is described. In Section 4, a comparison of four designed classifiers will be given followed by a dis- cussion. Lastly in Section 5, conclusions will be drawn from the results and suggestions for future work will be given.

Multimodal Classification of Driver Glancepr10/publications/icmi16.pdf · This yielded a total of 23400 samples. In order to syn-chronize car information with the driver camera,

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Multimodal Classification of Driver Glancepr10/publications/icmi16.pdf · This yielded a total of 23400 samples. In order to syn-chronize car information with the driver camera,

Multimodal Classification of Driver Glance

Author 1Institution 1Address 1Address 1Email 1

Author 2Institution 2Address 2Address 2Email 2

Author 3Institution 3Address 3Address 3Email 3

ABSTRACTThis paper presents a multimodal approach to in-vehicleclassification of driver glances. Driver glance is a strongpredictor of cognitive load and is a useful input to many ap-plications in the automotive domain. Six descriptive glanceregions are defined and a classifier is trained on video record-ings of drivers from a single low-cost camera. Visual featuressuch as head orientation, eye gaze and confidence ratings areextracted by tracking software, then statistical methods areused to perform failure analysis and calibration on the vi-sual features. Non-visual features such as steering wheelangle and indicator position are extracted from a RaceLogicVBOX system. We evaluate our approach on a dataset of13 drivers, each consisting of 60 seconds at 30 frames persecond. We compare our multimodal approach to separateunimodal approaches using both Support Vector Machine(SVM) and Random Forests (RF) classifiers. RF Mean De-crease in Gini Index is used to rank selected features whichgives insight into the selected features and improves the clas-sifier performance. We demonstrate that our multimodalapproach yields significantly higher results than unimodalapproaches. The final model achieves an average F1 score of68.3% across the six classes.

KeywordsMultimodal Classification; Glance Classification; Driver Dis-traction; Vehicle CAN data

1. INTRODUCTIONDriver distraction has emerged as one of the leading causes

of vehicle related accidents [1]. The increasing adoption ofdigital interactions inside the vehicle threatens to aggravatethis issue. Driver distraction can be modelled by consider-ing factors such as orientation, duration and history of thedriver’s gaze. This can then be used for notification schedul-ing and other safety related applications.

Liang et al. [1] published research on the 100-Car Natu-ralistic Driving Study comparing 24 driver distraction algo-

Permission to make digital or hard copies of all or part of this work for personal orclassroom use is granted without fee provided that copies are not made or distributedfor profit or commercial advantage and that copies bear this notice and the full cita-tion on the first page. Copyrights for components of this work owned by others thanACM must be honored. Abstracting with credit is permitted. To copy otherwise, or re-publish, to post on servers or to redistribute to lists, requires prior specific permissionand/or a fee. Request permissions from [email protected].

WOODSTOCK ’97 El Paso, Texas USAc© 2016 ACM. ISBN 123-4567-24-567/08/06. . . $15.00

DOI: 10.475/123 4

rithms. These algorithms relied heavily on three characteris-tics: glance location, duration and history. Although basedon limited manually-labelled data, the research highlightsthe theoretical importance of differentiating the glance lo-cations as categories rather than raw angles. This is sincedriving related glances such as side mirrors are less risk in-ducing than glances at in-vehicle systems [1]. The researchdemonstrates the importance of considering a short window(less than 3 seconds) of off-road glances to predict crash risk.This motivates the need for a qualitative classifier versus aquantitative one.

We propose a qualitative 6-class driver glance classifica-tion system. The labels Forward, Down, Left, Right, LeftBlind Spot, and Right Blind Spot are chosen as useful de-scriptors of the driver’s glance. A baseline glance classifier isfirst implemented that classifies the glance direction basedon the yaw and pitch of the driver’s head. This baselinemodel assumes a perfect head tracker and linearly separableregions. However due to poor calibration, visual noise andlimitations of the tracker, more predictors are required todetermine the driver’s glance accurately. We propose a mul-timodal approach combining visual cues with non-visual carparameters to classify driver glances. Visual features suchas head position, orientation, confidence and eye gaze orien-tations are extracted. Statistical methods are implementedto filter, calibrate and extrapolate the visual information.Non-visual car parameters such as steering wheel angle andindicator position are extracted from CAN (Controller AreaNetwork) data collected by a RaceLogic VBOX system.

The main contributions of this paper can be summarisedas follows:

1. Proposing a multimodal approach for driver glanceclassification that combines driver’s facial features withcar parameters.

2. Demonstrating that our multimodal approach performssignificantly better than single modalities.

3. Presenting a ranking of predictive importance of the vi-sual and non-visual features considered, which informsfuture work in this field.

This paper will begin with a discussion of the related workin Section 2. In Section 3, the construction and preprocess-ing of the dataset is described. In Section 4, a comparisonof four designed classifiers will be given followed by a dis-cussion. Lastly in Section 5, conclusions will be drawn fromthe results and suggestions for future work will be given.

Page 2: Multimodal Classification of Driver Glancepr10/publications/icmi16.pdf · This yielded a total of 23400 samples. In order to syn-chronize car information with the driver camera,

Figure 1: Overview of the process in generating the driver glance classifiers.

2. RELATED WORKThe problem of modelling driver distraction using multiple

modalities such as visual cues, audio, CAN data, biometricsensors and GPS has been investigated widely in recent years[1, 2, 3, 5, 6, 7]. However, attempts to classify driver glancegenerally depend on specialised equipment. For example,Sodhi et al. [7] used a head-mounted eye-tracking deviceto evaluate driver distraction based on eye positions andpupil diameters. Kutila et al. [6] utilised a stereo cameraset-up combined with lane-tracking data in order to predictcognitive workload. They classified driver glance into fourcategories: “Road Ahead”, “Left Mirror”, “Right Mirror”,“Windscreen”. However the total cost of this system wasreported to be around e35 000. Ji et al. [2] used a configu-ration of infrared LEDs to illuminate the driver’s pupils forbetter head and gaze tracking. Their system also consistedof two cameras and extracted additional visual cues such aseyelid movement. The use of infrared LEDs enabled accu-rate eye gaze tracking, but the range of head movement wasstill limited. Our new approach focuses on using a single,low-cost camera without depth information combined withCAN-Bus data.

3. METHODOLOGYIn this section the dataset, feature extraction methods,

and evaluation methods are described.

3.1 DatasetTrials were performed in a Land Rover Discovery fitted

with a standard, uncalibrated USB Mobius camera moni-toring the driver’s face and a RaceLogic VBOX. A drivingroute was specified of about 90 minutes. Sixty seconds ofnaturalistic driver footage from each of 13 drivers was se-lected at 30 frames per second and processed at a resolutionof 640x360 where significant head turning was naturally ex-hibited. The VBOX system recorded car information atonly 10 frames per second. The frames were annotated withCowLog 3.0.2 [8] serving as the ground truth. The classlabels Forward, Down, Left, Right, Left Blind Spot, RightBlind Spot relative to the camera’s perspective were chosenas significant regions which can then be further expanded onby other vehicle applications. The front car pillars and base

Figure 2: Top view allocation of driver glance re-gions. The directions are relative to the camera po-sitioned to the right of the right-seated driver.

of the windscreen were chosen as the separating boundaries.These regions are visualised in Figure 2. It must be notedthat some classes were difficult to disambiguate during fasthead turns as well as slight eye changes near the boundaries.A Rear View Mirror class was considered, but removed fromour taxonomy due to difficulty in distinguishing it from theForward class.

This yielded a total of 23 400 samples. In order to syn-chronize car information with the driver camera, the audioof the two devices were matched.

3.2 Feature Extraction

3.2.1 Visual FeaturesOver the past decade, many head and eye tracking meth-

ods have been proposed. In this paper, the Cambridge FaceTracker developed by Baltrusaitis et al. [4] was selected forvisual feature extraction. This was chosen due to its real-time operation and focus on performance in the presence ofpoor lighting conditions, occlusions and extreme poses. Themethod locates facial landmarks as a function of spatial re-lationships and certainties of other detected landmarks. Thesystem works from a single simple camera without a depthsensor, generalises well to different faces and returns a con-fidence measure which is useful in our subsequent analysis.

Page 3: Multimodal Classification of Driver Glancepr10/publications/icmi16.pdf · This yielded a total of 23400 samples. In order to syn-chronize car information with the driver camera,

The framework extracted the following features from thedata: Head Position (Horizontal, Vertical, Depth), HeadOrientation (Pitch, Yaw, Roll), Head Orientation Confi-dence, Eye Orientations (X1,X2,Y1,Y2,Z1,Z2) and an EyesDetected Flag.

3.2.2 Non-Visual FeaturesThe non-visual data was captured at 10 frames per second

with a RaceLogic VBOX. Each VBOX instance was dupli-cated twice in order to match the visual frame rate. Thesystem provides .vbo files for each set of readings whichwere then converted to .csv files. The following featureswere selected from this system: Velocity, Accelerator PedalPosition, Brake Pressure, Steering Wheel Angle, IndicatorPosition, Gear Lever Position. These features are hypothe-sized to be correlated with driver glance behaviour.

3.3 Tracker Failure AnalysisDespite the tracker’s robustness, it often lost track of its

target. The success rate is further limited by the detectablerange of head orientations. The tracker provided a proba-bilistic confidence measure for each head tracking estima-tion. Out of all the recorded frames, 20.4% had confidencelevels of less than 50% and 6.6% had confidence levels lessthan 1%. A low pass filter was implemented to determine thelocal head rotational velocity in order to extrapolate missingand low confidence (tuned to a threshold of 9%) values. Thelow pass filter was required to smooth out the noise in thereadings that could otherwise forecast spurious values.

3.4 CalibrationEach driver was positioned differently relative to the cam-

era. In order to regularize different driver positions androtations, the average head position and rotation for eachdriver is subtracted from the readings. This average is ac-quired by considering a sliding window of 60 seconds fromwhich to calibrate the data. Furthermore only frames with ahigh head tracking confidence (tuned to a threshold of 60%)are considered in the calibration calculations to compensatefor spurious readings. This ensures that the calibration isuser-independent and adjusts with driver position changes.

4. EVALUATIONIn this section, the implementation details are discussed

followed by a comparison of the different constructed non-linear classifiers. Visual and non-visual modalities are firstconsidered separately. The multimodal approach then con-siders the combination of these features. For each classifier,both Support Vector Machines (SVM) and Random Forests(RF) classifiers are trained. The individual class F1 scoresare reported as well the arithmetic mean.

4.1 Implementation DetailsTwo methods were chosen to train the classifiers. Ran-

dom Forests were chosen as they provide variable impor-tance techniques and are simpler to optimise. Support Vec-tor Machines were chosen as they generally perform betteron smaller datasets.

In order to design a classifier that generalises to unseendrivers, the system is trained with Leave-One-Out Cross Val-idation. For each combination, the classifier is trained on 12drivers and tested on the remaining one.

The number of features was optimised by stepwise back-ward elimination feature selection guided by Random ForestMean Decrease in Gini Index. In both classifiers the objec-tive function was chosen to be the cross-validated arithmeticmean of the test class F1 scores. This function ensures thatthe size of the classes do not affect their weighting in the op-timisation function. Class sub-sampling is performed beforetraining the classifiers due to the largely imbalanced labelfrequencies. Specifically, the Left and Right Blind Spot la-bels contain 288 and 662 instances respectively out of thetotal of 23 400.

The Radial Basis Function was chosen as the SVM ker-nel. The model parameters were chosen by an iterative loga-rithmic grid search to optimize the cross-validated objectivefunction. The RF model was tuned by first maximising thecross-validated objective function by adjusting the allocatednumber of features per candidate split, then adjusting thenumber of decision trees in the forest via a linear search.

4.2 Baseline algorithmAs a baseline we chose a Rule-based classifier using the

visual yaw and pitch of the drivers head, since head poseis a good proxy for gaze direction. Without knowledge ofthe shortcoming of available trackers, one might expect thisclassifier to be sufficient. For example, if the head pitch ex-ceeded a threshold, the frame would be labelled as Down.The process was then continued for the remaining classes us-ing the head yaw. Threshold values were chosen by inspec-tion from annotated training footage and class distributions.Calibration is omitted, but failure analysis is performed sothat blind spots can be extrapolated. The performance ofthe classifier is tabulated in Table 1.

The baseline approach struggles to differentiate betweenthe Left and Left Blind Spot given the head yaw and pitch.It assumes a linear separable boundary and does not accountfor the noise present in the features. The other approachesuse machine learning methods that build more sophisticated,non-linear boundaries between the classes that account forthe presence of noise and the additional available features.

4.3 Visual ClassifierFor this classifier additional visual information is incorpo-

rated from the tracker such as the remaining head rotationvalue, eye information, confidence values and head angularvelocities. Failure analysis is included to extrapolate miss-ing values and calibration is applied to account for driverposture changes. Feature selection is performed to removeinsignificant features. The resulting performance of the clas-sifiers are tabulated in Table 1.

The RF and SVM classifiers produced comparable resultsexcept for the Left Blind Spot label. The following variableswere considered and their variable importances were rankedin order of mean decrease in RF Gini Index (most significantto least significant): head yaw, head pitch, horizontal headposition, eye orientation X2, head roll, eye orientation X1,head confidence level, vertical head position, eye orientationY1, eye orientation Y2, change in head yaw, head depth, eyeorientation Z1, change in head roll, change in head pitch,eyes detected flag. The classifier performed best when thefeatures after the vertical head position were disregarded.As expected from the baseline approach assumptions, thehead yaw and pitch are the most important variables. Theeye orientation X1 component also scores highly as expected

Page 4: Multimodal Classification of Driver Glancepr10/publications/icmi16.pdf · This yielded a total of 23400 samples. In order to syn-chronize car information with the driver camera,

Algorithm Modalities Forward Down Left Right Left Blind Spot Right Blind Spot Mean

Baseline Visual 68.4 54.8 3.2 43.5 6.85 33.5 35.0

SVM Visual 84.3 59.4 88.0 47.4 60.1 52.3 65.2RF Visual 81.5 57.0 88.3 53.5 19.3 49.4 58.2

SVM Non-Visual 66.5 38.6 18.9 17.8 0.0 18.0 26.6RF Non-Visual 66.1 33.6 36.9 20.7 22.2 9.2 31.5

SVM Multimodal 82.5 65.5 90.4 45.1 71.5 54.7 68.3RF Multimodal 80.8 64.2 86.9 55.2 49.6 51.3 64.7

Table 1: F1 scores [%] for each glance class for the baseline, SVM and RF classifiers.

since the glance classification can shift solely on the changein eye positions. The horizontal head position scores well onthe list due to a tendency of drivers to translate horizontallywhen turning their head. The confidence level aids the clas-sification since it is likely to be inversely correlated to theangle between the drivers glance and the camera’s principleaxis.

4.4 Non-Visual ClassifierThis classifier is trained purely on non-visual VBOX data.

The performance of the classifiers are tabulated in Table 1.The non-visual classifier performs surprisingly well given thefeatures and limited training data. This indicates the impor-tance of non-visual CAN data in glance estimation. Further-more, these features do not require any failure analysis orcalibration and contain negligible noise. The variables con-sidered were calculated to have a variable importance rank-ing as follows (most significant to least significant): SteeringWheel Angle, Velocity, Brake Pressure, Gear Lever Position,Accelerator Pedal Position, Indicator Position. The RF clas-sifier yielded the best performance which was achieved usingonly the first three features.

4.5 Multimodal ClassifierThis classifier merges the non-visual information from the

CAN data with the visual information. The performance ofthe classifiers are tabulated in Table 1.

In this case the SVM performs better than the RF clas-sifier. The variables considered were calculated to have avariable importance ranking as follows (most significant toleast significant): Head yaw, horizontal head position, headpitch, head confidence, head roll, eye component X1, verti-cal head position, steering wheel angle, gear lever position.The remaining features did not improve the cross validatedoptimisation function.

By inspecting the confusion matrix for the SVM classifierin Table 2, it is possible to interpret the ambiguities be-tween the chosen classes. As can be seen from the confusionmatrix, neighbouring classes are often confused, whereasglances that are distant are less likely to be mistaken.

4.6 DiscussionAs can be seen from the Table 1, the SVM multimodal

approach surpasses all of the other classifiers considered.The significance is verified with a paired samples t-test.The multimodal SVM results for each driver were signifi-cantly more accurate than the unimodal visual SVM results(t(12) = 1.87, p = 0.0428).

The main bottleneck for the performance of the classifiersis the success rate of the tracker. Tracking failure increased

PGT

F D L R LB RB Precision

F 2655 152 99 239 0 42 83.3D 114 1104 32 327 0 47 68.0L 129 25 2381 0 43 0 92.4R 286 266 26 661 0 72 50.4

LB 9 0 143 0 245 0 61.7RB 57 201 7 396 0 496 42.9

Recall 81.7 63.2 88.6 40.7 85.1 75.5

Table 2: Confusion matrix for the multimodal classi-fier. Rows represent the predicted values. Columnsrepresent the ground truth values.

with the angle of the driver glance to the camera principleaxis. Augmenting the visual features with the CAN datahelped to achieve better performance in these cases. A sec-ond camera at the adjacent front corner of the car could beincorporated to further raise the performance of the Rightand Right Blind Spot classes. The size of the training setwas another limitation. It is expected that a larger datasetwould significantly improve the classifiers.

5. CONCLUSIONS AND FUTURE WORKWe have presented a novel multimodal approach to clas-

sify driver glance behaviour into six classes. We demonstratethat augmenting the visual features with vehicle CAN dataoutperformed the separate single modality alternatives. Vi-sual features such as head and eye gaze orientation are suc-cessfully combined with non-visual features such as steeringwheel angle and gear lever position. The multimodal SVMclassifier returns F1 scores of 82.5%, 90.4% and 71.5% forthe Forward, Left and Left Blind Spot regions respectively.The shortcoming of the other labels are primarily due to thelarge angle between the driver glance and the camera princi-ple axis. This can be ameliorated by incorporating a similarsecond camera at the adjacent front corner of the car duringdata collection.

For future work, we would like to investigate the additionof a Rear View Mirror class given more data and a secondcamera. As for non-visual modalities, we would like to in-corporate GPS data by considering the expected behaviourgiven the driver location.

6. ACKNOWLEDGMENTS[Space held for acknowledgements after anonymous re-

view.]

Page 5: Multimodal Classification of Driver Glancepr10/publications/icmi16.pdf · This yielded a total of 23400 samples. In order to syn-chronize car information with the driver camera,

7. REFERENCES[1] Y. Liang, J. D. Lee, and L. Yekhshatyan, “How

dangerous is looking away from the road? algorithmspredict crash risk from glance patterns in naturalisticdriving,” Human Factors: The Journal of the HumanFactors and Ergonomics Society, vol. 54, no. 6,pp. 1104–1116, 2012.

[2] Q. Ji and X. Yang, “Real-time eye, gaze, and face posetracking for monitoring driver vigilance,” Real-TimeImaging, vol. 8, no. 5, pp. 357–377, 2002.

[3] C. Busso and J. Jain, “Advances in multimodaltracking of driver distraction,” in Digital SignalProcessing for In-Vehicle Systems and Safety,pp. 253–270, Springer, 2012.

[4] T. Baltrusaitis, P. Robinson, and L.-P. Morency,“Constrained local neural fields for robust faciallandmark detection in the wild,” in Proceedings of theIEEE International Conference on Computer VisionWorkshops, pp. 354–361, 2013.

[5] P. Angkititrakul, M. Petracca, A. Sathyanarayana,and J. H. Hansen, “Utdrive: driver behavior andspeech interactive systems for in-vehicle

environments,” in Intelligent Vehicles Symposium,2007 IEEE, pp. 566–569, IEEE, 2007.

[6] M. Kutila, M. Jokela, G. Markkula, and M. R. Rue,“Driver distraction detection with a camera visionsystem,” in Image Processing, 2007. ICIP 2007. IEEEInternational Conference on, vol. 6, pp. VI–201, IEEE,2007.

[7] M. Sodhi, B. Reimer, and I. Llamazares, “Glanceanalysis of driver eye movements to evaluatedistraction,” Behavior Research Methods, Instruments,& Computers, vol. 34, no. 4, pp. 529–538, 2002.

[8] L. Hanninen and M. Pastell, “Cowlog: Open-sourcesoftware for coding behaviors from digital video,”Behavior Research Methods, vol. 41, no. 2,pp. 472–476, 2009.

[9] T. Harris, “Support vector machines.” File:svmInputSpace2.gif. Accessed: May 2016.

[10] Sideswipe, “Random forest.” File:20131204173330.png. Accessed: May 2016.

[11] Formfonts, “Disco3 plan 2d model.” File:land-rover-discovery-plan-view disco3 plan.jpg.Accessed: May 2016.