Upload
buicong
View
216
Download
0
Embed Size (px)
Citation preview
1
Vision-based systems for driver
monitoring and de-identification
Human Sensing
Fernando De la Torre
Outline
• Driver distraction detection
• System
• Evaluation
• Video de-identification
• System
• Preliminary evaluation
• Conclusions
Outline
• Driver distraction detection
• System
• Evaluation
• Video de-identification
• System
• Preliminary evaluation
• Conclusions
Distracted driving has become a large problem due to the increasing use of cell phones
National Safety Council: 23% of all crashes involve cell phone use
Amount of drivers texting while driving:
• U.K: 45% at least one time
• Australia: 17 % regularly
• United States: 27% adults once at least
Related problems:
• Amount of time with Eyes-Off-The-Road (EOTR) increase by up to 400%
• 28% more lane excursions and 140% more incorrect lane changes
• Chances of a crash increase by up to 8 to 23 times
Recent study in U.K. concludes that
Texting while driving could be more dangerous than driving drunk
“Mobile Phone: A growing problem of driver distraction”, World Health Organization (2011)
Problem of texting while driving
5
Goals:
� Driver distraction detection, e.g.,
Eyes-Off-Road (EOR)
Challenges
� Low cost camera
� Pre-defined camera location
� Abrupt change in pose & illumination
� Night and day
� Real-time gaze and head pose
tracking
� Glasses support
6
Goals and Challenges
Goals:
� Driver distraction detection, e.g.,
Eyes-Off-Road (EOR)
Challenges
� Low cost camera
� Pre-defined camera location
� Abrupt change in pose & illumination
� Night and day
� Real-time gaze and head pose
tracking
� Glasses support
7
Webcam $60
Goals and Challenges
Goals:
� Driver distraction detection, e.g.,
Eyes-Off-Road (EOR)
Challenges
� Low cost camera
� Pre-defined camera location
� Abrupt change in pose & illumination
� Night and day
� Real-time gaze and head pose
tracking
� Glasses support
8
Goals and Challenges
Goals:
� Driver distraction detection, e.g.,
Eyes-Off-Road (EOR)
Challenges
� Low cost camera
� Pre-defined camera location
� Abrupt change in pose & illumination
� Night and day
� Real-time gaze and head pose
tracking
� Glasses support
9
Goals and Challenges
Goals:
� Driver distraction detection, e.g.,
Eyes-Off-Road (EOR)
Challenges
� Low cost camera
� Pre-defined camera location
� Abrupt change in pose & illumination
� Night and day
� Real-time gaze and head pose
tracking
� Glasses support
10
Goals and Challenges
Goals:
� Driver distraction detection, e.g.,
Eyes-Off-Road (EOR)
Challenges
� Low cost camera
� Pre-defined camera location
� Abrupt change in pose & illumination
� Night and day
� Real-time gaze and head pose
tracking
� Glasses support
11
Goals and Challenges
Goals:
� Driver distraction detection, e.g.,
Eyes-Off-Road (EOR)
Challenges
� Low cost camera
� Pre-defined camera location
� Abrupt change in pose & illumination
� Night and day
� Real-time gaze and head pose
tracking
� Glasses support
12
Goals and Challenges
13
EOR detection
from head pose Prediction
System Overview
driver imagefacial landmark
trackinghead pose gaze
Estimation
14
Eyes Off the Road (Ray Tracing)
Eyes Off the Road Estimation:
The EOR estimation is based on a 3D ray tracing method that uses the geometry of the scene.
Outline
• Driver distraction detection
• System
• Evaluation
• Video de-identification
• System
• Preliminary evaluation
• Conclusions
16
Results
VTTI has developed a protocol to evaluate the performance of the system:
• Extended Off-Road glances (2 sec)
• Brief check glances
Locations of interest selected:
Outline
• Driver distraction detection
• System
• Evaluation
• Video de-identification
• System
• Preliminary evaluation
• Conclusions
Project Scope• Video De-identification in the Automobile Environment
• Remove person-specific facial features
• Preserve head pose, gaze and facial expression
Outline
• Driver distraction detection
• System
• Evaluation
• Video de-identification
• System
• Preliminary evaluation
• Conclusions
Evaluate identity protection
• Face recognition experiments on the images in CMU Multi-PIE
database (multipie.org) and frames in NDS videos.
• FAT produced the lowest recognition accuracies: best identity protection.
• An average face was used as mask, no distinctive facial features.
• Face recognition classifiers were trained and tested in the same image sets
or videos, an unlikely case in real applications.
Results:
Facial Action Unit Coding
Nose Wrinkler (AU9)Upper lid raiser (AU5) Lip Tightener (AU23)Outer Brow Raiser (AU2)
Evaluate data utility
• Head pose estimation experiments on the images in CMU Multi-
PIE database (multipie.org) and frames in NDS videos
• Facial Action Unit recognition on frames in NDS videos
• FAT preserved the most recognizable AU#12 in de-identified frames.
• Tracker needs to be improved for landmarks on the mouth.
• The head pose angles were preserved in the FAT de-identified
frames
Results:
Conclusions
• It is a difficult problem
– Low spatial and temporal resolution
– Noisy camera (specially at night)
– Large changes in pose
• Preliminary results are encouraging, especially during the day
• Currently improving
– Tracking at night
– Stabilizing tracking in mouth and eye region