It is well known that drivers who talk on their phones become distracted, putting themselves and others at risk. Despite legislation, people still use their phones while driving, often feeling that they are in control of the situation. Our solution is to monitor the drivers’ speech and other information for evidence that they are distracted and then take appropriate action. Using a driving simulator, the Distracted Driving Project has gathered speech and driving data in conditions that vary in their degree of cognitive load.The speech and driving data are time-synced so that coincidences of events in speech and high cognitive load in driving can be modeled.With funding from Yahoo!, we have mined this data to build a model and, from that, a set of algorithms to create a first distraction detector. This detector listens to the driver conversing on the phone and, from information present in the speech signal alone, detects when the driver is becoming distracted. In order to increase robustness and avoid false alarms, the detector needs to model a larger set of variables, which include information about the driving environment. For this, we will add information from the accelerometer, and other non-speech audio signals.From this, we will create anew model and corresponding set of algorithms. This new robust detector is intended to make everyday use of a distraction detector by the general public much more acceptable. We deliberately do not include any of the car system sensors in our detection algorithm since integration of phone and car systems will often not be available. This is especially the case for younger drivers with older model cars. The second step in gaining drivers’ trust is to determine what the system should do when it decides that the driver is distracted. A human passenger conversing with a driver can be aware of the difficulty of the driving situation and modify their discussion with the driver appropriately. We will investigate how a phone app can be driving-situation aware and reactive in a similar way. Potential strategies are simply stopping the app, or explicitly indicating that it is going to stop through explicit speech or implicit signaling.We will identify the most effective and graceful methods for a system to disengage from the dialog.
This past year, the Distracted Driving Project took the first steps in modeling distraction by creating a speech database that can be used to model normal and distracted driver speech. We set up a driving simulator connected to a Wizard of Oz setup that imitated a phone app reading email to a driver. The driving conditions varied in difficulty from an easy straightaway to soft curves to tight curves. At the same time, the emails being read to the driver varied from easy to deal with (Mom asks how you are) to more challenging, requiring more of the driver’s cognitive load (John wants to know five things you would like to get for your birthday, he’ll choose one). Fifty drivers will have been recorded by the end of 2014. A recorded sample of low cognitive load on a straightaway can be found at https://dialrc.org/tim_control. A recorded sample of high cognitive load in a tight curve can be found at https://dialrc.org/tim_hesit. The data we have collected so far clearly shows detectable different speech behavior in easier and more difficult driving situations. The data, with funding from Yahoo!, is being modeled to represent control and distracted conditions uniquely from information in the speech signal time synced to the driving conditions. For this we a reusing automatic speech processing tools such as OpenSmile. Based on these models, we have begun to derive detection algorithms that are performing correctly on the data we have collected so far. While performance (precision in distraction detection based on speech) has been our first step, in order to have a detection system that drivers will actually accept to use takes two more steps. First we need to make the system more robust so that false alarms are minimized. The second is to determine what action is appropriate for the system to take when it detects distraction.This is the goal for our UTCFY2015 project. In order to make the system more robust, we will complement the speech signal information with non-speech data. Accelerometer information can be obtained from any smartphone. Sudden accelerations decelerations may be indicators of a dangerous driving condition. These may take place while a driver is speaking, at a time they are not speaking, and/or before distraction shows up in the driver’s speech. Other types of information such as GPScan also help increase the detector’s robustness.GPS information can be used to tell us the car’s speed and location relative to some areas where driving is difficult. We will record and several types of signals such as these and include them in our models and algorithms, finally testing them for relative effectiveness. This will entail collecting a new dataset from both the driving simulator we have already set up and aboard real vehicles(in controlled conditions that do not present a danger to the driver). We will test whether the addition of each of the signals together and individually contributes to a decreased false positive rate and modify our algorithms accordingly.This will make our system more robust to false alarms, a strong first step in gaining drivers’ trust and thus adoption of the system. The second step concerns system interaction with the driver. Here we will discover how to gracefully intervene when distraction is detected. Not all types of intervention that we create will be acceptable to the average driver. In our Yahoo! project, we are creating a system that reads the driver’s email to them(they can dictate answers to the mail as well). The goal here is to have the detection system gracefully shut down the mail app. That is, the system stops reading mail or listening to the driver dictate mail and saves the last few actions that happened in the dialog so that the driver can come back to exactly what they were doing with no loss of information at a later time. Yet there is no proof that having the system rather than the driver decide when to shut down the application is acceptable to drivers in general This may in fact be more distracting if, for example, the driver ends up paying more attention to wondering what happened to the app. There are several possible actions that we can set up in a smartphone-based system/app that can be taken when distraction is detected: - the app can simply be shut down gracefully; - the system can activate some warning alarm; - the system can send one or more successive messages to the driver either aurally visually or in some tactile form (vibrating steering wheel, for example). The first solution involves a system-driven decision about what to do and when. The two other solutions give the driver the choice. We will test the acceptability of several versions of each of these options. While acceptability is the main concern, we will also measure whether each option tested adds to the driver’s cognitive load, thus distracting them further. We will also investigate whether we should offer only one solution for all drivers or if the driver should be given the choice of the type of warning they prefer. Finally, we will test these options for acceptability and added cognitive load both with our driving simulator and in a real vehicle. We will employ one LTI MLT graduate student, Tiancheng Zhao, for this project.
One year.
We have recently met with representatives of Mercedes. While they were very interested in our work, they cannot work with us yet. To partner with us, they need proof that the detection works robustly and that it is acceptable to drivers. The results of this UTCFY2015 project will enable us to approachcar manufacturers and others to find a means to deploy our system.The distraction detection that we havedevelopedwith Yahoo! specifically for their newsfeed and mailapps should be deployed some time in 2015.Our techniques are specifically designed to use whatever information is available, thus if we only have an audio stream (through the telephone system) we can use that stream to detect distraction and reactappropriately. If we have an app on the driver’s smartphone we can use sensors on the phone itself to aid indetection. Finally,if our system is integratedwith the car's sensors we can also utilize them.
- Two sets of driving data enriched with accelerometer and other information. - A robust distraction detection algorithm. After a series of offline tests, it will be tested on the data we have collected by comparing it to human annotation of the degree of the driver’s distraction and determining whether the system performs similarly to the humans. We will measure the Cohen Kappa of agreement using tenfold cross-validation.The system will also be tested for correct performance in both a simulator and a real vehicle. The driver will be asked to assess system performance. - A system that gracefully alerts drivers that they are distracted. This will be assessed by having drivers use the system with the simulator and in real vehicles to see if the types of alarms are acceptable to them –if they would use them. Fifty drivers will be enlisted for this task.
Name | Affiliation | Role | Position | |
---|---|---|---|---|
awb@cs.cmu.edu | Black, Alan | LTI | Co-PI | Faculty - Research/Systems |
max@cs.cmu.edu | Eskenazi, Maxine | LTI/SCS | PI | Faculty - Research/Systems |
tingyaoh@andrew.cmu.edu | Hu, Ting Yao | MLT | Other | Student - Masters |
tianchez@andrew.cmu.edu | Zhao, Tiancheng | MLT | Other | Student - Masters |
Type | Name | Uploaded |
---|---|---|
Final Report | UTC_project_83_Detecting_Distraction_Final_report.pdf | July 2, 2018, 4:33 a.m. |
No match sources!
No partners!