TY - JOUR
T1 - Multi-task Deep Learning with Optical Flow Features for Self-Driving Cars
AU - Hu, Yuan
AU - Shum, Hubert
AU - Ho, Edmond S. L.
N1 - Funding information:
The project was supported in part by the Royal Society (Ref: IES\R2\181024 and IES\R1\191147).
PY - 2020/12/1
Y1 - 2020/12/1
N2 - The control of self‐driving cars has received growing attention recently. Although existing research shows promising results in the vehicle control using video from a monocular dash camera, there has been very limited work on directly learning vehicle control from motion‐based cues. Such cues are powerful features for visual representations, as they encode the per‐pixel movement between two consecutive images, allowing a system to effectively map the features into the control signal. The authors propose a new framework that exploits the use of a motion‐based feature known as optical flow extracted from the dash camera and demonstrates that such a feature is effective in significantly improving the accuracy of the control signals. The proposed framework involves two main components. The flow predictor, as a self‐supervised deep network, models the underlying scene structure from consecutive frames and generates the optical flow. The controller, as a supervised multi‐task deep network, predicts both steer angle and speed. The authors demonstrate that the proposed framework using the optical flow features can effectively predict control signals from a dash camera video. Using the Cityscapes data set, the authors validate that the system prediction has errors as low as 0.0130 rad/s on steer angle and 0.0615 m/s on speed, outperforming existing research.
AB - The control of self‐driving cars has received growing attention recently. Although existing research shows promising results in the vehicle control using video from a monocular dash camera, there has been very limited work on directly learning vehicle control from motion‐based cues. Such cues are powerful features for visual representations, as they encode the per‐pixel movement between two consecutive images, allowing a system to effectively map the features into the control signal. The authors propose a new framework that exploits the use of a motion‐based feature known as optical flow extracted from the dash camera and demonstrates that such a feature is effective in significantly improving the accuracy of the control signals. The proposed framework involves two main components. The flow predictor, as a self‐supervised deep network, models the underlying scene structure from consecutive frames and generates the optical flow. The controller, as a supervised multi‐task deep network, predicts both steer angle and speed. The authors demonstrate that the proposed framework using the optical flow features can effectively predict control signals from a dash camera video. Using the Cityscapes data set, the authors validate that the system prediction has errors as low as 0.0130 rad/s on steer angle and 0.0615 m/s on speed, outperforming existing research.
U2 - 10.1049/iet-its.2020.0439
DO - 10.1049/iet-its.2020.0439
M3 - Article
SN - 1751-956X
VL - 14
SP - 1845
EP - 1854
JO - IET Intelligent Transport Systems
JF - IET Intelligent Transport Systems
IS - 13
ER -