Abstract
The global personal mobility market has been rapidly expanding due to its convenience. However, the increasing number of accidents involving personal mobility devices has become a growing concern, including falls, collisions with objects, and riders being struck by moving vehicles or objects. In this paper, we propose a deep learning-based safe driving system that considers both user and road images to address this issue. Our system employs CNN-based models to detect whether the user is 1) wearing a helmet and 2) looking ahead in user-side images. At the same time, the roadside image recognizes whether the user is 3) driving on the sidewalk and 4) near the intersection. These tasks are simultaneously performed in parallel to identify the overall situation, which is done by determining the final speed as the minimum speed of speed values extracted from all tasks. Additionally, we employ knowledge distillation techniques to compress the models and enable real-time inference on edge devices, resulting in a fast and accurate system that is well-suited to the characteristics of personal mobility.
Original language | English |
---|---|
Title of host publication | 2023 20th International Conference on Ubiquitous Robots, UR 2023 |
Publisher | Institute of Electrical and Electronics Engineers Inc. |
Pages | 217-222 |
Number of pages | 6 |
ISBN (Electronic) | 9798350335170 |
DOIs | |
State | Published - 2023 |
Event | 20th International Conference on Ubiquitous Robots, UR 2023 - Honolulu, United States Duration: 25 Jun 2023 → 28 Jun 2023 |
Publication series
Name | 2023 20th International Conference on Ubiquitous Robots, UR 2023 |
---|
Conference
Conference | 20th International Conference on Ubiquitous Robots, UR 2023 |
---|---|
Country/Territory | United States |
City | Honolulu |
Period | 25/06/23 → 28/06/23 |
Bibliographical note
Publisher Copyright:© 2023 IEEE.