Por favor, use este identificador para citar o enlazar este ítem:https://uvadoc.uva.es/handle/10324/68168
Título
A Self-Adaptive Automatic Incident Detection System for Road Surveillance Based on Deep Learning
Autor
Año del Documento
2024
Documento Fuente
Sensors 2024, 24, 1822
Abstract
first_page
settings
Order Article Reprints
Open AccessArticle
A Self-Adaptive Automatic Incident Detection System for Road Surveillance Based on Deep Learning
by César Bartolomé-Hornillos
1, Luis M. San-José-Revuelta
1 [ORCID] , Javier M. Aguiar-Pérez
1 [ORCID] , Carlos García-Serrada
2, Eduardo Vara-Pazos
2 and Pablo Casaseca-de-la-Higuera
1,* [ORCID]
1
ETSI Telecomunicación, Universidad de Valladolid, 47011 Valladolid, Spain
2
Construcciones y Obras Llorente, S.A., 47012 Valladolid, Spain
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(6), 1822; https://doi.org/10.3390/s24061822
Submission received: 7 February 2024 / Revised: 2 March 2024 / Accepted: 8 March 2024 / Published: 12 March 2024
(This article belongs to the Section Vehicular Sensing)
Download
keyboard_arrow_down
Browse Figures
Versions Notes
Abstract
We present an automatic road incident detector characterised by a low computational complexity for easy implementation in affordable devices, automatic adaptability to changes in scenery and road conditions, and automatic detection of the most common incidents (vehicles with abnormal speed, pedestrians or objects falling on the road, vehicles stopped on the shoulder, and detection of kamikaze vehicles). To achieve these goals, different tasks have been addressed: lane segmentation, identification of traffic directions, and elimination of unnecessary objects in the foreground. The proposed system has been tested on a collection of videos recorded in real scenarios with real traffic, including areas with different lighting. Self-adaptability (plug and play) to different scenarios has been tested using videos with significant scene changes. The achieved system can process a minimum of 80 video frames within the camera’s field of view, covering a distance of 400 m, all within a span of 12 s. This capability ensures that vehicles travelling at speeds of 120 km/h are seamlessly detected with more than enough margin. Additionally, our analysis has revealed a substantial improvement in incident detection with respect to previous approaches. Specifically, an increase in accuracy of 2–5% in automatic mode and 2–7% in semi-automatic mode. The proposed classifier module only needs 2.3 MBytes of GPU to carry out the inference, thus allowing implementation in low-cost devices.
Revisión por pares
SI
Patrocinador
This work was supported in part by the European Regional Development Fund (ERDF-FEDER) through the Instituto de Competitividad Empresarial (Junta de Castilla y León, Spain) under Grants (04/18/VA/0008, 04/18/VA/0013, 04/20/VA/0130) and in part by Construcciones y Obras Llorente S.A. The work was also partly supported by the EU Horizon 2020 Research and Innovation Programme under the Marie Sklodowska-Curie grant agreement No 101008297. This article reflects only the authors’ view. The European Union Commission is not responsible for any use that may be made of the information it contains
Idioma
eng
Tipo de versión
info:eu-repo/semantics/publishedVersion
Derechos
openAccess
Collections
Files in this item
Except where otherwise noted, this item's license is described as Attribution-NonCommercial-NoDerivatives 4.0 Internacional