Cut-In Maneuver Detection With Self-Supervised Contrastive Video Representation Learning
Loading...
Date
2023
Authors
Nalçakan, Yağız
Baştanlar, Yalın
Journal Title
Journal ISSN
Volume Title
Publisher
Springer
Open Access Color
Green Open Access
Yes
OpenAIRE Downloads
OpenAIRE Views
Publicly Funded
No
Abstract
The detection of the maneuvers of the surrounding vehicles is important for autonomous vehicles to act accordingly to avoid possible accidents. This study proposes a framework based on contrastive representation learning to detect potentially dangerous cut-in maneuvers that can happen in front of the ego vehicle. First, the encoder network is trained in a self-supervised fashion with contrastive loss where two augmented videos of the same video clip stay close to each other in the embedding space, while augmentations from different videos stay far apart. Since no maneuver labeling is required in this step, a relatively large dataset can be used. After this self-supervised training, the encoder is fine-tuned with our cut-in/lane-pass labeled datasets. Instead of using original video frames, we simplified the scene by highlighting surrounding vehicles and ego-lane. We have investigated the use of several classification heads, augmentation types, and scene simplification alternatives. The most successful model outperforms the best fully supervised model by ∼ 2% with an accuracy of 92.52%
Description
Keywords
Contrastive representation learning, Driver assistance systems, Vehicle maneuver classification
Fields of Science
Citation
WoS Q
Q3
Scopus Q
Q2

OpenCitations Citation Count
2
Source
Signal Image and Video Processing
Volume
17
Issue
Start Page
2915
End Page
2923
PlumX Metrics
Citations
Scopus : 3
Captures
Mendeley Readers : 5
SCOPUS™ Citations
3
checked on Apr 27, 2026
Page Views
653
checked on Apr 27, 2026
Downloads
53
checked on Apr 27, 2026
Google Scholar™


