Cloud-Anwendung für zeitlich veränderliche Fahrbahnzustandsinformationen für verschiedene Fahrzeugklassen basierend auf fusionierten Fahrzeugdaten von Fahrzeugen verschiedener Klassen (LKW/PKW)
Detection, localization, and classification of road conditions in real-time from the in-vehicle front camera. Estimation of vehicle specific road friction by fusion with information from other sensors (such as vibroacoustic, meteorologic, or traction control responses).
A cloud based service increases traffic safety and efficiency for driving assistance and automated driving applications.
Identifikation dynamik- und sicherheitsrelevanter Trailerzustände für automatisiert fahrende Lastkraftwagen
(Identification of Dynamic and Safety-relevant Trailer States for Automated Moving Trucks)
The dynamic properties of a truck are determined to a large extent by its trailer. Yet the trailer is barely taken into account in current truck sensor systems. The goal of IdenT is the development of an intelligent trailer sensor network and a cloud-based data platform for reliable real-time estimation of the state of trailer components crucial for automated driving applications. Using a backward-facing camera VISCODA provides its project partners with highly accurate visual odometry and a geometrically reconstructed scene for sensor fusion on-board and in the cloud. Through a combination of rule-based (geometric) and data-driven (machine learning) algorithms, VISCODA additionally detects and localizes road users behind the rear of the trailer.
Resource Adaptive Sceneanalysis and -reconstruction (Ressourcenadaptive Szenenanalyse und -rekonstruktion) [2019-2020]
Research project in cooperation with the Institut für Informationsverarbeitung (Leibniz Universität Hannover / LUH) targeting optimal 3D scene reconstruction accuracy and object detection reliability with resource limitations such as hardware specific computation power.
Dieses Projekt wird mit Mitteln des Europäischen Fonds für regionale Entwicklung gefördert.
Multi Camera Multiple Object Tracking [2017-2019]
For the H2020 5G PPP phase 2 project 5GCAR, VISCODA develops algorithms for localization and tracking of vehicles in image sequences using a multi camera setup.
For the trajectory planning in autonomous driving, the accurate localization of the vehicles is required. Accurate localizations of the ego-vehicle will be provided by the next generation of connected cars using 5G. Until all cars participate in the network, un-connected cars have to be considered as well. These cars are localized via static cameras positioned next to the road. The demonstrated scenario is a lane merge where a car on the accelaration lane merges into the traffic on the main lane.
To achieve high accuracy in the vehicle localization, the highly accurate calibration of the cameras is required. The camera based system consists of vehicle detection, localization, and tracking. It provides accurate vehicle desciptions which are used for computing trajectory recommendations for all participating vehicles with the aim of an automatic cooperative maneuver, the lane merge.
- 5GCAR final demonstration
- 5GCAR Pre-demonstration (demo video for MWC 2019)
- Explanation of 5GCAR use cases (demo video for MWC 2018)
K. Cordes and H. Broszio: "Vehicle Lane Merge Visual Benchmark",
International Conference on Pattern Recognition (ICPR), IEEE, Jan. 2021
- K. Antonakoglou, N. Brahmi, T. Abbas, A.E. Fernandez Barciela, M. Boban, K. Cordes, M. Fallgren, L. Gallo, A. Kousaridas, Z. Li, T. Mahmoodi, E. Ström, W. Sun, T. Svensson, G. Vivier, J. Alonso-Zarate: "On the Needs and Requirements Arising from Connected and Automated Driving",
J. Sens. Actuator Netw. 2020, 9(2):24.
OpenAccess: abstract, html, pdf
K. Cordes, N.Nolte, N. Meine, and H. Broszio: "Accuracy Evaluation of Camera-based Vehicle Localization",
International Conference on Connected Vehicles and Expo (ICCVE), IEEE, pp. 1-7, Nov. 2019
Paper at ieeexplore
- "The 5GCAR Demonstrations", Sep. 2019
Deliverable D5.2 (pdf)
- B. Cellarius, K. Cordes, T. Frye, S. Saur, J. Otterbach, M. Lefebvre, F. Gardes, J. Tiphène, M. Fallgren: "Use Case Representations of Connected and Automated Driving",
European Conference on Networks and Communications (EuCNC), June 2019
Extended abstract (pdf)
- K. Cordes and H. Broszio: "Constrained Multi Camera Calibration for Lane Merge Observation",
International Conference on Computer Vision Theory and Applications (VISAPP), SciTePress, pp. 529-536, Feb. 2019
Paper preprint (pdf)
- "5GCAR Demonstration Guidelines", May 2018
Deliverable D5.1 (pdf)
- M. Fallgren, M. Dillinger, A. Servel, Z. Li, B. Villeforceix, T. Abbas, N. Brahmi, P. Cuer, T. Svensson, F. Sanchez, J. Alonso-Zarate, T. Mahmoodi, G. Vivier, M. Narroschke: "On the Fifth Generation Communication Automotive Research and Innovation Project 5GCAR - The Vehicular 5G PPP Phase 2 Project",
European Conference on Networks and Communications (EuCNC), June 2017
Extended abstract (pdf)
Object Motion Estimation [ongoing]
For the estimation of motion models of moving objects in video, a motion segmentation technique is utilized. Motion segmentation is the task of classifying the feature trajectories in an image sequence to different motions. Hypergraph based approaches use a specific graph to incorporate higher order similarities for the estimation of motion clusters. They follow the concept of hypothesis generation and validation.
Our approach uses a simple but effective model for incorporating motion-coherent affinities. The hypotheses generated from the resulting hypergraph lead to a significant decrease of the segmentation error.
- K. Cordes, C. Ray’onaldo, H. Broszio: "Motion-Coherent Affinities for Hypergraph Based Motion Segmentation",
International Conference on Computer Analysis of Images and Patterns (CAIP), Springer LNCS volume 10424, pp. 121-132, August 2017
Paper preprint (pdf)
Presentation slides (pdf)
Scene Reconstruction from Video [ongoing]
For the products VooCAT/CineCAT, various algorithms for camera calibration, tracking, structure from motion, and video segmentation are developed. The basis for the scene estimation is the usage of corresponding image features which arise from a 3D structure being mapped to diﬀerent camera image planes. By using a statistical error model which describes the errors in the position of the detected feature points, a Maximum Likelihood estimator can be formulated that simultaneously estimates the camera parameters and the 3D positions of the image features.
The camera path, the video segmentation, and the reconstructed scene are essential for the integration virtual objects into the video. The point cloud is used for 3D measurements, such as distances to or between different objects of the scene.
- M. Narroschke, N. Meine, H. Broszio, M. Korzec: "Verfahren zur Erfassung von im Fahrzeugumfeld befindlichen Bordsteinen", DE102016215840A1, 2018-03-01
- H. Broszio, D. Liebehenschel, C. Last, M. Korzec, K. Cordes: "Verfahren zur Eigenschattenerkennung", DE102016216462A1, 2018-03-01
- N. Meine, A. Haja, H.Broszio: "Method for continuous estimation of driving surface plane of motor vehicle", DE102011118171A1, 2013-05-16
- N. Meine, A. Haja, H.Broszio: "Method for determining e.g. wall in rear area of passenger car during parking", DE102011113099A1, 2013-03-14
- N. Meine, A. Haja, H.Broszio: "Method for determining angle between towing vehicle and trailer", DE102011113197A1, 2013-03-14
- N. Meine, A. Haja, H.Broszio: "Method for detecting obstacle in car surroundings", DE102010009620A1, 2011-09-01