Refine
Document Type
- Conference Proceeding (17) (remove)
Conference Type
- Konferenzartikel (16)
- Konferenz-Abstract (1)
Has Fulltext
- no (17)
Is part of the Bibliography
- yes (17) (remove)
Keywords
- Kalman Filter (2)
- Air Pollution (1)
- EKF-SLAM (1)
- Environmental monitoring (1)
- Inertial (1)
- Monocular Depth Estimation (1)
- Optimization (1)
- RoboCup (1)
- autonomous systems (1)
- camera-based navigation (1)
Institute
- Fakultät Elektrotechnik, Medizintechnik und Informatik (EMI) (ab 04/2019) (17) (remove)
Open Access
- Closed Access (8)
- Open Access (6)
- Bronze (3)
- Closed (3)
This study focuses on the autonomous navigation and mapping of indoor environments using a drone equipped only with a monocular camera and height measurement sensors. A visual SLAM algorithm was employed to generate a preliminary map of the environment and to determine the drone's position within the map. A deep neural network was utilized to generate a depth image from the monocular camera's input, which was subsequently transformed into a point cloud to be projected into the map. By aligning the depth point cloud with the map, 3D occupancy grid maps were constructed by using ray tracing techniques to get a precise depiction of obstacles and the surroundings. Due to the absence of IMU data from the low-cost drone for the SLAM algorithm, the created maps are inherently unscaled. However, preliminary tests with relative navigation in unscaled maps have revealed potential accuracy issues, which can only be overcome by incorporating additional information from the given sensors for scale estimation.
Modern industrial production is heavily dependent on efficient workflow processes and automation. The steady flow of raw materials as well as the separation of vital parts and semi-finished products are at the core of these automated procedures. Commonly used systems for this work are bowl feeders, which separate the parts and material by a combination of mechanical vibration and friction. The production of these tools, especially the design of the ramping spiral, is delicate and time-consuming work, as the shape, slope, and material must be carefully adjusted for the corresponding parts. In this work, we propose an automated approach, making use of optimization procedures from artificial intelligence, to design the spiral ramps of the bowl feeders. Therefore, the whole system and considered parts are physically simulated and the optimized geometry is subsequently exported into a CAD system for the actual building, respectively printing. The employment of evolutionary optimization gives the need to develop a mathematical model for the whole setup and find an efficient representation of integral features.
The precise positioning of mobile systems is a prerequisite for any autonomous behavior, in an industrial environment as well as for field robotics. The paper describes the set up for an experimental platform and its use for the evaluation of simultaneous localization and mapping (SLAM) algorithms. Two approaches are compared. First, a local method based on point cloud matching and integration of inertial measurement units is evaluated. Subsequent matching makes it possible to create a three-dimensional point cloud that can be used as a map in subsequent runs. The second approach is a full SLAM algorithm, based on graph relaxation models, incorporating the full sensor suite of odometry, inertial sensors, and 3D laser scan data.
Die Positionierung mobiler Systeme mit hoher Genauigkeit ist eine Voraussetzung für intelligentes autonomes Verhalten, sowohl in der Feldrobotik als auch in industriellen Umgebungen. Dieser Beitrag beschreibt den Aufbau einer Roboterplattform und ihre Verwendung für den Test und die Bewertung von Kalman-Filter-Konfigurationen. Der Aufbau wurde mit einem mobilen Roboter Husky A200 und einem LiDAR-Sensor (Light Detection and Ranging) realisiert. Zur Verifizierung des vorgeschlagenen Aufbaus wurden fünf verschiedene Szenarien ausgearbeitet. Mit denen wurden die Filter auf ihre Leistungsfähigkeit hinsichtlich der Genauigkeit der Positionsbestimmung getestet.
Evaluierung von Kalman Filter Konfigurationen zur Roboterlokaliserung mittels Sensordatenfusion
(2023)
In dieser Arbeit werden drei verschiedene Konfigurationen der von Tom Moore, für das Robot Operating System, entwickelte Kalman-Filter vorgestellt. Diese bilden die Grundlage für eine Lokalisierung mittels Sensorfusion in dem verwendeten ROS-Framework. Ziel dieser Arbeit ist der Aufbau und die Verifikation einer Lokalisierung für ein mobiles Robotersystem Husky A200 der Firma Clearpath Robotics. Hierzu wurden die Möglichkeiten des bestehenden Systems untersucht und mehrere Versionen von Lokalisierungsfiltern konfiguriert. Am an Ende, wird eine Verifikation der Ergebnisse in verschiedenen Szenarien gegeneinandergestellt. Hierzu werden die Ergebnisse einer Variante des Extended Kalman-Filters in 2D (EKF2D), eine Variante des Unscented Kalman-Filter in 2D (UKF2D) und eine Variante des Extended Kalman-Filters in 3D (EKF3D) verifiziert und verglichen. Die Untersuchungen ergaben das der EKF2D die besten und robustesten Ergebnisse für eine Lokalisierung erbringt, trotz, im Vergleich zu der UKF2D Variante, 17,3 % höhere Endpositionsabweichung aufweist. Die in diesem Projekt gewählte EKF3D Konfigurationsvariante eignet sich, wegen seinen starken Ungenauigkeiten in der Höhenbestimmung nicht für eine aussagekräftige Positionsbestimmung.
Mit der Implementierung sowie einer anschließenden aussagekräftigen Evaluierung, soll das, visuelle-inertiale Kartierungs- und Lokalisierungssystem maplab analysiert werden. Hierbei basiert die Kartierung bzw. Lokalisierung auf der Detektion von Umgebungsmerkmalen. Neben der Möglichkeit der Kartenerstellung besteht ferner die Option, mehrere Karten zu fusionieren und somit weitreichende Gebiete zu kartieren sowie für weitere Datenauswertungen zu nutzen. Aufgrund der Durchführung und Bewertung der Ergebnisse in unterschiedlichen Anwendungsszenarien zeigt sich, dass maplab besonders zur Kartierung von Räumen bzw. kleinen Gebäudekomplexen geeignet ist. Die Möglichkeit der Kartenfusionierung bietet weiterhin die Option, den Informationsgehalt von Karten zu erhöhen, welches die Effektivität für eine anschließende Lokalisierung steigert. Bei wachsender Kartierungsgröße hingegen zeigt sich jedoch eine Vergrößerung geometrischer Inkonsistenzen.
A novel approach for synchronization and calibration of a camera and an inertial measurement unit (IMU) in the research-oriented visual-inertial mapping-and localization-framework maplab is presented. Mapping and localization are based on detecting different features in the environment. In addition to the possibility of creating single-case maps, the included algorithms allow merging maps to increase mapping accuracy and obtain large-scale maps. Furthermore, the algorithms can be used to optimize the collected data. The preliminary results show that after appropriate calibration and synchronization maplab can be used efficiently for mapping, especially in rooms and small building environments.
The visual-inertial mapping and localization system maplab is analyzed by its implementation and subsequent evaluation. The mapping or localization is based on environmental feature detection. In addition to creating maps, there is also the option of fusion of several maps and thus mapping extensive areas and using them for further analysis of data. In this way, various software tools can be used to optimize the existing data sets.
Two sensor components are needed: an inertial measuring unit (IMU) and a monochrome camera, which are combined by a hardware rig and put into operation for the analysis of the visual-inertial system. System calibration is crucial for precision and system functioning and is based on nonlinear dynamic state estimation. This ensures the best possible estimate of the position of the environmental feature and the map. Maplab is particularly suitable for mapping rooms or small building complexes as the implementation and evaluation of the results in different application scenarios show. Special emphasis is laid on the evaluation of larger scenarios, in which is shown, that the system is struggling to keep up geometric consistencies and thus provide an accurate map.
Object Detection and Mapping with Unmanned Aerial Vehicles Using Convolutional Neural Networks
(2021)
Significant progress has been made in the field of deep learning through intensive research over the last decade. So-called convolutional neural networks are an essential component of this research. In this type of neural network, the mathematical convolution operator is used to extract characteristics or anomalies. The purpose of this work is to investigate the extent to which it is possible in certain initial settings to input aerial recordings and flight data of Unmanned Aerial Vehicles (UAVs) in the architecture of a neural network and to detect and map an object. Using the calculated contours or dimensions of the so-called bounding boxes, the position of the objects can be determined relative to the current UAV location.
The paper describes a systematic approach for a precise short-time cloud coverage prediction based on an optical system. We present a distinct pre-processing stage that uses a model based clear sky simulation to enhance the cloud segmentation in the images. The images are based on a sky imager system with fish-eye lens optic to cover a maximum area. After a calibration step, the image is rectified to enable linear prediction of cloud movement. In a subsequent step, the clear sky model is estimated on actual high dynamic range images and combined with a threshold based approach to segment clouds from sky. In the final stage, a multi hypothesis linear tracking framework estimates cloud movement, velocity and possible coverage of a given photovoltaic power station. We employ a Kalman filter framework that efficiently operates on the rectified images. The evaluation on real world data suggests high coverage prediction accuracy above 75%.