Refine
Document Type
- Bachelor Thesis (1)
- Article (unreviewed) (1)
Keywords
- Maschinelles Sehen (2) (remove)
Institute
Open Access
- Closed (1)
- Diamond (1)
- Open Access (1)
Garbage in, Garbage out: How does ambiguity in data affect state-of-the-art pedestrian detection?
(2024)
This thesis investigates the critical role of data quality in computer vision, particularly in the realm of pedestrian detection. The proliferation of deep learning methods has emphasised the importance of large datasets for model training, while the quality of these datasets is equally crucial. Ambiguity in annotations, arising from factors like mislabelling, inaccurate bounding box geometry and annotator disagreements, poses significant challenges to the reliability and robustness of the pedestrian detection models and their evaluation. This work aims to explore the effects of ambiguous data on model performance with a focus on identifying and separating ambiguous instances, employing an ambiguity measure utilizing annotator estimations of object visibility and identity. Through accurate experimentation and analysis, trade-offs between data cleanliness and representativeness, noise removal and retention of valuable data emerged, elucidating their impact on performance metrics like the log average miss-rate, recall and precision. Furthermore, a strong correlation between ambiguity and occlusion was discovered with higher ambiguity corresponding to greater occlusion prevalence. The EuroCity Persons dataset served as the primary dataset, revealing a significant proportion of ambiguous instances with approximately 8.6% ambiguity in the training dataset and 7.3% in the validation set. Results demonstrated that removing ambiguous data improves the log average miss-rate, particularly by reducing the false positive detections. Augmentation of the training data with samples from neighbouring classes enhanced the recall but diminished precision. Error correction of wrong false positives and false negatives significantly impacts model evaluation results, as evidenced by shifts in the ECP leaderboard rankings. By systematically addressing ambiguity, this thesis lays the foundation for enhancing the reliability of computer vision systems in real-world applications, motivating the prioritisation of developing robust strategies to identify, quantify and address ambiguity.
Zur Erkennung bestimmter Fehler, wie zum Beispiel Ätzfehler, Oberflächen-Fremdkörper, Verschmutzungen, bei der Leiterplattenherstellung ist eine Echtfarben-Bildverarbeitung notwendig, über die hier ein Überblick gegeben wird. Zur Farbsegmentierungs sind verschiedene Methoden geeignet, die einmal im Rot-Grün-Blau-Raum oder im HSI-Raum (Hue, Saturation, Intensity - Färbung, Sättigung, Intensität) untersucht und verglichen werden. Clusterfindungsverfahren und Lookup-Tabelle bereiten Schwierigkeiten bei der Erfassung der Farbkanten, zum Beispiel der Kante zwischen dem Kupfer und dem Beschichtungsmaterial. Hier hilft ein geeigneter Algorithmus, der mit Vorsegmentierung arbeitet. Eine Verfeinerung ist mit Hilfe von Kantenfiltern möglich, zum Beispiel das Color-Sobel-Magnitude-Filter.