IMLA - Institute for Machine Learning and Analytics
Refine
Document Type
- Conference Proceeding (68)
- Article (unreviewed) (19)
- Article (reviewed) (15)
- Part of a Book (3)
- Report (3)
- Book (2)
- Contribution to a Periodical (2)
- Doctoral Thesis (1)
Conference Type
- Konferenzartikel (64)
- Konferenz-Abstract (1)
- Konferenz-Poster (1)
- Konferenzband (1)
- Sonstiges (1)
Language
- English (103)
- German (9)
- Multiple languages (1)
Is part of the Bibliography
- yes (113)
Keywords
- Deep learning (16)
- Deep Leaning (12)
- Künstliche Intelligenz (11)
- Machine Learning (11)
- Artificial Intelligence (7)
- Deep Learning (5)
- Robustness (4)
- image classification (4)
- Data Science (3)
- Deep Reinforcement Learning (3)
Institute
Open Access
- Open Access (82)
- Bronze (25)
- Diamond (23)
- Closed (18)
- Closed Access (12)
- Gold (5)
- Hybrid (4)
- Grün (3)
Advancing digitization has greatly increased the need for usable and secure identity management solutions in e-commerce. Digital identity wallets offer a promising solution by allowing users to securely store their personal data at one place on their smartphones and share it on demand in a self-determined way. In order to ensure the acceptance of digital identity wallets, the digital sovereignty of various stakeholders must be safeguarded. This not only promotes trust in digital technologies and services, but also protects against misuse and unauthorized access, which is crucial for maintaining privacy and security in the digital world. This article describes the elicitation of requirements in the area of digital sovereignty when integrating digital identity wallets into e-commerce. In particular, the requirements of three stakeholder groups—online shoppers, online shop operators, and intermediaries—were analyzed. For this purpose, the study used the MERDigS method for eliciting requirements in the area of digital sovereignty. As a result, 31 unique requirements through workshops and 16 unique requirements through interviews were elicited and grouped into seven key requirements. In addition, the quality of the method results and the method feasibility were evaluated using various evaluation criteria.
VisualTorch is a library designed for visualizing neural network architectures in PyTorch. It offers support for multiple visualization styles, such as layered-style, graph-style, and the newly added LeNet-like visualization. When provided with a sequential or custom PyTorch model, alongside the input shape and visualization specifications, VisualTorch automatically translates the model structure into an architectural diagram. The resulting diagram can be refined using various configurations, including style, color, opacity, size, and a legend. VisualTorch is particularly valuable for projects involving PyTorch-based neural networks. By facilitating the generation of graphics with a single function call, it streamlines the process of visualizing neural network architectures. This ensures that the produced results are suitable for publication with minimal additional modifications. Moreover, owing to its diverse customization options, VisualTorch empowers users to generate polished figures suitable for publication.
In this paper, we describe a first publicly available fine-grained product recognition dataset based on leaflet images. Using advertisement leaflets, collected over several years from different European retailers, we provide a total of 41.6k manually annotated product images in 832 classes. Further, we investigate three different approaches for this fine-grained product classification task, Classification by Image, Classification by Text, as well as Classification by Image and Text. The last both approaches use the text extracted directly from the leaflet product images. We show, that the combination of image and text as input improves the classification of visual difficult to distinguish products. The final model leads to an accuracy of 96.4% with a Top-3 score of 99.2%.
https://github.com/ladwigd/Leaflet-Product-Classification
Das Forschungsprojekt KINLI (Künstliche Intelligenz für Nachhaltige Lebensmittelqualität in Lieferketten) zielt durch den Einsatz von Künstlicher Intelligenz (KI) im Bereich der Aufzucht und Mast von Nutztieren sowie in der Verarbeitung von Fleisch auf eine nachhaltige Gestaltung von Lebensmittellieferketten ab. Neben der optimierten Gesundheit und dem Wohl von Nutztieren sollen auch Lebensmittelqualität und -sicherheit durch KI prädiktiv gefördert werden.
This study applies semi-supervised learning to automate the differen-tiation of mold colonies, thereby reducing the time and cost associated with airquality assessments. EfficientNet V2 and Normalization-Free Net (NfNet) weretrained on a dataset of mold colony images, created in a semi-supervised way.NfNet demonstrated superior performance, particularly on non-padded images,with explainable AI techniques enhancing interpretability. The models exhibitedgeneralization capabilities to environmental samples, indicating the potential forautomating mold identification and streamlining air quality monitoring, therebyreducing manual effort and costs. Future work will focus on refining species han-dling and integrating the system into laboratory workflows.
Urban geothermal energy production plays a critical role in achieving global climate objectives. However, drilling operations in densely populated areas generate significant noise pollution, posing challenges to community acceptance and regulatory compliance. This research presents an artificial intelligence-driven approach to dynamically reduce noise emissions during geothermal drilling. We integrate Deep Reinforcement Learning (DRL) with generative neural network models to provide real-time recommendations for optimal drilling parameters. Specifically, the Drill-LSTM model forecasts future machine states, while the Sound-GAN framework predicts sound propagation based on varying operational conditions. These models feed into a DRL-Agent that learns to balance drilling efficiency with noise minimization. Additionally, an interactive assistance system GUI presents predictions, forecasts, and recommendations to human operators, facilitating informed decision-making. Our system demonstrates significant potential in reducing noise levels, enhancing operational efficiency, and fostering greater acceptance of urban geothermal projects. Future work will focus on refining the models and validating the system in real-world drilling scenarios.
We have developed a methodology for the systematic generation of a large image dataset of macerated wood references, which we used to generate image data for nine hardwood genera. This is the basis for a substantial approach to automate, for the first time, the identification of hardwood species in microscopic images of fibrous materials by deep learning. Our methodology includes a flexible pipeline for easy annotation of vessel elements. We compare the performance of different neural network architectures and hyperparameters. Our proposed method performs similarly well to human experts. In the future, this will improve controls on global wood fiber product flows to protect forests.
Wood species identification plays a crucial role in various industries, from ensuring the legality of timber products to advancing ecological conservation efforts. This paper introduces WoodYOLO, a novel object detection algorithm specifically designed for microscopic wood fiber analysis. Our approach adapts the YOLO architecture to address the challenges posed by large, high-resolution microscopy images and the need for high recall in localization of the cell type of interest (vessel elements). Our results show that WoodYOLO significantly outperforms state-of-the-art models, achieving performance gains of 12.9% and 6.5% in F2 score over YOLOv10 and YOLOv7, respectively. This improvement in automated wood cell type localization capabilities contributes to enhancing regulatory compliance, supporting sustainable forestry practices, and promoting biodiversity conservation efforts globally.
Recent work in neural networks for image classification has seen a strong tendency towards increasing the spatial context during encoding. Whether achieved through large convolution kernels or self-attention, models scale poorly with the increased spatial context, such that the improved model accuracy often comes at significant costs. In this paper, we propose a module for studying the effective filter size of convolutional neural networks (CNNs). To facilitate such a study, several challenges need to be addressed: (i) we need an effective means to train models with large filters (potentially as large as the input data) without increasing the number of learnable parameters, (ii) the employed convolution operation should be a plug-and-play module that can replace conventional convolutions in a CNN and allow for an efficient implementation in current frameworks, (iii) the study of filter sizes has to be decoupled from other aspects such as the network width or the number of learnable parameters, and (iv) the cost of the convolution operation itself has to remain manageable i.e. we can not naïvely increase the size of the convolution kernel. To address these challenges, we propose to learn the frequency representations of filter weights as neural implicit functions, such that the better scalability of the convolution in the frequency domain can be leveraged. Additionally, due to the implementation of the proposed neural implicit function, even large and expressive spatial filters can be parameterized by only a few learnable weights. Interestingly, our analysis shows that, although the proposed networks could learn very large convolution kernels, the learned filters are well localized and relatively small in practice when transformed from the frequency to the spatial domain. We anticipate that our analysis of individually optimized filter sizes will allow for more efficient, yet effective, models in the future.