Refine
Year of publication
- 2023 (115)
- 2022 (106)
- 2017 (92)
- 2019 (86)
- 2021 (85)
- 2018 (84)
- 2014 (81)
- 2016 (79)
- 2015 (77)
- 2020 (72)
- 2013 (59)
- 2011 (43)
- 2012 (42)
- 2010 (33)
- 2009 (25)
- 2007 (20)
- 2008 (19)
- 2005 (18)
- 2006 (15)
- 2004 (13)
- 1995 (10)
- 2001 (10)
- 2003 (7)
- 1991 (6)
- 1993 (6)
- 1997 (6)
- 2002 (6)
- 1990 (5)
- 1992 (5)
- 1994 (5)
- 1996 (5)
- 1998 (5)
- 2000 (4)
- 1989 (3)
- 1999 (3)
- 1986 (1)
- 1987 (1)
- 1988 (1)
Document Type
- Conference Proceeding (1253) (remove)
Conference Type
- Konferenzartikel (950)
- Konferenz-Abstract (156)
- Konferenzband (77)
- Sonstiges (42)
- Konferenz-Poster (32)
Language
- English (934)
- German (317)
- Multiple languages (1)
- Russian (1)
Keywords
- Mikroelektronik (62)
- RoboCup (32)
- Gamification (12)
- Machine Learning (12)
- injury (10)
- Biomechanik (9)
- Finite-Elemente-Methode (9)
- Kommunikation (9)
- Assistive Technology (8)
- Produktion (8)
Institute
- Fakultät Elektrotechnik und Informationstechnik (E+I) (bis 03/2019) (453)
- Fakultät Maschinenbau und Verfahrenstechnik (M+V) (286)
- Fakultät Elektrotechnik, Medizintechnik und Informatik (EMI) (ab 04/2019) (213)
- Fakultät Wirtschaft (W) (164)
- Fakultät Medien und Informationswesen (M+I) (bis 21.04.2021) (120)
- ivESK - Institut für verlässliche Embedded Systems und Kommunikationselektronik (113)
- INES - Institut für nachhaltige Energiesysteme (59)
- IMLA - Institute for Machine Learning and Analytics (46)
- ACI - Affective and Cognitive Institute (40)
- Fakultät Medien (M) (ab 22.04.2021) (33)
Open Access
- Open Access (560)
- Closed Access (456)
- Closed (223)
- Bronze (214)
- Diamond (29)
- Grün (13)
- Gold (6)
- Hybrid (6)
In 2015, Google engineer Alexander Mordvintsev presented DeepDream as technique to visualise the feature analysis capabilities of deep neural networks that have been trained on image classification tasks. For a brief moment, this technique enjoyed some popularity among scientists, artists, and the general public because of its capability to create seemingly hallucinatory synthetic images. But soon after, research moved on to generative models capable of producing more diverse and more realistic synthetic images. At the same time, the means of interaction with these models have shifted away from a direct manipulation of algorithmic properties towards a predominance of high level controls that obscure the model's internal working. In this paper, we present research that returns to DeepDream to assess its suit-ability as method for sound synthesis. We consider this research to be necessary for two reasons: it tackles a perceived lack of research on musical applications of DeepDream, and it addresses DeepDream's potential to combine data driven and algorithmic approaches. Our research includes a study of how the model architecture, choice of audio data-sets, and method of audio processing influence the acoustic characteristics of the synthesised sounds. We also look into the potential application of DeepDream in a live-performance setting. For this reason, the study limits itself to models consisting of small neural networks that process time-domain representations of audio. These models are resource-friendly enough to operate in real time. We hope that the results obtained so far highlight the attractiveness of Deep-Dream for musical approaches that combine algorithmic investigation with curiosity driven and open ended exploration.
This paper describes the authors' first experiments in creating an artificial dancer whose movements are generated through a combination of algorithmic and interactive techniques with machine learning. This approach is inspired by the time honoured practice of puppeteering. In puppeteering, an articulated but inanimate object seemingly comes to live through the combined effects of a human controlling select limbs of a puppet while the rest of the puppet's body moves according to gravity and mechanics. In the approach described here, the puppet is a machine-learning-based artificial character that has been trained on motion capture recordings of a human dancer. A single limb of this character is controlled either manually or algorithmically while the machine-learning system takes over the role of physics in controlling the remainder of the character's body. But rather than imitating physics, the machine-learning system generates body movements that are reminiscent of the particular style and technique of the dancer who was originally recorded for acquiring training data. More specifically, the machine-learning system operates by searching for body movements that are not only similar to the training material but that it also considers compatible with the externally controlled limb. As a result, the character playing the role of a puppet is no longer passively responding to the puppeteer but makes movement decisions on its own. This form of puppeteering establishes a form of dialogue between puppeteer and puppet in which both improvise together, and in which the puppet exhibits some of the creative idiosyncrasies of the original human dancer.
Generative machine learning models for creative purposes play an increasingly prominent role in the field of dance and technology. A particularly popular approach is the use of such models for generating synthetic motions. Such motions can either serve as source of ideation for choreographers or control an artificial dancer that acts as improvisation partner for human dancers. Several examples employ autoencoder-based deep-learning architectures that have been trained on motion capture recordings of human dancers. Synthetic motions are then generated by navigating the autoencoder's latent space. This paper proposes an alternative approach of using an autoencoder for creating synthetic motions. This approach controls the generation of synthetic motions on the level of the motion itself rather than its encoding. Two different methods are presented that follow this principle. Both methods are based on the interactive control of a single joint of an artificial dancer while the other joints remain under the control of the autoencoder. The first method combines the control of the orientation of a joint with iterative autoencoding. The second method combines the control of the target position of a joint with forward kinematics and the application of latent difference vectors. As illustrative example of an artistic application, this latter method is used for an artificial dancer that plays a digital instrument. The paper presents the implementation of these two methods and provides some preliminary results.
Strings P
(2021)
Strings is an audiovisual performance for an acoustic violin and two generative instruments, one for creating synthetic sounds and one for creating synthetic imagery. The three instruments are related to each other conceptually , technically, and aesthetically by sharing the same physical principle, that of a vibrating string. This submission continues the work the authors have previously published at xCoAx 2020. The current submission briefly summarizes the previous publication and then describes the changes that have been made to Strings. The P in the title emphasizes, that most of these changes have been informed by experiences collected during rehearsals (in German Proben). These changes have helped Strings to progress from a predominantly technical framework to a work that is ready for performance.
In anisotropic media, the existence of leaky surface acoustic waves is a well-known phenomenon. Very recently, their analogs at the apex of an elastic silicon wedge have been found in experiments using laser-ultrasonics. In addition to a wedge-wave (WW) pulse with low speed, a pseudo-wedge wave (p-WW) pulse was found with a velocity higher than the velocity of shear bulk waves, propagating in the same direction. With a probe-beam-deflection technique, the propagation of the WW pulses was monitored on one of the faces of the wedge at variable distance from the apex. In this way, their depth structure and the leakage of the p-WW could be visualized directly. Calculations were carried out using a method based on a representation of the displacement field in Laguerre functions. This method has been validated by calculating the surface density of states in anisotropic media and comparing the results with those obtained from the surface Green's tensor. The approach has then been extended to the continuum of acoustic modes in infinite wedges with fixed wave-vector along the apex. These calculations confirmed the measured speeds of the WW and p-WW pulses.
Strings
(2020)
This article presents the currently ongoing development of an audiovisual performance work with the title Strings. This work provides an improvisation setting for a violinist, two laptop performers, and two generative systems. At the core of Strings lies an approach that establishes a strong correlation among all participants by means of a shared physical principle. The physical principle is that of a vibrating string. The article discusses how this principle is used in both natural and simulated forms as main interaction layer between all performers and as natural or generative principle for creating audio and video.
Anisotropy has been found to play an important role for the existence of edge-localized acoustic modes as well as for nonlinear effects in rectangular edges. For a certain propagation geometry in silicon, the effective second-order nonlinearity for wedge waves was determined numerically from second-order and third-order elastic moduli and compared with the nonlinearity for Rayleigh waves propagating in the direction of the apex on one of the two surfaces forming the edge. In the presence of weak dispersion resulting from modifications of the wedge tip or coating of the adjacent surfaces, solitary pulses are predicted to exist and their shape was calculated.
Previous studies of the hyphenation of gas chromatographic separation and spectrophotometric detection in the ultraviolet wavelength range between 168 and 330 nm showed a high potential for applications where the analysis of complex samples is required. Within this paper the development of a state-of-the-art detection system for compounds in the vapour phase is described, offering an improved behaviour compared to previous systems: Dependent on the requirements of established detection systems hyphenated with gas chromatography, the main components of the system have to be designed for optimum performance and reliability of the spectrophotometric detector: A deuterium lamp as a broadband light source has been selected for improved stability in the measurements. A new-type absorption cell based on fiber-optics has been developed considering the dynamic necessary to compete with existing techniques. In addition, the influence of the volume of the cell on the chromatogram needs to be analyzed. Tests for determining the performance of the absorption cell in terms of chemical and thermal influences have been carried out. A new spectrophotometer with adequate spectral resolution in the wavelength range, offering improved stability and dynamic for an efficient use in this application was developed. Furthermore, the influence of each component on the performance, reliability and stability of the sensor system will be discussed. An overview and outlook over the potential applications in the environmental, scientific and medical field will be given.
In thin-layer chromatography, fiber-bundle arrays have been introduced for spectral absorption measurements in the UV-region. Using all-silica fiber bundles, the exciting light will be detected after re-emission on the plate with a fiberoptic spectrometer. In addition, fluorescence light can be detected which will be masked by the re-emitted light. Therefore, it is helpful to separate the absorption and fluorescence on the TLC-plate. A modified three-array assembly has been developed: using one array for detection, the two others are used for excitation with broadband band deuterium-light and with UV-LEDs adjusted to the substances under test. As an example, the quantification of glucosamine in nutritional supplements or spinach leaf extract will be described. Using simply heating of the amino-plate for derivation, the reaction product of Glucosamine can be detected sensitively either by light absorption or by fluorescence, using the new fiber-optic assembly. In addition, the properties of the new 3-row fiber-optic array and the commercially available UV-LEDs will be shown, in the interesting wavelength region for excitation of fluorescence, from 260 nm to 360 nm. The squint angle having an influence on coupling efficiency and spatial resolution will be measured with the inverse farfield method. Some properties of UV-LEDs for analytical applications will be described and discussed, too.
Most E-Learning projects tend to separate learning activities from everyday work. This paper presents an approach where closer integration between learning and work is achieved by integrating multimedia services into manufacturing processes. The goal of E-Learning services integration in manufacturing is, through the development of new multimedia solutions, to accelerate and enhance the ability of manufacturing industry to capitalise on the emergence of a powerful global information infrastructure. In this paper we suggest to combine the areas of media streaming services and manufacturing processes, by providing electronic learning offerings as collections of media streaming services. The key components of our approach are 1) an xml based streaming service specification language, 2) automated configuration of distributed E-Learning streaming applications, 3) web services for searching, registration, and creation of E-Learning streaming services.
Integrating voice / video communication into business processes can accelerate resolution time, reduce mistakes, and establish a full audit-trail of the interactions. Some VoIP service providers offer website based or plugin based solutions, which are, however, difficult to integrate with other applications. A promising approach to overcome these disadvantages is the development of appropriate Web Services to allow applications interacting with a VoIP system. We propose a generic framework for VoIP applications consisting of an XML-based service specification language and a set of reusable Web Service components. Service providers using the proposed service-oriented architecture can offer to their customers a protocol-neutral Web Service interface, thus enabling the deployment of a general and integrated VoIP solution.
In this paper we suggest to combine the areas of media streaming services, mobile devices, and manufacturing processes to support monitoring, controlling and supervising production processes in order to achieve high levels of efficiency and environmentally friendly production. It contains a comprehensive and detailed explanation of the proposed E-Learning streaming framework, especially the adaption of streaming services to mobile environments. The key components of our approach are 1) an XML-based streaming service specification language, 2) adaption of multimedia E-Learning services to mobile environments, and 3) a media delivery platform for searching, registration, and creation of streaming services for mobile devices.
The central purpose of this paper is to present a novel framework supporting the specification and the implementation of media streaming services using XML and Java Media Framework (JMF). It provides an integrated service development environment comprising of a streaming service model, a service specification language and several implementation and retrieval tools. Our approach is based on a clear separation of a streaming service specification, and its implementation by a distributed JMF application and can be used for different streaming paradigms, e.g. push and pull services.
This paper presents an approach where closer integration between learning and work is achieved by integrating multimedia services into manufacturing processes. The goal of E-Learning services integration in manufacturing processes is, through the development of new multimedia services, to accelerate and enhance the ability of manufacturing industry to capitalise on the emergence of a powerful global information infrastructure. In this paper we suggest to combine the areas of media streaming services and manufacturing processes, by providing electronic learning offerings as collections of media streaming services. The key components of our approach are 1) an xml based streaming service specification language, 2) automated configuration of distributed E-Learning streaming applications, 3) Web Services for searching, registration, and creation of E-Learning streaming services.
We propose a new streaming media service development environment comprising of a streaming media service model, a XML based service specification language and several implementation and configuration management tools. In our project, the described approach is used for integration of streaming based eLearning services in manufacturing processes of a subcontractor to the automotive industry. The key components of our approach are 1) an xml based streaming service specification language, 2) a set of web services for searching, registration, and creation of streaming services, 3) caching and replication policies based on timing information derived from the service specifications.
The goal of eLearning services integration in manufacturing is, through the development of new multimedia solutions, to accelerate and enhance the ability of the manufacturing industry to capitalise on the emergence of a powerful global information infrastructure. The key components of our approach are: (1) an XML based streaming service specification language; (2) automatic configuration of distributed eLearning streaming service implementations; (3) a set of Web services for searching, registration, and creation of streaming services; (4) caching and replication policies based on timing information derived from the service specifications. We also introduce a new concept for cache management during runtime, e.g., content is distributed to cache servers located at the edge of a network close to the client.
The central purpose of this paper is to present a novel framework supporting the specification, the implementation and retrieval of media streaming services. It provides an integrated service development environment comprising of a streaming service model, a service specification language and several implementation and retrieval tools. Our approach is based on a clear separation of a streaming service specification, and its implementation by a distributed application and can be used for different streaming paradigms, e.g. push and pull services.
This paper treats the Brillouin backscattering in a single mode optical fiber and its implications on the Brillouin Ring Laser Gyroscope (BRLG). The BRLG consists of a fiber ring cavity in which stimulated Brillouin scattering is induced and provides two resonant counterpropagating backscattered waves. If this cavity is rotating around its axis, the backscattered waves get different resonant frequencies because of the Sagnac effect. The frequency difference is proportional to the rotation rate (Omega) by inducing a frequency offset between the counterpropagating waves. Some reported Brillouin spectra exhibit several peaks, which means that one pump wave provides at least two backscattered waves with distinguishable frequencies. In order to understand this multi-backscattering and to take advantage of it for the BRLG, we present results of a simulation of the Brillouin backscattering in a single mode optical fiber.
We aim to debate and eventually be able to carefully judge how realistic the following statement of a young computer scientist is: “I would like to become an ethical correctly acting offensive cybersecurity expert”. The objective of this article is not to judge what is good and what is wrong behavior nor to present an overall solution to ethical dilemmas. Instead, the goal is to become aware of the various personal moral dilemmas a security expert may face during his work life. For this, a total of 14 cybersecurity students from HS Offenburg were asked to evaluate several case studies according to different ethical frameworks. The results and particularities are discussed, considering different ethical frameworks. We emphasize, that different ethical frameworks can lead to different preferred actions and that the moral understanding of the frameworks may differ even from student to student.
The prototype of an optical gyro encoder (OGE) has been successfully tested on the NTT telescope in September '93. The OGE consists of a ring laser gyro and a fiber optic gyro with their input axis parallel. The gyro outptu signals are compensated for earth rotation and misalignment and are subsequently integrated to get the angles. An adaptive digital control loop locks the fiber optic gyro to the laser gyro data. Thus the combined output has the precision of the laser gyro and the low noise of the fiber optic gyro. Specifically, the bias stability is better than 2 X 10-3 deg/h, the scale factor accuracy better than 1 ppm, the random walk coefficient better than 5 X 10-4 deg/(root)h and the resolution better than 3 X 10-4 arcsec. The OGE has been mounted in the altitude and in the azimuthy axis of the telescope. The data were compared with the telescope disk encoder data. The test data show that the pointing accuracy is about 1 arcsec and the tracking accuracy 0.1 arcsec over a time of 30 seconds. This accuracy is sufficient for the very large telescope, for instance.
An investigation is underway regarding the usefulness of altazimuth-mounting telescopes' incorporation of laser gyros for pointing and fiber gyros with extremely small random-walk coefficient for telescope inertial stabilization during tracking. A star tracker is expected to help stabilize long-term gyro bias. Gyro and telescope specifications have been derived by means of computer simulations and systems analyses.
Artificial intelligence (AI), and in particular machine learning algorithms, are of increasing importance in many application areas but interpretability and understandability as well as responsibility, accountability, and fairness of the algorithms' results, all crucial for increasing the humans' trust into the systems, are still largely missing. Big industrial players, including Google, Microsoft, and Apple, have become aware of this gap and recently published their own guidelines for the use of AI in order to promote fairness, trust, interpretability, and other goals. Interactive visualization is one of the technologies that may help to increase trust in AI systems. During the seminar, we discussed the requirements for trustworthy AI systems as well as the technological possibilities provided by interactive visualizations to increase human trust in AI.
The use of artificial intelligence continues to impact a broad variety of domains, application areas, and people. However, interpretability, understandability, responsibility, accountability, and fairness of the algorithms' results - all crucial for increasing humans' trust into the systems - are still largely missing. The purpose of this seminar is to understand how these components factor into the holistic view of trust. Further, this seminar seeks to identify design guidelines and best practices for how to build interactive visualization systems to calibrate trust.
Variable refrigerant flow (VRF) and variable air volume (VAV) systems are considered among the best heating, ventilation, and air conditioning systems (HVAC) thanks to their ability to provide cooling and heating in different thermal zones of the same building. As well as their ability to recover the heat rejected from spaces requiring cooling and reuse it to heat another space. Nevertheless, at the same time, these systems are considered one of the most energy-consuming systems in the building. So, it is crucial to well size the system according to the building’s cooling and heating needs and the indoor temperature fluctuations. This study aims to compare these two energy systems by conducting an energy model simulation of a real building under a semi-arid climate for cooling and heating periods. The developed building energy model (BEM) was validated and calibrated using measured and simulated indoor air temperature and energy consumption data. The study aims to evaluate the effect of these HVAC systems on energy consumption and the indoor thermal comfort of the building. The numerical model was based on the Energy Plus simulation engine. The approach used in this paper has allowed us to reach significant quantitative energy saving along with a high level of indoor thermal comfort by using the VRF system compared to the VAV system. The findings prove that the VRF system provides 46.18% of the annual total heating energy savings and 6.14% of the annual cooling and ventilation energy savings compared to the VAV system.
Learning programming fundamentals is considered as one of the most challenging and complex learning activities. Some authors have proposed visual programming language (VPL) approaches to address part of the inherent complexity [1]. A visual programming language lets users develop programs by combining program elements, like loops graphically rather than by specifying them textually. Visual expressions, spatial arrangements of text and graphic symbols are used either as syntax elements or secondary notation. VPLs are normally used for educational multimedia, video games, system development, and data warehousing/business analytics purposes. For example, Scratch, a platform of Massachusetts Institute of Technology, is designed for kids and after school programs.
Design of mobile software applications is considered as one of the most challenging application domains due to the build in sensors as part of a mobile device, like GPS, camera or Near Field Communication (NFC). Sensors enable creation of context-aware mobile applications in which applications can discover and take advantage of contextual information, such as user location, nearby people and objects, and the current user activity. As a consequence, context-aware mobile applications can sense clues about the situational environment making mobile devices more intelligent, adaptive, and personalized. Such context aware mobile applications seem to be motivating and attractive case studies, especially for programming beginners (“my own first app”).
In this work, we introduce a use-case centered approach as well as clear separation of user interface design and sensor-based program development. We provide an in-depth discussion of a new VPL based teaching method, a step by step development process to enable programming beginners the creation of context aware mobile applications. Finally, we argue that addressing challenges for programming beginners by our teaching approach could make programming teaching more motivating, with an additional impact on the final software quality and scalability.
The key contributions of our study are the following:
- An overview of existing attempts to use VPL approaches for mobile applications
- A use case centered teaching approach based on a clear separation of user interface design and sensor-based program development
- A teaching case study enabling beginners a step by step creation of context-aware mobile applications based on the MIT App Inventor (a platform of Massachusetts Institute of Technology)
- Open research challenges and perspectives for further development of our teaching approach
References:
[1] Idrees, M., Aslam, F. (2022). A Comprehensive Survey and Analysis of Diverse Visual Programming Languages, VFAST Transactions on Software Engineering, 2022, Volume 10, Number 2, pp 47-60.
3D Bin Picking with an innovative powder filled gripper and a torque controlled collaborative robot
(2023)
A new and innovative powder filled gripper concept will be introduced to a process to pick parts out of a box without the use of a camera system which guides the robot to the part. The gripper is a combination of an inflatable skin, and a powder inside. In the unjammed condition, the powder is soft and can adjust to the geometry of the part which will be handled. By applying a vacuum to the inflatable skin, the powder gets jammed and transforms to a solid shaped form in which the gripper was brought before applying the vacuum. This physical principle is used to pick parts. The flexible skin of the gripper adjusts to all kinds of shapes, and therefore, can be used to realize 3D bin picking. With the help of a force controlled robot, the gripper can be pushed with a consistent force on flexible positions depending of the filling level of the box. A Kuka LBR iiwa with joint torque sensors in all of its seven axis’ was used to achieve a constant contact pressure. This is the basic criteria to achieve a robust picking process.
While most ultrafast time-resolved optical pump-probe experiments in magnetic materials reveal the spatially homogeneous magnetization dynamics of ferromagnetic resonance (FMR), here we explore the magneto-elastic generation of GHz-to-THz frequency spin waves (exchange magnons). Using analytical magnon oscillator equations, we apply time-domain and frequency-domain approaches to quantify the results of ultrafast time-resolved optical pump-probe experiments in free-standing ferromagnetic thin films. Simulations show excellent agreement with the experiment, provide acoustic and magnetic (Gilbert) damping constants and highlight the role of symmetry-based selection rules in phonon-magnon interactions. The analysis is extended to hybrid multilayer structures to explore the limits of resonant phonon-magnon interactions up to THz frequencies.
An international study summarizes the threat situation in the OT environment under the heading "Growing security threats" [1]. According to this study, attacks on automation systems are likely to increase in the future. Accordingly, an automation system must be able to protect the integrity of the transmitted information in the future. This requirement is motivated, among other things, by the fact that the network-side isolation of industrial communication systems is no longer considered sufficient as the sole protective measure. This paper uses the example of PROFINET to show how the future requirements for a real-time communication protocol can be met and how they can be derived from the IEC 62443 standard.
Polyarticulated active prostheses constitute a promising solution for upper limb amputees. The bottleneck for their adoption though, is the lack of intuitive control. In this context, machine learning algorithms based on pattern recognition from electromyographic (EMG) signals represent a great opportunity for naturally operating prosthetic devices, but their performance is strongly affected by the selection of input features. In this study, we investigated different combinations of 13 EMG-derived features obtained from EMG signals of healthy individuals performing upper limb movements and tested their performance for movement classification using an Artificial Neural Network. We found that input data (i.e., the set of input features) can be reduced by more than 50% without any loss in accuracy, while diminishing the computing time required to train the classifier. Our results indicate that input features must be properly selected in order to optimize prosthetic control.
Differentiation between human and non-human objects can increase efficiency of human-robot collaborative applications. This paper proposes to use convolutional neural networks for classifying objects in robotic applications. The body temperature of human beings is used to classify humans and to estimate the distance to the sensor. Using image classification with convolutional neural networks it is possible to detect humans in the surroundings of a robot up to five meters distance with low-cost and low-weight thermal cameras. Using transfer learning technique we trained the GoogLeNet and MobilenetV2. Results show accuracies of 99.48 % and 99.06 % respectively.
Detecting Images Generated by Deep Diffusion Models using their Local Intrinsic Dimensionality
(2023)
Diffusion models recently have been successfully applied for the visual synthesis of strikingly realistic appearing images. This raises strong concerns about their potential for malicious purposes. In this paper, we propose using the lightweight multi Local Intrinsic Dimensionality (multiLID), which has been originally developed in context of the detection of adversarial examples, for the automatic detection of synthetic images and the identification of the according generator networks. In contrast to many existing detection approaches, which often only work for GAN-generated images, the proposed method provides close to perfect detection results in many realistic use cases. Extensive experiments on known and newly created datasets demonstrate that the proposed multiLID approach exhibits superiority in diffusion detection and model identification.Since the empirical evaluations of recent publications on the detection of generated images are often mainly focused on the "LSUN-Bedroom" dataset, we further establish a comprehensive benchmark for the detection of diffusion-generated images, including samples from several diffusion models with different image sizes.The code for our experiments is provided at https://github.com/deepfake-study/deepfake-multiLID.
Following their success in visual recognition tasks, Vision Transformers(ViTs) are being increasingly employed for image restoration. As a few recent works claim that ViTs for image classification also have better robustness properties, we investigate whether the improved adversarial robustness of ViTs extends to image restoration. We consider the recently proposed Restormer model, as well as NAFNet and the "Baseline network" which are both simplified versions of a Restormer. We use Projected Gradient Descent (PGD) and CosPGD for our robustness evaluation. Our experiments are performed on real-world images from the GoPro dataset for image deblurring. Our analysis indicates that contrary to as advocated by ViTs in image classification works, these models are highly susceptible to adversarial attacks. We attempt to find an easy fix and improve their robustness through adversarial training. While this yields a significant increase in robustness for Restormer, results on other networks are less promising. Interestingly, we find that the design choices in NAFNet and Baselines, which were based on iid performance, and not on robust generalization, seem to be at odds with the model robustness.
The identification of vulnerabilities is an important element in the software development life cycle to ensure the security of software. While vulnerability identification based on the source code is a well studied field, the identification of vulnerabilities on basis of a binary executable without the corresponding source code is more challenging. Recent research [1] has shown how such detection can generally be enabled by deep learning methods, but appears to be very limited regarding the overall amount of detected vulnerabilities. We analyse to what extent we could cover the identification of a larger variety of vulnerabilities. Therefore, a supervised deep learning approach using recurrent neural networks for the application of vulnerability detection based on binary executables is used. The underlying basis is a dataset with 50,651 samples of vulnerable code in the form of a standardised LLVM Intermediate Representation. Te vectorised features of a Word2Vec model are used to train different variations of three basic architectures of recurrent neural networks (GRU, LSTM, SRNN). A binary classification was established for detecting the presence of an arbitrary vulnerability, and a multi-class model was trained for the identification of the exact vulnerability, which achieved an out-of-sample accuracy of 88% and 77%, respectively. Differences in the detection of different vulnerabilities were also observed, with non-vulnerable samples being detected with a particularly high precision of over 98%. Thus, our proposed technical approach and methodology enables an accurate detection of 23 (compared to 4 [1]) vulnerabilities.
The importance of machine learning (ML) has been increasing dramatically for years. From assistance systems to production optimisation to healthcare support, almost every area of daily life and industry is coming into contact with machine learning. Besides all the benefits ML brings, the lack of transparency and difficulty in creating traceability pose major risks. While solutions exist to make the training of machine learning models more transparent, traceability is still a major challenge. Ensuring the identity of a model is another challenge, as unnoticed modification of a model is also a danger when using ML. This paper proposes to create an ML Birth Certificate and ML Family Tree secured by blockchain technology. Important information about training and changes to the model through retraining can be stored in a blockchain and accessed by any user to create more security and traceability about an ML model.
Public educational institutions are increasingly confronted with a decline in the number of applicants, which is why competition between colleges and universities is also intensifying. For this reason, it is important to position oneself as an institution in order to be perceived by the various target groups and to differentiate oneself from the competition. In this context, the brand and thus its perception and impact play a decisive role, especially in view of the desired communication of the institution's own values and its self-image, the brand identity. To this end, emotions serve as an approach to creating positive stimulation and brand loyalty.
Visual programming languages (VPL) let users develop software programs by combining visual program elements, like lists of objects, loops or conditional statements rather than by specifying them textually.
Humanoid robots programming is a very attractive and motivating application domain for students, especially for programming beginners. Humanoid robots are constructed in such a way that they mimic the human body by using actuators that perform like muscles. Typically, a humanoid robot consists of sensors and actuators, i.e. torso, a head, two arms, and two legs, though some humanoid robots may replicate only part of the body, for example, from the waist up. In some cases, humanoid robots are equipped with heads designed to replicate additional human facial features such as eyes. Additional sensors are needed by a robot to gather information about the conditions of the environment to allow the robot to make necessary decisions about its position or certain actions that the situation requires, e.g. an arm movement or an open/close hand action. Other examples for sensor are reflective infrared sensors used to detect objects in proximity.
In this work, we introduce a use-case centered approach based on sensors and actors of a robot and a workflow model to visually describe the sequence of actions including conditional actions or concurrent actions. We provide an in-depth discussion of a new VPL based teaching method for programming humanoid robots based on VPLs. Open research challenges, limits and perspectives for further development of our teaching approach are discussed as well.
Sensors and actuators enable creation of context-aware applications in which applications can discover and take advantage of contextual information, such as user location, nearby people and objects. In this work, we use a general context definition, which can be applied to various devices, e.g., robots and mobile devices. Developing context-based software applications is considered as one of the most challenging application domains due to the sensors and actuators as part of a device. We introduce a new development approach for context-based applications by using use-case descriptions and Visual Programming Languages (VPL). The introduction of web-based VPLs, such as Scratch and Snap, has reinvigorated the usefulness of VPLs. We provide an in-depth discussion of our new VPL based method, a step by step development process to enable development of context-based applications. Two case studies illustrate how to apply our approach to different problem domains: Context-based mobile apps and context-based humanoid robot applications.
The main advantage of mobile context-aware applications is to provide effective and tailored services by considering the environmental context, such as location, time, nearby objects and other data, and adapting their functionality according to the changing situations in the context information without explicit user interaction. The idea behind Location-Based Services (LBS) and Object-Based Services (OBS) is to offer fully-customizable services for user needs according to the location or the objects in a mobile user's vicinity. However, developing mobile context-aware software applications is considered as one of the most challenging application domains due to the built-in sensors as part of a mobile device. Visual Programming Languages (VPL) and hybrid visual programming languages are considered to be innovative approaches to address the inherent complexity of developing programs. The key contribution of our new development approach for location and object-based mobile applications is a use case driven development approach based on use case templates and visual code templates to enable even programming beginners to create context-aware mobile applications. An example of the use of the development approach is presented and open research challenges and perspectives for further development of our approach are formulated.
Due to globalization and the resulting increase in competition on the market, products must be produced more and more cheaply, especially in series production, because buyers expect new variants or even completely new products in ever shorter cycles. Injection molding is the most important production process for manufacturing plastic components in large quantities. However, the conventional production of a mold is extremely time-consuming and costly, which creates a contradiction to the fast pace of the market. Additive tooling is an area of application of additive manufacturing, which in the field of injection molding is preferably used for the prototype production of mold inserts. This allows injection molding tools to be produced faster and more cheaply than through the subtractive manufacturing of metal tools. Material Jetting processes using polymers (MJT-UV/P), also called Polyjet Modeling (PJM), have a great potential for use in additive tooling. Due to the poorer mechanical and thermal properties compared to conventional mold insert materials, e.g. steel or aluminum, the previously used design principles cannot be applied. Accordingly, new design guidelines are necessary, which are developed in this paper. The necessary information is obtained with the help of a systematic literature research. The design guidelines are mapped in a uniform design guide, which is structured according to the design process of injection molds. The guidelines do not only refer to the constructive design of the injection mold or the polymer mold insert, but to the entire design process and describe the four phases of planning, conception, development and realization. Particular attention is paid to the special geometric designs of a polymer mold insert and the thermomechanical properties of the mold insert materials. As a result, design guidelines are available that are adapted to the special requirements of additive tooling of molds inserts made of plastics for injection molding.
The increasing diffusion of rapidly developing AI technologies led to the idea of the experiment to combine TRIZ-based automated idea generation with the natural language processing tool ChatGPT, using the chatbot to interpret the automatically generated elementary solution principles. The article explores the opportunities and benefits of a novel AI-enhanced approach to teaching systematic innovation, analyses the learning experience, identifies the factors that affect students' innovation and problem-solving performance, and highlights the main difficulties students face, especially in interdisciplinary problems.