Refine
Year of publication
Document Type
- Conference Proceeding (105) (remove)
Conference Type
- Konferenzartikel (48)
- Konferenz-Abstract (32)
- Sonstiges (20)
- Konferenz-Poster (5)
Language
- English (105) (remove)
Is part of the Bibliography
- yes (105)
Keywords
- RoboCup (31)
- injury (9)
- Biomechanik (6)
- biomechanics (5)
- running (5)
- ACL (4)
- Machine Learning (4)
- Mobile Learning (3)
- Robustness (3)
- Sound Synthesis (3)
- sport (3)
- Android (2)
- Generative Art (2)
- Improvisation (2)
- Mobile Applications (2)
- Simulation-based Interaction (2)
- Visual Programming (2)
- content adaptation (2)
- deep learning (2)
- device independent learning (2)
- overuse (2)
- 3D User Interface (1)
- 3D analysis (1)
- 3D bin picking (1)
- 3D printing (1)
- Actuators (1)
- Additive Tooling (1)
- Adversarial Attacks (1)
- Adversarial Robustness (1)
- Alexander von Humboldt (1)
- Aliasing (1)
- Archives (1)
- Audiovisual Performance (1)
- AudiovisualPerformance (1)
- Automation (1)
- Biomedizin (1)
- Blended Learning (1)
- CNNs (1)
- Computer Vision (1)
- Context-based Services (1)
- Deep Learning (1)
- Digital Library (1)
- Eye tracking (1)
- Fault detection (1)
- Flashcards (1)
- GIS (1)
- Google Maps (1)
- Hemodynamic monitoring (1)
- Humanoid Robots (1)
- Intelligentes Stromnetz (1)
- Internet portal (1)
- KAM (1)
- Language learning (1)
- Location-Based Services (1)
- Location-based Services (1)
- M-learning (1)
- Makespan (1)
- Microelectronics (1)
- Model Calibration (1)
- Nyquist-Shannon (1)
- Object-Based Services (1)
- Online Student Services (1)
- Online simulation (1)
- Ontology-based Classification (1)
- Particle emissions (1)
- Plant commissioning (1)
- Produktion (1)
- Recommendation Techniques (1)
- Robot Applications (1)
- Sampling (1)
- Sensors (1)
- Software (1)
- Transistor (1)
- User behavior (1)
- Virtual Research Environment (1)
- accuracy (1)
- adversarial (1)
- algorithm-based data analysis (1)
- artificial dancer (1)
- atrial fibrillation (1)
- autoML (1)
- autoattack (1)
- biomechanical stimulation (1)
- cantilever (1)
- collaborative learning (1)
- content synchronization (1)
- dance and technology (1)
- device detection (1)
- dickkopf 3 (1)
- digital library (1)
- dynamic hyperlinks (1)
- energy harvesting (1)
- flashcard (1)
- flexible job shop (1)
- force (1)
- force controlled robot (1)
- generalized content (1)
- high-speed cameras (1)
- impedance cardiography (1)
- information network (1)
- interconnected data (1)
- joint torque sensors (1)
- language learning (1)
- laser triangulation (1)
- lid (1)
- mahalanobis (1)
- maintenance scheduling (1)
- measurement (1)
- mobile web design guidelines (1)
- motion synthesis (1)
- multi-material (1)
- multidimensional flashcard (1)
- neural architecture search (1)
- non interfering measurements (1)
- optical on-line/in-line measurements (1)
- optimization (1)
- powder filled gripper (1)
- pruning (1)
- pulmonary vein isolation (1)
- railway system (1)
- remote laboratory (1)
- repeatability (1)
- risk factor (1)
- robotics (1)
- service engineering (1)
- service research (1)
- spectraldefense (1)
- vacuum (1)
- vibration (1)
Institute
- Fakultät Elektrotechnik und Informationstechnik (E+I) (bis 03/2019) (40)
- Fakultät Maschinenbau und Verfahrenstechnik (M+V) (37)
- Fakultät Elektrotechnik, Medizintechnik und Informatik (EMI) (ab 04/2019) (22)
- Fakultät Wirtschaft (W) (9)
- IMLA - Institute for Machine Learning and Analytics (9)
- IBMS - Institute for Advanced Biomechanics and Motion Studies (ab 16.11.2022) (5)
- Fakultät Medien (M) (ab 22.04.2021) (4)
- Fakultät Medien und Informationswesen (M+I) (bis 21.04.2021) (2)
- POIM - Peter Osypka Institute of Medical Engineering (2)
- CRT - Campus Research & Transfer (1)
Open Access
- Bronze (105) (remove)
The main advantage of mobile context-aware applications is to provide effective and tailored services by considering the environmental context, such as location, time, nearby objects and other data, and adapting their functionality according to the changing situations in the context information without explicit user interaction. The idea behind Location-Based Services (LBS) and Object-Based Services (OBS) is to offer fully-customizable services for user needs according to the location or the objects in a mobile user's vicinity. However, developing mobile context-aware software applications is considered as one of the most challenging application domains due to the built-in sensors as part of a mobile device. Visual Programming Languages (VPL) and hybrid visual programming languages are considered to be innovative approaches to address the inherent complexity of developing programs. The key contribution of our new development approach for location and object-based mobile applications is a use case driven development approach based on use case templates and visual code templates to enable even programming beginners to create context-aware mobile applications. An example of the use of the development approach is presented and open research challenges and perspectives for further development of our approach are formulated.
Sensors and actuators enable creation of context-aware applications in which applications can discover and take advantage of contextual information, such as user location, nearby people and objects. In this work, we use a general context definition, which can be applied to various devices, e.g., robots and mobile devices. Developing context-based software applications is considered as one of the most challenging application domains due to the sensors and actuators as part of a device. We introduce a new development approach for context-based applications by using use-case descriptions and Visual Programming Languages (VPL). The introduction of web-based VPLs, such as Scratch and Snap, has reinvigorated the usefulness of VPLs. We provide an in-depth discussion of our new VPL based method, a step by step development process to enable development of context-based applications. Two case studies illustrate how to apply our approach to different problem domains: Context-based mobile apps and context-based humanoid robot applications.
A system for the on-line/in-line measurement of soot particle sizes and concentrations in the undiluted exhaust gas of diesel engines was developed and successfully tested. The unit uses the individual attenuations of three different laser wavelengths and is combined with an optical cell (white principle) with adjustable path lengths from 2.5 to 15 meters.
Time Resolved Measurements of Soot Concentrations and Mean Particle Sizes during EUDC and ECE Cycles
(2002)
The purpose of this study was to describe the effects of running speed and slope on metatarsophalangeal (MTP) joint kinematics. 22 male and female runners underwent 3D motion analysis on an instrumented treadmill at three different speeds (2.5 m/s, 3.0 m/s, 3.5 m/s). At each speed, participants ran at seven slope conditions (downhill: -15%, -10%, -5%, level, and uphill: +5%, +10%, +15%). We found a significant main effect (p < 0.001) of running speed and slope on peak MTP dorsiflexion and a running speed by slope interaction effect (p < 0.001) for peak MTP dorsiflexion velocity. These findings highlight the need to consider running intensity and environmental factors like running surface inclination when considering MTP joint mechanics and technological aids to support runners.
Established robot manufacturers have developed methods to determine and optimize the accuracy of their robots. These methods vary from robot manufacturers to their competitors. Due to the lack of published data, a comparison of robot performance is difficult. The aim of this article is to find methods to evaluate important characteristics of a robot with an accurate and cost-effective setup. A laser triangulation sensor and geometric referenced spheres were used as a base to compare the robot performance.
Objective: Dickkopf 3 (DKK3) has been identified as a urinary biomarker. Values above 4000 pg/mg creatinine (Cr) were linked with a higher risk of short-term decline of kidney function (J Am Soc Nephrol 29: 2722–2733). However, as of today, there is little experience with DKK3 as a risk marker in everyday clinical practice. We used algorithm-based data analysis to evaluate the potential dependence of DKK3 in a cohort from a large single center in Germany.
Method: DKK3 was measured in all CKD patients in our center October 1 st 2018 till Dec. 31 2019, together with calculated GFR (eGFR) and urinary albumin/creatinine ratio (UACR). Kidney transplant patients were excluded. Until the end of follow-up Dec 31 st 2021, repeated measurements were performed for all parameters. Data analysis was performed using MD-Explorer (BioArtProducts, Rostock, Germany) and Python with multiple libraries. Linear regression models were applied in patients for DKK3, eGFR and UACR. Comparison of the models was performed with a twosided Kolmogorov-Smirnov test.
Results: 1206 DKK3 measurements were performed in 1103 patients (621 male, age 70yrs, eGFR 29,41 ml/min/1.73qm, UACR 800 mg/g). 134 patients died during follow-up. DKK3 mean was 2905 pg/mg Cr (max. 20000, 75 % percentile 3800). 121 pts had DKK3 > 4000. At the end of follow-up 7 % of patients with DKK3 < 4000 (initial eGFR 17.6) versus 39.6 % of patients with DDK3 > 4000 (initial eGFR 15.7) underwent dialysis. Compared to eGFR and UACR at baseline, DKK3 > 4000 performed best to predict eGFR loss over the next 12 months.
Conclusion: In this cohort of CKD patients, DKK3 > 4000 at baseline predicted the eGFR slope better than eGFR or UACR at baseline. DKK3 > 4000 reflected a higher risk of progression towards ESRD in patients with similar baseline eGFR levels.
Strings
(2020)
This article presents the currently ongoing development of an audiovisual performance work with the title Strings. This work provides an improvisation setting for a violinist, two laptop performers, and two generative systems. At the core of Strings lies an approach that establishes a strong correlation among all participants by means of a shared physical principle. The physical principle is that of a vibrating string. The article discusses how this principle is used in both natural and simulated forms as main interaction layer between all performers and as natural or generative principle for creating audio and video.
Security in IT systems, particularly in embedded devices like Cyber Physical Systems (CPSs), has become an important matter of concern as it is the prerequisite for ensuring privacy and safety. Among a multitude of existing security measures, the Transport Layer Security (TLS) protocol family offers mature and standardized means for establishing secure communication channels over insecure transport media. In the context of classical IT infrastructure, its security with regard to protocol and implementation attacks has been subject to extensive research. As TLS protocols find their way into embedded environments, we consider the security and robustness of implementations of these protocols specifically in the light of the peculiarities of embedded systems. We present an approach for systematically checking the security and robustness of such implementations using fuzzing techniques and differential testing. In spite of its origin in testing TLS implementations we expect our approach to likewise be applicable to implementations of other cryptographic protocols with moderate efforts.
In the railway technical centers, scheduling the maintenance activities is a very complex task, it consists in ordering, in the time, all the maintenance operations on the workstations, while respecting the number of resources, precedence constraints, and the workstations' availabilities. Currently, this process is not completely automatic. For improving this situation, this paper presents a mathematical model for the maintenance activities scheduling in the case of railway remanufacturing systems. The studied problem is modeled as a flexible job-shop, with the possibility for a job to be executed several times on a stage. MILP formulation is implemented with the Makespan as an objective, representing the time for remanufacturing the train. The aim is to create a generic model for optimizing the planning of the maintenance activities and improving the performance of the railway technical centers. At last, numerical results are presented, discussing the impact of the instances size on the computing time to solve the described problem.
The concept of m-learning which differs from other forms of e-learning covers a wide range of possibilities opened up by the convergence of new mobile technologies, wireless communication structure and distance learning development. This process of converging has launched some new goals to support m-learning where heterogeneity of devices, their operating systems (Linux, Windows, Symbian, Android etc) and supported markup languages (WML, XHTML etc), adaptive content, preferences or characteristics of user have become some of the major problems to be solved. To facilitate the learning process even more and to establish literally anytime anywhere learning, learning material/content should be available to the user always even if the user is in offline. Multiple devices used by the same user should also be synchronized among themselves and with server to provide updated learning content and to give a freedom to the user to choose any device as per his/her convenience. In this paper software architecture has been proposed to solve these problems and has been implemented by using a multidimensional flashcard learning system which synchronizes among all the devices that are being used by the user.
Today's network landscape consists of quite different network technologies, wide range of end-devices with large scale of capabilities and power, and immense quantity of information and data represented in different formats. Research on 3D imaging, virtual reality and holographic techniques will result in new user interfaces (UI) for mobile devices and will increase their diversity and variety. A lot of efforts are being done in order to establish open, scalable and seamless integration of various technologies and content presentation for different devices including mobile considering individual situation of the end user. This is very difficult because various kinds of devices used by different users or in different times/parallel by the same user which are not predictable and have to be recognized by the system in order to identify device capabilities. Not only the devices but also Content and User Interfaces are big issues because they could include different kinds of data format like text, image, audio, video, 3D Virtual Reality data and other upcoming formats. A very suitable and useful example of the use of such a system is mobile learning because of the large amount of varying devices with significantly different features and functionalities. This is true not only to support different learners, e.g. all learners within one learning community, but also to support the same learner using different equipment parallel and/or at different times. Those applications may be significantly enhanced by including virtual reality content presentation. Whatever the purposes are, it is impossible to develop and adapt content for all kind of devices including mobiles individually due to different capabilities of the devices, cost issues and author‘s requirement. A solution should be found to enable the automation of the content adaptation process.
The research project Ko-TAG [2], as part of the research initiative Ko-FAS [1], funded by the German Ministry of Economics and Technologies (BMWi), deals with the development of a wireless cooperative sensor system that shall pro-vide a benefit to current driver assistance systems (DAS) and traffic safety applications (TSA). The system’s primary function is the localization of vulnerable road users (VRU) e.g. pedestrians and powered two-wheelers, using communication signals, but can also serve as pre-crash (surround) safety system among vehicles. The main difference of this project, compared to previous ones that dealt with this topic, e.g. the AMULETT project, is an underlying FPGA based Hardware-Software co-design. The platform drives a real-time capable communication protocol that enables highly scalable network topologies fulfilling the hard real-time requirements of the single localization processes. Additionally it allows the exchange of further data (e.g. sensor data) to support the accident pre-diction process and the channel arbitration, and thus supports true cooperative sensing. This paper gives an overview of the project’s current system design as well as of the implementations of the key HDL entities supporting the software parts of the communication protocol. Furthermore, an approach for the dynamic reconfiguration of the devices is described, which provides several topology setups using a single PCB design.
The services sector is also called “tertiary sector” and has become increasingly important in the last few decades. The process of this occurring structural change is characterized by a significant increase in employment in the services sector. On the other hand, the former economic importance in traditional areas, such as agriculture and forestry, as well as manufacturing, is declining.
To have an overview it is important to look and to analyse the different research studies.
This paper presents a method for supporting the application of Additive Tooling (AT)-based validation environments in integrated product development. Based on a case study, relevant process steps, activities and possible barriers in the realisation of an injection-moulded product are identified and analysed. The aim of the method is to support the target-oriented application of Additive Tooling to obtain physical prototypes at an early stage and to shorten validation cycles.
Autonomous humanoid robots require light weight, high torque and high speed actuators to be able to walk and run. For conventional gears with a fixed gear ratio the product of torque and velocity is constant. On the other hand desired motions require maximum torque and speed. In this paper it is shown that with a variable gear ratio it is possible to vary the relation between torque and velocity. This is achieved by introducing systems of rods and levers to move the joints of our humanoid robot ”Sweaty II”. On the basis of a variable gear ratio low speed and high torque can be achieved for those joint angles, which require this motion mode, whereas high speed and low torque can be realized for those joint angles, where it is favorable for the desired motion.
The efficient support of Hardwae-In-theLoop (HIL) in the design process of hardwaresoftware-co-designed systems is an ongoing challenge. This paper presents a network-based integration of hardware elements into the softwarebased image processing tool „ADTF“, based on a high-performance Gigabit Ethernet MAC and a highly-efficient TCP/IP-stack. The MAC has been designed in VHDL. It was verified in a SystemCsimulation environment and tested on several Altera FPGAs.
Effect of downhill running on biomechanical risk factors associated with iliotibial band syndrome
(2022)
The purpose of this study was to identify the influence of downhill running on biomechanical risk factors for iliotibial band syndrome. We conducted a 3D motion analysis of 22 females and males running on an instrumented treadmill at four different inclinations (0%, -5%, -10%, -15%) at a speed of 3.5 m/s. We found significant differences for biomechanical risk factors associated with iliotibial band syndrome. Peak knee flexion angle at initial ground contact (p < .001), peak knee adduction angle (p = .005), and iliotibial band strain (p < .001) systematically increased with increasing slope. Downhill running increases biomechanical risk factors for iliotibial band syndrome. Our results highlight the need to consider the individual running environment in assessing overuse injury risk in runners.
This paper explores the potential of an m-learning environment by introducing the concept of mLab, a remote laboratory environment accessible through the use of handheld devices.
We are aiming to enhance the existing e-learning platform and internet-assisted laboratory settings, where students are offered in-depth tutoring, by providing compact tuition and tools for controlling simulations that are made available to learners via handheld devices. In this way, students are empowered by having access totheir simulations from any place and at any time.
The Metering Bus, also known as M-Bus, is a European standard EN13757-3 for reading out metering devices, like electricity, water, gas, or heat meters. Although real-life M-Bus networks can reach a significant size and complexity, only very simple protocol analyzers are available to observe and maintain such networks. In order to provide developers and installers with the ability to analyze the real bus signals easily, a web-based monitoring tool for the M-Bus has been designed and implemented. Combined with a physical bus interface it allows for measuring and recording the bus signals. For this at first a circuit has been developed, which transforms the voltage and current-modulated M-Bus signals to a voltage signal that can be read by a standard ADC and processed by an MCU. The bus signals and packets are displayed using a web server, which analyzes and classifies the frame fragments. As an additional feature an oscilloscope functionality is included in order to visualize the physical signal on the bus. This paper describes the development of the read-out circuit for the Wired M-Bus and the data recovery.
This study aimed to compare a simplified calculation of the knee abduction moment with the traditional inverse dynamics calculation when athletes perform fake-cut maneuvers with different complexities. In the simplified calculation, we multiply the force vector with its lever arm to the knee, projected onto the local coordinate system of the proximal thigh, hence neglecting the inertial contributions from distal segments. We found very strong ranking consistency using Spearman’s rank correlation coefficient when using the simplified method compared to the traditional calculation. Independent of the tasks, the simplified method resulted in higher moments than the inverse dynamics. This was caused by ignoring the moment caused by segment linear acceleration generating a counteracting moment by about 7%. An alternative to the complex calculations of inverse dynamics can be used to investigate the contributions of the GRF magnitude and its lever arm to the knee.
Recently, RobustBench (Croce et al. 2020) has become a widely recognized benchmark for the adversarial robustness of image
classification networks. In it’s most commonly reported sub-task, RobustBench evaluates and ranks the adversarial robustness of trained neural networks on CIFAR10 under AutoAttack (Croce and Hein 2020b) with l∞ perturbations limited to ϵ = 8/255. With leading scores of the currently best performing models of around 60% of the baseline, it is fair to characterize this benchmark to be quite challenging. Despite it’s general acceptance in recent literature, we aim to foster discussion about the suitability of RobustBench as a key indicator for robustness which could be generalized to practical applications. Our line of argumentation against this is two-fold and supported by excessive experiments presented in this paper: We argue that I) the alternation of data by AutoAttack with l∞, ϵ = 8/255 is unrealistically strong, resulting in close to perfect detection rates of adversarial samples even by simple detection algorithms and human observers.
We also show that other attack methods are much harder to detect while achieving similar success rates. II) That results on low resolution data sets like CIFAR10 do not generalize well to higher resolution images as gradient based attacks appear to become even more detectable with increasing resolutions.
Machine-to-machine communication is continuously extending to new application fields. Especially smart metering has the potential to become the first really large-scale M2M application. Although in the future distributed meter devices will be mainly connected via dedicated primary communication protocols, like ZigBee, Wireless
M-Bus or alike, a major percentage of all meters will be connected via point to point communication using GPRS or UMTS platforms. Thus, such meter devices have to be extremely cost and energy efficient, especially if the devices are battery based and powered several years by a single battery. This paper presents the development of an automated measurement unit for power and time, thus energy characteristics can be recorded. The measurement unit includes a hardware platform for the device
under test (DUT) and a database based software environment for a smooth execution and analysis of the measurements.
3D Bin Picking with an innovative powder filled gripper and a torque controlled collaborative robot
(2023)
A new and innovative powder filled gripper concept will be introduced to a process to pick parts out of a box without the use of a camera system which guides the robot to the part. The gripper is a combination of an inflatable skin, and a powder inside. In the unjammed condition, the powder is soft and can adjust to the geometry of the part which will be handled. By applying a vacuum to the inflatable skin, the powder gets jammed and transforms to a solid shaped form in which the gripper was brought before applying the vacuum. This physical principle is used to pick parts. The flexible skin of the gripper adjusts to all kinds of shapes, and therefore, can be used to realize 3D bin picking. With the help of a force controlled robot, the gripper can be pushed with a consistent force on flexible positions depending of the filling level of the box. A Kuka LBR iiwa with joint torque sensors in all of its seven axis’ was used to achieve a constant contact pressure. This is the basic criteria to achieve a robust picking process.
A platform of an electronic capsule is being developed for multi-task medical assistant application. It includes a near field telemetry unit for bidirectional communication system of 115 KHz low carrier frequency for inductive data transmission suited for human body energy transfer. The system triggers an actuator for drug delivery in various time and release forms via wireless external control, it has the ability to record temperature, measure pH of the body (additional sensors), and retrieve data to the outside. It consists of a 32bit processor, memory, external peripheries, and detection facility. The complete system is designed to fit small-size mass medical application with low power consumption, size of 7x25mm. The system is designed, simulated and emulated on FPGA. A final layout of the complete chip design is still under progress.
Estimating the Robustness of Classification Models by the Structure of the Learned Feature-Space
(2022)
Over the last decade, the development of deep image classification networks has mostly been driven by the search for the best performance in terms of classification accuracy on standardized benchmarks like ImageNet. More recently, this focus has been expanded by the notion of model robustness, \ie the generalization abilities of models towards previously unseen changes in the data distribution. While new benchmarks, like ImageNet-C, have been introduced to measure robustness properties, we argue that fixed testsets are only able to capture a small portion of possible data variations and are thus limited and prone to generate new overfitted solutions. To overcome these drawbacks, we suggest to estimate the robustness of a model directly from the structure of its learned feature-space. We introduce robustness indicators which are obtained via unsupervised clustering of latent representations from a trained classifier and show very high correlations to the model performance on corrupted test data.
Sweaty has already participated several times in RoboCup soccer competitions (Adult Size). Now the work is focused coordinating the play of two robots. Moreover, we are working on stabilizing the gait by adding additional sensor information. An ongoing work is the optimization of the control strategy by balancing between impedance and position control. By minimizing the jerk, gait and overall gameplay should improve significantly.
Android is an operating system which was developed for use in smart mobile phones and is the current leader in this market. A lot of efforts are being spent to make Android available to the embedded world, as well. Many embedded systems do not have a local GUI and are therefore called headless devices. This paper presents the results of an analysis of the general suitability of Anroid in headless embedded systems and ponders the advantages and disadvantages. It focuses on the hardware related issues, i.e. to what extent Android supports hardware peripherals normally used in embedded systems.
Many commonly well-performing convolutional neural network models have shown to be susceptible to input data perturbations, indicating a low model robustness. Adversarial attacks are thereby specifically optimized to reveal model weaknesses, by generating small, barely perceivable image perturbations that flip the model prediction. Robustness against attacks can be gained for example by using adversarial examples during training, which effectively reduces the measurable model attackability. In contrast, research on analyzing the source of a model’s vulnerability is scarce. In this paper, we analyze adversarially trained, robust models in the context of a specifically suspicious network operation, the downsampling layer, and provide evidence that robust models have learned to downsample more accurately and suffer significantly less from aliasing than baseline models.
Over the last years, Convolutional Neural Networks (CNNs) have been the dominating neural architecture in a wide range of computer vision tasks. From an image and signal processing point of view, this success might be a bit surprising as the inherent spatial pyramid design of most CNNs is apparently violating basic signal processing laws, i.e. Sampling Theorem in their down-sampling operations. However, since poor sampling appeared not to affect model accuracy, this issue has been broadly neglected until model robustness started to receive more attention. Recent work in the context of adversarial attacks and distribution shifts, showed after all, that there is a strong correlation between the vulnerability of CNNs and aliasing artifacts induced by poor down-sampling operations. This paper builds on these findings and introduces an aliasing free down-sampling operation which can easily be plugged into any CNN architecture: FrequencyLowCut pooling. Our experiments show, that in combination with simple and Fast Gradient Sign Method (FGSM) adversarial training, our hyper-parameter free operator substantially improves model robustness and avoids catastrophic overfitting. Our code is available at https://github.com/GeJulia/flc_pooling
Despite the success of convolutional neural networks (CNNs) in many academic benchmarks for computer vision tasks, their application in the real-world is still facing fundamental challenges. One of these open problems is the inherent lack of robustness, unveiled by the striking effectiveness of adversarial attacks. Current attack methods are able to manipulate the network's prediction by adding specific but small amounts of noise to the input. In turn, adversarial training (AT) aims to achieve robustness against such attacks and ideally a better model generalization ability by including adversarial samples in the trainingset. However, an in-depth analysis of the resulting robust models beyond adversarial robustness is still pending. In this paper, we empirically analyze a variety of adversarially trained models that achieve high robust accuracies when facing state-of-the-art attacks and we show that AT has an interesting side-effect: it leads to models that are significantly less overconfident with their decisions, even on clean data than non-robust models. Further, our analysis of robust models shows that not only AT but also the model's building blocks (like activation functions and pooling) have a strong influence on the models' prediction confidences. Data & Project website: https://github.com/GeJulia/robustness_confidences_evaluation
Despite the success of convolutional neural networks (CNNs) in many academic benchmarks for computer vision tasks, their application in the real-world is still facing fundamental challenges. One of these open problems is the inherent lack of robustness, unveiled by the striking effectiveness of adversarial attacks. Adversarial training (AT) is often considered as a remedy to train more robust networks. In this paper, we empirically analyze a variety of adversarially trained models that achieve high robust accuracies when facing state-of-the-art attacks and we show that AT has an interesting side-effect: it leads to models that are significantly less overconfident with their decisions even on clean data than non-robust models. Further, our analysis of robust models shows that not only AT but also the model's building blocks (like activation functions and pooling) have a strong influence on the models' prediction confidences.
In this contribution, we present a novel 3D printed multi-material, electromagnetic vibration harvester. The harvester is based on a cantilever design and utilizes an embedded constantan wire within a matrix of polyethylene terephthalate glycol (PETG). A prototype has been manufactured with a combination of a fused filament fabrication (FFF) printer and a robot with a custom-made tool.
Recent work has investigated the distributions of learned convolution filters through a large-scale study containing hundreds of heterogeneous image models. Surprisingly, on average, the distributions only show minor drifts in comparisons of various studied dimensions including the learned task, image domain, or dataset. However, among the studied image domains, medical imaging models appeared to show significant outliers through "spikey" distributions, and, therefore, learn clusters of highly specific filters different from other domains. Following this observation, we study the collected medical imaging models in more detail. We show that instead of fundamental differences, the outliers are due to specific processing in some architectures. Quite the contrary, for standardized architectures, we find that models trained on medical data do not significantly differ in their filter distributions from similar architectures trained on data from other domains. Our conclusions reinforce previous hypotheses stating that pre-training of imaging models can be done with any kind of diverse image data.
Electrode modelling and simulation of diagnostic and pulmonary vein isolation in atrial fibrillation
(2022)
The idea of this game is to use a flashcard system to create a short story in a foreign language. The story is developed by a group of people by exchanging sentences via a flashcard system. This way, people can learn from each other without fear of making mistakes because the group members are anonymous.
This paper shows the results of the evaluation of two sets of mobile web design guidelines concerning mobile learning. The first set of guidelines is concerned with the usage of text on mobile device screens. The second set is concerned with the usage of images on mobile devices. The evaluation is performed by eye tracking (objective) as well as questionnaires and interviews (subjective) respectively.
Electronic pills, smart capsules or miniaturized microsystems swallowed by human beings or animals for various biomedical and diagnostic applications are growing rapidly in the last years. This paper searched out the important existing electronic pills in the market and prototypes in research centers. Further objective of this research is to develop a technology platform with enhanced feature to cover the drawback of most
capsules. The designed telemetry unit is a synchronous bidirectional communication block using continuous phase DQPSK of 115 kHz low carrier frequency for inductive data transmission suited for human body energy transfer. The communication system can assist the electronic pill to trigger an actuator for drug delivery, to record temperature, or to measure pH of the body. It consists additionally to a 32bit processor, memory, external peripheries, and detection facility. The complete system is designed to fit small-size mass medical application with low power consumption, size of 7x25mm. The system is designed, simulated and emulated on FPGA.
In this paper, a new method is demonstrated for online remote simulation of photovoltaic systems. The required communication technology for the data exchange is introduced and the methods of PV generator parameter extraction for the simulation models are analysed. The method shown for parameter extraction from the manufacturer data is especially useful for the commissioning procedure, where the measured installed power is transferred to standard test conditions using the simulation model and can then be easily compared with the design power. At a simulation accuracy of 2% using the software environment INSEL ® any problems with the PV generator can reliably be detected. Online simulation of a grid connected PV generator is then carried out during the operation of the photovoltaic plant. The visualisation includes both the monitored and the simulated online data sets, so that a very efficient fault detection scheme is available. The method is implemented and validated on several grid connected photovoltaic power plants in Germany. It is excellently suited to provide automatic and real time fault detection and significantly improve the commissioning procedure for photovoltaic plants of all sizes.
Sweaty has already participated four times in RoboCup soccer competitions (Adult Size) and came second three times. While 2016 Sweaty needed a lot of luck to be finalist, 2017 Sweaty was a serious adversary in the preliminary rounds. In 2018 Sweaty showed up in the final with some lack of experience and room for improvements, but not without any chance. This paper describes the intended improvements of the humanoid adult size robot Sweaty in order to qualify for the RoboCup 2019 adult size competition.
Sweaty has already participated several times in RoboCup soccer competitions (Adult Size). Now the work is focused on stabilizing the gait. Moreover, we would like to overcome the constraints of a ZMP-algorithm that has a horizontal footplate as precondition for the simplification of the equations. In addition we would like to switch between impedance and position control with a fuzzy-like algorithm that might help to minimize jerks when Sweaty’s feet touch the ground.
This paper describes the magmaOffenburg 3D simulation team trying to qualify for RoboCup 2011. While last year’s TDP focused on the tool set created for 3D simulation in this year we describe the further improvement in this tools as well as some new features we implemented focusing on heterogeneous robot models which seem to be used in RoboCup 2012.
An additional tool was written to simply generate situation-dependent strategies. Furthermore some tools, described last year, are now integrated in one single GUI to easy things up.
This paper describes the magmaOffenburg 3D simulation team trying to qualify for RoboCup 2012. While last year’s TDP focused on the tool set created for 3D simulation and the support for heterogeneous robot models, this year we focus on the different ways how robot behavior can be defined in the magmaOffenburg framework and how those behaviors can be improved by learning.
This paper describes the magmaOffenburg 3D simulation team trying to qualify for RoboCup 2013. While last year’s TDP focused on different ways how robot behavior can be defined in the magmaOffenburg framework this year we focus on how we statistically evaluate new features on distributed systems. We also show some results gained through such analysis.
After having described many different aspects of our team software in previous years, in this paper we take the freedom to describe the magmaChallenge framework provided by the magmaOffenburg team. The framework is used as a benchmark tool to run different challenges like the running challenge in 2014 or the kick accuracy challenge in 2015. This description should serve as a documentation to simplify the maintenance by the community and to add new benchmarks in the future.
This paper describes the magmaOffenburg 3D simulation team trying to qualify for RoboCup 2010. While last year’s TDP focused on decisions making using extended behavior networks and on its software architecture and implementation in this year we describe the tool set that was created for RoboCup 3D. It contians a GUI for agent- and world state visualization, for evaluation of localization algorithms and benchmarks in general, a visual editor for Extended Behavior Networks creation and debugging, a live movement tool to interact with the joints and finally a tool for editing behavior motor files.
This paper describes the magmaOffenburg 3D simulation team trying to qualify for RoboCup 2009. It focuses on two distinctive features of the team: decisions making using extended behavior networks and its software architecture and implementation in Java to open the simulation for the Java community.
Technology and computer applications influence our daily lives and questions arise concerning the role of artificial intelligence and decision-making algorithms. There are warning voices, that computers can, in theory, emulate human intelligence-and exceed it. This paper points out that a replacement of humans by computers is unlikely, because human thinking is characterized by cognitive heuristics and emotions, which cannot simply be implemented in machines operating with algorithms, procedural data processing or artificial neural networks. However, we are going to share our responsibilities with superior computer systems, which are tracking and surveying all of our digital activities, whereas we have no idea of the decision-making processes inside the machines. It is shown that we need a new digital humanism defining rules of computer responsibilities to avoid digital totalism and comprehensive monitoring and controlling of individuals within the planet Earth.
This article sets the focus on methods of information technology in the Humboldt Portal, which represents an ongoing research project to develop a virtual research environment on the Internet for the legacy of Alexander von Humboldt. Based on the experiences of developing and providing the Humboldt Digital Library (www.avhumboldt.net) for more than a decade, we defined a working plan to create an Internet portal for comprehensive access to Humboldt’s writings, no matter if documents are provided as PDF files, scan images or XML-TEI documents on external archives (Google Books, Internet Archive, Deutsches Textarchiv, Bibliotheque National de France). Going far beyond services of a digital library we will provide an information network with multimedia assets, which are containing objects like terms, paragraphs, data tables, scan images, or illustrations, together with correlated properties like thematic linkage to other objects, relevant keywords with optional synonyms and dynamic hyperlinks to related translations in different languages. So the Humboldt Portal can contribute to the key question, how to present interconnected data in an appropriate form using information technologies on the Web.
More than 200 years ago, the scientist Alexander von Humboldt noted in his travel diaries that "everything is interconnectedness", when he was fascinated by nature and the phenomena observed. The view of nature has become much more detailed through the knowledge of phenomena and natural processes, which led to a more precise view of nature shaped by Humboldt. Technological progress and the artificial intelligence of highly developed computer systems are upsetting this view and changing the established world view through a new, unprecedented interaction between man and machinery. Thus we need digital axioms and comprehensive rules and laws for such autonomous acting systems that determine human interaction between cybernetic systems and biological individuals. This digital humanism should encompass our relationship to nature, our handling of the complexity and diversity of nature and the technological influences on society in order to avoid technical colonialism through supercomputers.
This paper describes the Sweaty II humanoid adult size robot trying to qualify for the RoboCup 2017 adult size humanoid competition. Sweaty came 2nd in RoboCup 2016 adult size league. The paper describes the main characteristics of Sweaty that made this success possible, and improvements that have been made or are planned to be implemented for RoboCup 2017.
This paper describes the new Sweaty humanoid adult size robot trying to qualify for the RoboCup 2014 adult size humanoid competition. The robot is built from scratch to eventually allow it to run. One characteristic is that to prevent the motors from overheating, water evaporation is used for cooling. The robot is literally sweating which has given it its name. Another characteristic is, that the motors are not directly connected to the frame but by means of beams. This allows a variable transmission ratio depending on the angle.
This paper describes the new Sweaty II humanoid adult size robot trying to qualify for the RoboCup 2016 adult size humanoid competition. Based on experiences during RoboCup 2014, the Sweaty robot has been completely redesigned to a new robot Sweaty II. A major change is the use of linear actuators for the legs. Another characteristic is its indirect actuation by means of rods. This allows a variable transmission ratio depending on the angle of a joint.
Existing approaches solving multi-vehicle pickup and delivery problems with soft time windows typically use common benchmark sets to verify their performance. However, there is a gap from these benchmark sets to real world problems with respect to instance size and problem complexity. In this paper we show that a combination of existing approaches together with improved heuristics is able to deal with the instance sizes and complexity of real world problems. The cost savings potential of the heuristics is compared to human dispatching plans generated from the data of a European carrier.
Meeting the requirements of smart grids local, decentralized subnets will offer additional potentials to stabilize and compensate the utility grid mainly on the low voltage level. In a quite complex configuration these decentralized energy systems are combined power, heat and cooling power distributions. According to the regional and local availability of renewable energy sources advanced energy management concepts should consider climatic conditions as well as the state of the interacting utility grid and consumption profiles. The approach uses demonstrational setups to develop a forecast based energy management for trigeneration subnets by taking into account the running conditions of local electrical and thermal energy conversion units. This should lead to the best coverage of the demand and supporting/stabilizing the utility grid at the same time. For the first of three demonstrational projects the priorities of the subnet are given with the maximization of the CHP operation to substitute a major part of the heating and cooling power delivered by electric heaters or compression chillers.
The title expresses goals the Kansas Geological Survey (KGS) has been working toward for some time. This report extends concepts and objectives developed while working on an earlier effort for effective interactive digital maps on the Internet. That work was reported to the 1998 DMT Workshop in Champaign, Illinois (Ross, 1998). The current project goes beyond previous efforts that focused on methods for serving the contents of a geographic information system (GIS); the points, lines, and polygons representing features of the digital geologic map and the data in the attribute tables of the GIS describing those features.
“Today’s network landscape consists of quite different network technologies, wide range of end-devices with large scale of capabilities and power, and immense quantity of information and data represented in different formats” [9]. A lot of efforts are being done in order to establish open, scalable and seamless integration of various technologies and content presentation for different devices including mobile considering individual situation of the end user. This is very difficult because various kinds of devices used by different users or in different times/parallel by the same user which is not predictable and have to be recognized by the system in order to know device capabilities. Not only the devices but also Content and User Interfaces are big issues because they could include different kinds of data format like text, image, audio, video, 3D Virtual Reality data and upcoming other formats. Language Learning Game (LLG) is such an example of a device independent application where different kinds of devices and data formats, as a content of a flashcard is used for a collaborative learning. The idea of this game is to create a short story in a foreign language by using mobile devices. The story is developed by a group of participants by exchanging sentences/data via a flashcard system. This way the participants can learn from each other by knowledge sharing without fear of making mistakes because the group members are anonymous. Moreover they do not need a constant support from a teacher.
Flashcards are a well known and proven method to learn and memorise. Such a way of learning is perfectly suited for “learning on the way,” but carrying all the flashcards could be awkward. In this scenario, a mobile device (mobile phone) is an adequate solution. The new mobile device operating system Android from Google allows for writing multimedia-enriched applications.
In this paper, we propose a unified approach for network pruning and one-shot neural architecture search (NAS) via group sparsity. We first show that group sparsity via the recent Proximal Stochastic Gradient Descent (ProxSGD) algorithm achieves new state-of-the-art results for filter pruning. Then, we extend this approach to operation pruning, directly yielding a gradient-based NAS method based on group sparsity. Compared to existing gradient-based algorithms such as DARTS, the advantages of this new group sparsity approach are threefold. Firstly, instead of a costly bilevel optimization problem, we formulate the NAS problem as a single-level optimization problem, which can be optimally and efficiently solved using ProxSGD with convergence guarantees. Secondly, due to the operation-level sparsity, discretizing the network architecture by pruning less important operations can be safely done without any performance degradation. Thirdly, the proposed approach finds architectures that are both stable and well-performing on a variety of search spaces and datasets.
High mobility, electrolyte-gated transistors (EGTs) show high DC performance at low voltages (< 2 V). To model those EGTs, we have used different models for the below and the above threshold regime with appropriate interpolation to ensure continuity and smoothness over all regimes. This empirical model matches very well with our measured results obtained by the electrical characterization of EGTs.
Digital libraries are providing an increasing amount of data, which is normally structured in a classical way by documents and described by metadata as keywords. The data, even in scientific systems such as digital libraries and virtual research environments, will contain a great amount of noise or information unnecessary for our personal interests. Although there has been a lot of progress in the field of information retrieval, search techniques and other content finding methods, there is still much to be done in the field of information retrieval based on user behavior. This paper presents an approach deployed in the Humboldt Digital Library (HDL) to facilitate the retrieval of relevant information to the users of the system, making recommendations of paragraphs based on their profile and the behavior of other users who share similar profiles. The Humboldt digital library represents an innovative system of open access to the legacy of Alexander von Humboldt in a digital form on the Internet (www.avhumboldt.net). It contributes to the key question, how to present interconnected data in a proper form using information technologies.
Strings P
(2021)
Strings is an audiovisual performance for an acoustic violin and two generative instruments, one for creating synthetic sounds and one for creating synthetic imagery. The three instruments are related to each other conceptually , technically, and aesthetically by sharing the same physical principle, that of a vibrating string. This submission continues the work the authors have previously published at xCoAx 2020. The current submission briefly summarizes the previous publication and then describes the changes that have been made to Strings. The P in the title emphasizes, that most of these changes have been informed by experiences collected during rehearsals (in German Proben). These changes have helped Strings to progress from a predominantly technical framework to a work that is ready for performance.
In 2015, Google engineer Alexander Mordvintsev presented DeepDream as technique to visualise the feature analysis capabilities of deep neural networks that have been trained on image classification tasks. For a brief moment, this technique enjoyed some popularity among scientists, artists, and the general public because of its capability to create seemingly hallucinatory synthetic images. But soon after, research moved on to generative models capable of producing more diverse and more realistic synthetic images. At the same time, the means of interaction with these models have shifted away from a direct manipulation of algorithmic properties towards a predominance of high level controls that obscure the model's internal working. In this paper, we present research that returns to DeepDream to assess its suit-ability as method for sound synthesis. We consider this research to be necessary for two reasons: it tackles a perceived lack of research on musical applications of DeepDream, and it addresses DeepDream's potential to combine data driven and algorithmic approaches. Our research includes a study of how the model architecture, choice of audio data-sets, and method of audio processing influence the acoustic characteristics of the synthesised sounds. We also look into the potential application of DeepDream in a live-performance setting. For this reason, the study limits itself to models consisting of small neural networks that process time-domain representations of audio. These models are resource-friendly enough to operate in real time. We hope that the results obtained so far highlight the attractiveness of Deep-Dream for musical approaches that combine algorithmic investigation with curiosity driven and open ended exploration.
This paper describes the authors' first experiments in creating an artificial dancer whose movements are generated through a combination of algorithmic and interactive techniques with machine learning. This approach is inspired by the time honoured practice of puppeteering. In puppeteering, an articulated but inanimate object seemingly comes to live through the combined effects of a human controlling select limbs of a puppet while the rest of the puppet's body moves according to gravity and mechanics. In the approach described here, the puppet is a machine-learning-based artificial character that has been trained on motion capture recordings of a human dancer. A single limb of this character is controlled either manually or algorithmically while the machine-learning system takes over the role of physics in controlling the remainder of the character's body. But rather than imitating physics, the machine-learning system generates body movements that are reminiscent of the particular style and technique of the dancer who was originally recorded for acquiring training data. More specifically, the machine-learning system operates by searching for body movements that are not only similar to the training material but that it also considers compatible with the externally controlled limb. As a result, the character playing the role of a puppet is no longer passively responding to the puppeteer but makes movement decisions on its own. This form of puppeteering establishes a form of dialogue between puppeteer and puppet in which both improvise together, and in which the puppet exhibits some of the creative idiosyncrasies of the original human dancer.
The purpose of this study was to 1) compare knee joint kinematics and kinetics of fake-and-cut tasks of varying complexity in 51 female handball players and 2) present a case study of one athlete who ruptured her ACL three weeks post data collection. External knee joint moments and knee joint angles in all planes at the instance of the peak external knee abduction moment (KAM) as well as moment and angle time curves were analyzed. Peak KAMs and knee internal rotation moments were substantially higher than published values obtained during simple change-of-direction tasks and, along with flexion angles, differed significantly between the tasks. Introducing a ball reception and a static defender increased joint loads while they partially decreased again when anticipation was lacking. Our results suggest to use game-specific assessments of injury risk while complexity levels do not directly increase knee loading. Extreme values of several risk factors for a post-test injured athlete highlight the need and usefulness of appropriate screenings.
Landing heel first has been associated with elevated external knee abduction moments (KAM), thereby potentially increasing the risk of sustaining a non-contact ACL injury. Apart from the foot strike angle, knee valgus angle (VAL) and vertical center of mass velocity at initial ground contact (IC) have been associated with increased KAM in females across different sidestep cuts. While real-time biofeedback training has been proven effective for gait retraining [4], the highly dynamic, non-cyclical nature of cutting maneuvers makes real-time feedback unsuitable and alternative approaches necessary. This study aimed at assessing the efficacy of immediate software-aided feedback on cutting technique in reducing KAM during handball-specific cutting maneuvers.
The majority of anterior cruciate ligament (ACL) injuries in team sports are non-contact injuries, with cutting maneuvers identified as high-risk tasks. Young female handball players have been shown to be at greater risk for ACL injuries than males. One risk factor for ACL injuries is the magnitude of the knee abduction moment (KAM). Cutting technique variables on foot placement, overall approach and knee kinematics have been shown to influence the KAM. Since injury risk is believed to increase with increasing task complexity, the purpose of the study was to test the effect of task complexity on technique variables that influence the KAM in female handball players during fake-and-cut tasks.
Team description papers of magmaOffenburg are incremental in the sense that each year we address a different topic of our team and the tools around our team. In this year’s team description paper we focus on the architecture of the software. It is a main factor for being able to keep the code maintainable even after 15 years of development. We also describe how we make sure that the code follows this architecture.