Refine
Year of publication
Document Type
- Conference Proceeding (560) (remove)
Conference Type
- Konferenzartikel (342)
- Konferenz-Abstract (104)
- Konferenzband (68)
- Sonstiges (32)
- Konferenz-Poster (18)
Language
- English (361)
- German (198)
- Multiple languages (1)
Keywords
- Mikroelektronik (62)
- RoboCup (32)
- Machine Learning (9)
- injury (9)
- Ausbildung (6)
- Biomechanik (6)
- E-Learning (5)
- Herzkrankheit (5)
- Konstruktion (5)
- Produktion (5)
- biomechanics (5)
- running (5)
- ACL (4)
- Additive Manufacturing (4)
- Deep Leaning (4)
- Design (4)
- Licht (4)
- Optik (4)
- Produktentwicklung (4)
- Roboter (4)
- eco-innovation (4)
- machine learning (4)
- Additive Tooling (3)
- Artificial Intelligence (3)
- Augmented Reality (3)
- Communication Systems (3)
- Digitalisierung (3)
- Education (3)
- Generative Adversarial Network (3)
- Hochschuldidaktik (3)
- Information Systems (3)
- Katheter (3)
- Lehre (3)
- Mobile Learning (3)
- Photonik (3)
- Physik (3)
- Physikdidaktik (3)
- Robustness (3)
- Simulation (3)
- Sound Synthesis (3)
- Synchronisierung (3)
- TRIZ (3)
- VR (3)
- Virtuelle Realität (3)
- deep learning (3)
- research-oriented education (3)
- sport (3)
- Abtragung (2)
- Alexander von Humboldt (2)
- Android (2)
- Biogas (2)
- Brennstoffzelle (2)
- CRT (2)
- CST (2)
- Computer Vision (2)
- Deep Learning (2)
- Education in Optics and Photonics (2)
- Energieversorgung (2)
- Finite-Elemente-Methode (2)
- Gamification (2)
- Generative Art (2)
- HF-Ablation (2)
- Heart rhythm model (2)
- Herzrhythmusmodell (2)
- Herzschrittmacher (2)
- Humanoid Robots (2)
- IVD (2)
- Improvisation (2)
- Intelligentes Stromnetz (2)
- Kalman Filter (2)
- Kerbe (2)
- Maschinenbau (2)
- Mobile Applications (2)
- Modeling and simulation (2)
- Netzwerk (2)
- Neural networks (2)
- Optics and Photonics (2)
- Radiologie (2)
- Security (2)
- Sensortechnik (2)
- Signaltechnik (2)
- Simulation-based Interaction (2)
- Smart Grid (2)
- Software (2)
- Stability (2)
- Visual Programming (2)
- Vorhofflimmern (2)
- Wissenschaft (2)
- accountability (2)
- artificial dancer (2)
- artificial intelligence (2)
- atrial fibrillation (2)
- autoattack (2)
- cardiac ablation (2)
- content adaptation (2)
- cybersecurity (2)
- dance and technology (2)
- design methods (2)
- device independent learning (2)
- e-learning (2)
- efficient training (2)
- environmental education (2)
- explainability (2)
- fairness (2)
- heart rhythm model (2)
- interactive visualization (2)
- learning scenarios (2)
- motion synthesis (2)
- optics and photonics (2)
- overuse (2)
- process engineering (2)
- responsibility (2)
- sustainability (2)
- trust (2)
- understandability (2)
- 1.5-Degree target (1)
- 3D User Interface (1)
- 3D analysis (1)
- 3D bin picking (1)
- 3D print (1)
- 3D printing (1)
- 3D virtual reality (1)
- 3D-Modelling (1)
- 5G (1)
- AI aided Innovation (1)
- AR (1)
- AV nodal reentry tachycardia (1)
- AV reentrant tachycardia (1)
- Abwasser (1)
- Actuators (1)
- Adversarial Attacks (1)
- Adversarial Robustness (1)
- Aflatoxin (1)
- Ageing (1)
- Agile Business (1)
- Aktivierung (1)
- Aliasing (1)
- Analog-Schaltkreis (1)
- Anwenderschaltkreise (1)
- Applikation (1)
- Arbeitsmaschine (1)
- Archives (1)
- Art and Photonics (1)
- Artistic Research (1)
- Assistive Technologies (1)
- Astronomical events (1)
- Astronomical optics (1)
- Audiovisual Performance (1)
- AudiovisualPerformance (1)
- Augmented reality (1)
- Austausch (1)
- Automated Ideation (1)
- Automation (1)
- Automobilindustrie (1)
- Automotive engineering (1)
- Batteries (1)
- Bildverarbeitung (1)
- Bio-based materials (1)
- Bio-based plastics (1)
- Bioelektrochemische Brennstoffzelle (1)
- Bioimpedance measurement (1)
- Biologische Methanisierung (1)
- Biomassenutzung (1)
- Biomedizin (1)
- Blended Learning (1)
- Blockchain (1)
- Boiler (1)
- Building energy efficiency (1)
- Business Intelligence (1)
- CAAD (1)
- CAD (1)
- CAD-System (1)
- CAE (1)
- CCD-Kamera (1)
- CNNs (1)
- CRC (1)
- Car-to-Car-(C2C)-Communication (1)
- Cardiac Resynchronization Therapy (1)
- Challenges in Action Recognition (1)
- Chemie (1)
- Climate change (1)
- Cloud Computing (1)
- Cloud Security (1)
- Cloud Service Provider (1)
- Cloud User (1)
- Coal phase-out (1)
- Collaboration of Academia and Industry (1)
- Computer (1)
- Computersicherheit (1)
- Context-awareness (1)
- Context-based Services (1)
- Controlling (1)
- Creativity (1)
- Crowdsourcing-based ideation (1)
- Cryoballoon catheter ablation (1)
- Curricular concepts (1)
- Cyber Physical Systems, (1)
- DMD (1)
- Deaf-Blindness (1)
- Defibrillator (1)
- Degradability (1)
- Demand side flexibility (1)
- Design , Produktgestaltung (1)
- Design Based Research (1)
- Design Methods (1)
- Design education (1)
- Design-Structure-Matrix (1)
- Didaktisches Konzept (1)
- Dienstleistung (1)
- Digital Library (1)
- Digitale Turschilder (1)
- Digitaler Zwilling (1)
- Digitalization (1)
- Dimension 3 (1)
- Dimensional Modelling (1)
- Diskretisierung (1)
- Drehstrom (1)
- Druck (1)
- Durchhaltevermögen (1)
- E-Ink (1)
- E-Paper (1)
- EKF-SLAM (1)
- ERP (1)
- Ecercises (1)
- Ecodesign (1)
- Edge AI (1)
- Eigenvalues (1)
- Einführung (1)
- Eingebettetes System (1)
- Electrochemistry (1)
- Elektrizitätserzeugung (1)
- Elektrofahrzeug (1)
- Elektrokardiogramm (1)
- Embedded AI (1)
- Embedded Software (1)
- Embedded Systems (1)
- Energiemanagement (1)
- Energiepolitik (1)
- Energiewirtschaft (1)
- Energy Harvesting (1)
- Energy Harvesting (EH) (1)
- Energy Market (1)
- Energy Planning (1)
- Energy Storage Systems (1)
- Energy System Analysis (1)
- Energy System Optimization (1)
- Energy management (1)
- Energy policy (1)
- Energy systems modeling (1)
- Engineering Creativity (1)
- Engineering creativity (1)
- Engineering education in the age of digitalization (1)
- Entity Resolution (1)
- Enzym (1)
- Epilepsie (1)
- Ergebnissicherung (1)
- Erweiterte Realität <Informatik> (1)
- Erziehung (1)
- Esophageal catheter (1)
- Eye Tracking (1)
- Eye tracking (1)
- Eye-Tracking (1)
- FEM (1)
- FPGA (1)
- Fahrzeug (1)
- Faseroptik (1)
- Fault detection (1)
- Federated Learning (1)
- Fehler (1)
- Fehleranalyse (1)
- Fiber reinforcement (1)
- Finite-Differenzen Methode (1)
- Flashcards (1)
- Flexibilisierung des Lernortes (1)
- Flüssigkristall (1)
- Formzahlen (1)
- GIS (1)
- GPS-Empfänger (1)
- Gasanalyse (1)
- Gebäude (1)
- Gedruckte Elektronik (PE) (1)
- Generative Design (1)
- Gesellschaft (1)
- Gestaltung (1)
- Getriebewelle (1)
- Gewerbebetrieb (1)
- Gleichstrom (1)
- Google Maps (1)
- Grid Integration (1)
- HR (1)
- Halbkreisnut (1)
- Hardmakros (1)
- Haustechnik (1)
- Head-mounted displays (1)
- Heart Rhythm Simulation (1)
- Heat pumps (1)
- Heilkunde (1)
- Helmholtz coil (1)
- Hemodynamic monitoring (1)
- Herstellung (1)
- Herz (1)
- Herzmuskelkrankheit (1)
- Herzrhythmusstörung (1)
- His-Bundle Pacing (1)
- Hochfrequenzablation (1)
- Hochfrequenztechnik (1)
- Hochschule (1)
- Hochspannung (1)
- Holzpresssaft (1)
- Human Resources (1)
- Humanoider Roboter (1)
- Hybridantrieb (1)
- IDL (1)
- IYL (1)
- Identity Resolution (1)
- Immersive Technology (1)
- InceptionTime (1)
- Industrie 4.0 (1)
- Ingenieuerwissenschaften (1)
- Injection Molding (1)
- Injection Molds (1)
- Innovation (1)
- Integrated product development (1)
- Intellectual Properties (1)
- Intelligent Buildings (1)
- Interactive Film (1)
- International Day of Light (1)
- International Day of Light, IDL (1)
- International Year of Light (1)
- International Year of Light, IYL (1)
- Internet der Dinge (1)
- Internet of Things (1)
- Internet portal (1)
- KAM (1)
- Karbon (1)
- Kardiale Resynchronisationstherapie (1)
- Kardiologie (1)
- Kernspintomografie (1)
- Klausur (1)
- Klein- und Mittelbetrieb (1)
- Knowledge-based Innovation (1)
- Knowledge-based innovation (1)
- Kommunikation (1)
- Kryoballonkatheter (1)
- Kryptographie (1)
- LCA-Microcontroller (1)
- LPWAN (1)
- Landau-Lifshitz-Gilbert equations (1)
- Language learning (1)
- Leap Motion Controller (1)
- Left Atrial Appendage Closure (1)
- Lernmanagementsystem (1)
- Lernsoftware (1)
- Lerntrails (1)
- Lernumgebung (1)
- Lernzieltaxonomie (1)
- Lightweight design (1)
- Lithium-ion battery (1)
- Live Broadcasting (1)
- Location-Based Services (1)
- Location-based Services (1)
- Logic-Cell-Array (1)
- Lottozahlengenerators (1)
- Low-Cost Füllstandssensor (1)
- Low-Power (1)
- Low-Power-SoC-Systeme (1)
- Lückenskript (1)
- M-learning (1)
- MEMS (1)
- MLOps (1)
- MPC (1)
- Machine learning (1)
- Machine-to- Machine-(M2M)-Communication (1)
- Makespan (1)
- Management (1)
- Maschinenbaustudium (1)
- Mathematik (1)
- Mathematiktrails (1)
- Mathematikunterricht (1)
- Mathtrails (1)
- Medizintechnik (1)
- Mehrfachkerbwirkung (1)
- Mentor-Graphics (1)
- Methanol (1)
- Microelectronics (1)
- Mikrocontroller (1)
- Mikroprozessorkernels (1)
- Mode Collapse (1)
- Model Calibration (1)
- Model Search (1)
- Modellierung und Simulation (1)
- Modelling (1)
- Movement (1)
- Multi-Material 3D-Printing (1)
- NB-IoT (1)
- NFC (1)
- Nachhaltigkeit (1)
- Natural fibers (1)
- Naturwissenschaften (1)
- Network Test (1)
- New Product Development (1)
- New product development (1)
- Nickel (1)
- Niedrige Energie (1)
- Nonlinear Regression (1)
- Nyquist-Shannon (1)
- Object-Based Services (1)
- Octave Convolution (1)
- Onboarding (1)
- Online Student Services (1)
- Online simulation (1)
- Ontology-based Classification (1)
- Optimization and control (1)
- Oxidation (1)
- Particle emissions (1)
- Partnerschaft (1)
- Passivhaus (1)
- Pattern Recognition (1)
- Petri-Netz (1)
- Phontonik (1)
- Photonics (1)
- Photovoltaic (1)
- Physiological Pacing (1)
- Physiotherapy (1)
- Plant commissioning (1)
- Planung (1)
- PolyJet Modelling (1)
- Predictive Maintenance (1)
- Privatsphäre (1)
- Problemlösen (1)
- Process intensification (1)
- Product Design (1)
- Produktionsprozess (1)
- Projektmanagement (1)
- Prothesen (1)
- Prototyp (1)
- Prozessablauf (1)
- Prozessorkern (1)
- Präsentation (1)
- Prüfung (1)
- Pulmonalvenenisolation (1)
- Qualität (1)
- Qualitätskontrolle (1)
- Qualitätssicherung (1)
- Querbohrung (1)
- Rapid Tooling (1)
- Realtime simulation (1)
- Recommendation Techniques (1)
- Regularization (1)
- Rehabilitation (1)
- Reinforcement Learning (1)
- Representation Learning (1)
- ResNet (1)
- Road-Quality Prediction (1)
- Robot Applications (1)
- Robotic Soccer (1)
- Robotics (1)
- Robots (1)
- Russian Ukrainian war (1)
- SAP Analytics Cloud (1)
- SAP Data Warehouse Cloud (1)
- Sampling (1)
- Schaltkreis (1)
- Schaltungsdesign (1)
- Schlatungen (1)
- Screencast (1)
- Second-order Optimization (1)
- Selbsttest (1)
- Selbsttests (1)
- Sensors (1)
- Sensortechniik (1)
- Sequential Model-Based Optimization (1)
- Sicherheitstechnik (1)
- Simulationsversuch (1)
- Smart Gardening (1)
- Smart Home (1)
- Smart wearables (1)
- Social Robots (1)
- Soft- und Hardcore-Prozessoren (1)
- Soziale Roboter (1)
- Spannung (1)
- Spinal cord stimulation (1)
- Statistik (1)
- Statistikvideos (1)
- Stochastik (1)
- Studium (1)
- Supraventricular tachycardia (1)
- Synchronisationstherapie (1)
- System-on-Chip (1)
- TLS (1)
- TRIZ Inventive Principles (1)
- TRIZ methodology (1)
- TTCN3 (1)
- Tablet (1)
- Tablet Lückenskript (1)
- Technische Mechanik (1)
- Technischer Bericht (1)
- Telemetrie (1)
- Testumgebung (1)
- Thermische Solaranlage (1)
- Thermische Solarenergie (1)
- Thermologger (1)
- Time-series Classification (1)
- TinyML (1)
- Traceability (1)
- Transistor (1)
- Umwelt (1)
- Unsupervised Conditional Training (1)
- Unsupervised Learning (1)
- Unternehmensführung (1)
- User behavior (1)
- Variational Autoencoders (1)
- Vehicle safety (1)
- Verfahren (1)
- Vergleich (1)
- Verkehr (1)
- Versorgung (1)
- Verwaltung (1)
- Verzahnung (1)
- Videoclip (1)
- Virtual Reality (1)
- Virtual Research Environment (1)
- Virtual experiments (1)
- Virtual reality (1)
- Virtuelles Laboratorium (1)
- Visualization (1)
- Voxel (1)
- Voxelization (1)
- Wasserstand (1)
- Wearables (1)
- Weitsprung (1)
- Wertanalyse (1)
- Wireless Sensor Nodes (1)
- Wirtschaftswissenschaften (1)
- Wishbone (1)
- Wärmepumpe (1)
- Zellbibliothek (1)
- Zusatzstoff (1)
- accelerometer (1)
- accuracy (1)
- acoustic phonons (1)
- adversarial (1)
- algorithm-based data analysis (1)
- atrial flutter (1)
- attribute manipulation (1)
- autoML (1)
- benchmarking (1)
- biomechanical stimulation (1)
- biomimetics (1)
- business applications (1)
- cantilever (1)
- cifar (1)
- cloud (1)
- cloud computing (1)
- cluster (1)
- collaborative learning (1)
- content formatting (1)
- content synchronization (1)
- cross hole (1)
- cross-industry innovation (1)
- cryptography (1)
- curricular concepts (1)
- data architecture (1)
- data migration (1)
- data model (1)
- defense (1)
- design education (1)
- device detection (1)
- device independency (1)
- dickkopf 3 (1)
- differential mode-delay (1)
- digital games (1)
- digital library (1)
- digital twin (1)
- dynamic hyperlinks (1)
- eLearning (1)
- echocardiography (1)
- education and research (1)
- elektrische Felder (1)
- elektronischer Würfel (1)
- energy harvesting (1)
- equivalent circuit model (1)
- ethical frameworks (1)
- evaluation (1)
- exchange magnons (1)
- face editing (1)
- ferromagnetic resonance (1)
- flashcard (1)
- flexible job shop (1)
- force (1)
- force controlled robot (1)
- fourier (1)
- gamification (1)
- gan (1)
- generalized content (1)
- generation Y (1)
- generation Z (1)
- genetic algorithms (1)
- glass (1)
- grey-box model (1)
- gyroscope (1)
- heart rhythm simulation (1)
- heating and cooling (1)
- high-speed cameras (1)
- humanoid robot walking (1)
- iSign (1)
- imagenet (1)
- impedance cardiography (1)
- industry (1)
- inertial measurement unit (1)
- information network (1)
- innovation management (1)
- interactive media (1)
- interconnected data (1)
- joint torque sensors (1)
- kardiale Resynchronisationstherapie (1)
- language learning (1)
- laser material processing (1)
- laser triangulation (1)
- lid (1)
- lithium-ion battery (1)
- load profiles (1)
- local electricity markets (1)
- loneliness (1)
- m-learning (1)
- magnetization dynamics (1)
- magneto-acoustics (1)
- mahalanobis (1)
- maintenance scheduling (1)
- master data (1)
- measurement (1)
- media technology (1)
- media tyechnolog (1)
- mental health apps (1)
- metaverse (1)
- mobile learning (1)
- mobile web design guidelines (1)
- multi-material (1)
- multidimensional flashcard (1)
- multimode fibre (1)
- multimode fibre connectors (1)
- multiple notch-effective (1)
- natürliche Leistung (1)
- neural architecture search (1)
- neural ordinary differential equations (1)
- new product development (1)
- non interfering measurements (1)
- offensive security techniques (1)
- optical on-line/in-line measurements (1)
- optimization (1)
- peer-to-peer energy trading (1)
- pigment paste (1)
- powder filled gripper (1)
- power distribution (1)
- presentation skills (1)
- printing technologies (1)
- project-based learning (1)
- pruning (1)
- pulmonary vein isolation (1)
- quality function deployment (QFD) (1)
- railway system (1)
- reference data (1)
- remanufacturing (1)
- remote laboratory (1)
- remote model-based laboratory (1)
- repeatability (1)
- risk factor (1)
- robotics (1)
- scanning electron microscope (SEM) (1)
- scheduling (1)
- semicircular groove (1)
- service engineering (1)
- service research (1)
- signal averaging (1)
- simulation (1)
- smart grid (1)
- sparse backpropagation (1)
- spatial imagination (1)
- spectral defense (1)
- spectraldefense (1)
- stress concentration factors (1)
- style transfer (1)
- surface treatment (1)
- synthetical profiles (1)
- taxonomy (1)
- teaching and learning culture (1)
- technical drawings (1)
- temperature dependency (1)
- thermische Felder (1)
- transversal skills (1)
- ultrafast laser interaction with materials (1)
- vacuum (1)
- ventricular tachycardia (1)
- vibration (1)
- virtual lab (1)
- virtual reality (1)
- virtual worlds (1)
- wavelet transformation (1)
- Überwachung (1)
- übertragbare Wirkleistung (1)
Institute
- Fakultät Elektrotechnik und Informationstechnik (E+I) (bis 03/2019) (234)
- Fakultät Maschinenbau und Verfahrenstechnik (M+V) (139)
- Fakultät Elektrotechnik, Medizintechnik und Informatik (EMI) (ab 04/2019) (75)
- Fakultät Wirtschaft (W) (51)
- Fakultät Medien und Informationswesen (M+I) (bis 21.04.2021) (46)
- INES - Institut für nachhaltige Energiesysteme (30)
- IMLA - Institute for Machine Learning and Analytics (29)
- ivESK - Institut für verlässliche Embedded Systems und Kommunikationselektronik (19)
- Fakultät Medien (M) (ab 22.04.2021) (17)
- ACI - Affective and Cognitive Institute (8)
Open Access
- Open Access (560) (remove)
Plastics are used today in many areas of the automotive, aerospace and mechanical engineering industries due to their lightweight potential and ease of processing. Additive manufacturing is applied more and more frequently, as it offers a high degree of design freedom and eliminates the need for complex tools. However, the application of additively manufactured components made of plastics have so far been limited due to their comparatively low strength. For this reason, processes that offer additional reinforcement of the plastic matrix using fibers made of high-strength materials have been developed. However, these components represent a composite of different materials produced on the basis of fossil raw materials, which are difficult to recycle and generally not biodegradable.
Therefore, this paper will explore the potential for new composite materials whose matrix consists of a bio-based plastic. In this investigation, it is assumed that the matrix is reinforced with a fibrous material made of natural fiber to significantly increase the strength. This potential material should offer a lightweight yet strong structure and be biodegradable after use under controlled conditions. Therefore, the state of the art in the use of bio-based materials in 3D printing is first presented. In order to determine the economic boundary conditions, the growth potentials for bio-based materials are analyzed. Also, the recycling prospects for bio-based plastics will also be highlighted. The greenhouse gas emissions and land use to be expected when using bio-based materials are also estimated. Finally, the degradability of the composites is discussed.
Team description papers of magmaOffenburg are incremental in the sense that each year we address a different topic of our team and the tools around our team. In this year’s team description paper we focus on the architecture of the software. It is a main factor for being able to keep the code maintainable even after 15 years of development. We also describe how we make sure that the code follows this architecture.
Selbsttests in Lernmanagementsystemen (LMS) ermöglichen es Studierenden, den eigenen Lernfortschritt einzuschätzen. Im Gegensatz zur Einreichung und Korrektur vollständig ausformulierter Aufgabenlösungen nutzen LMS überwiegend die Eingabe der Lösung im Antwort-Auswahl-Verfahren (Single-Choice). Nach didaktischen Ansatz „Physik durch Informatik“ geben die Lernenden stattdessen ihre Aufgabenlösungen in einer Programmiersprache ins LMS ein, was eine automatisierte Rückmeldung erleichtert und das Erreichen einer höheren Kompetenzstufe fördert. Es wurden zehn LMS-Selbsttests erstellt, bei denen die Lösungen zu einer Lehrbuch-Aufgabenstellung jeweils durch Eingabe in einer Programmiersprache und von einer Kontrollgruppe im Antwort-Auswahl-Verfahren abgefragt wurden. Ergebnisse aus dem ersten Einsatz dieser Selbsttests für die Lehrveranstaltung Physik im Studiengang Biotechnologie werden vorgestellt.
Mathematik lässt sich in vielen Objekten finden. Sei es die lineare Steigung eines Handlaufs zum Schulgebäude oder die nahezu zylindrische Form einer Litfaßsäule in der Innenstadt. Das Bestreben, Schüler*innen diese Zusammenhänge entdecken zu lassen, steht im Zentrum des MathCityMap Projekts (Ludwig et al., 2013). Auf sogenannten mathematischen Wanderpfaden (bzw. Mathtrails) werden Schüler*innen durch eine App zu Mathematikaufgaben an realen Objekten bzw. in realen Situationen ihrer Umwelt geleitet. Um die Aufgaben zu lösen, werden Daten erhoben, z. B. durch Messungen oder Zählen. Entscheidend ist, dass die Aufgaben so gestellt sind, dass der Schritt der Datenbeschaffung nur vor Ort stattfinden kann und somit direkt mit dem Objekt bzw. der Situation verknüpft wird.
Enhancing engineering creativity with automated formulation of elementary solution principles
(2023)
The paper describes a method for the automated formulation of elementary creative stimuli for product or process design at different levels of abstraction and in different engineering domains. The experimental study evaluates the impact of structured automated idea generation on inventive thinking in engineering design and compares it with previous experimental studies in educational and industrial settings. The outlook highlights the benefits of using automated ideation in the context of AI-assisted invention and innovation.
Neural networks have a number of shortcomings. Amongst the severest ones is the sensitivity to distribution shifts which allows models to be easily fooled into wrong predictions by small perturbations to inputs that are often imperceivable to humans and do not have to carry semantic meaning. Adversarial training poses a partial solution to address this issue by training models on worst-case perturbations. Yet, recent work has also pointed out that the reasoning in neural networks is different from humans. Humans identify objects by shape, while neural nets mainly employ texture cues. Exemplarily, a model trained on photographs will likely fail to generalize to datasets containing sketches. Interestingly, it was also shown that adversarial training seems to favorably increase the shift toward shape bias. In this work, we revisit this observation and provide an extensive analysis of this effect on various architectures, the common L_2-and L_-training, and Transformer-based models. Further, we provide a possible explanation for this phenomenon from a frequency perspective.
Sweaty has already participated several times in RoboCup soccer competitions (Adult Size). Now the work is focused coordinating the play of two robots. Moreover, we are working on stabilizing the gait by adding additional sensor information. An ongoing work is the optimization of the control strategy by balancing between impedance and position control. By minimizing the jerk, gait and overall gameplay should improve significantly.
Established robot manufacturers have developed methods to determine and optimize the accuracy of their robots. These methods vary from robot manufacturers to their competitors. Due to the lack of published data, a comparison of robot performance is difficult. The aim of this article is to find methods to evaluate important characteristics of a robot with an accurate and cost-effective setup. A laser triangulation sensor and geometric referenced spheres were used as a base to compare the robot performance.
Additive manufacturing enables the production of lightweight and resilient components with extensive design freedom. In the low-cost sector, material extrusion (e.g. Fused Deposition Modeling - FDM) has been the main method used to date. Thus, robust 3D printers and inexpensive 3D materials (polymer filaments) can be used. However, the printing times for FDM are very long and the quality of the dimensions and surfaces is limited. Recently, new processes from the field of Vat polymerization have entered the market. For example, masked stereolithography (mSLA) offers a significant improvement in component quality and build speed through the use of resins and large-area curing at still reasonable costs. Currently, there is only limited knowledge available on the optimal design of components using this young process. In this contribution, design guidelines are developed to determine the possibilities and limitations of mSLA from a design point of view. For this purpose, a number of test geometries are designed and investigated to obtain systematic insights into important design features, such as wall thickness, grooves and holes. In addition, typical problems in additive manufacturing, such as the design of overhangs and fits or the hollowing of components, are investigated. The evaluation of practical 3D printing tests thus provides important parameters that can be transferred to design guidelines of components for additive manufacturing using mSLA.
In 2015, Google engineer Alexander Mordvintsev presented DeepDream as technique to visualise the feature analysis capabilities of deep neural networks that have been trained on image classification tasks. For a brief moment, this technique enjoyed some popularity among scientists, artists, and the general public because of its capability to create seemingly hallucinatory synthetic images. But soon after, research moved on to generative models capable of producing more diverse and more realistic synthetic images. At the same time, the means of interaction with these models have shifted away from a direct manipulation of algorithmic properties towards a predominance of high level controls that obscure the model's internal working. In this paper, we present research that returns to DeepDream to assess its suit-ability as method for sound synthesis. We consider this research to be necessary for two reasons: it tackles a perceived lack of research on musical applications of DeepDream, and it addresses DeepDream's potential to combine data driven and algorithmic approaches. Our research includes a study of how the model architecture, choice of audio data-sets, and method of audio processing influence the acoustic characteristics of the synthesised sounds. We also look into the potential application of DeepDream in a live-performance setting. For this reason, the study limits itself to models consisting of small neural networks that process time-domain representations of audio. These models are resource-friendly enough to operate in real time. We hope that the results obtained so far highlight the attractiveness of Deep-Dream for musical approaches that combine algorithmic investigation with curiosity driven and open ended exploration.
Digital, virtual environments and the metaverse are rapidly taking shape and will generate disruptive changes in the areas of ethics, privacy, safety, and how the relationships between human beings will be developed. To uncover some of some of the implications that will impact those areas, this study investigates the perceptions of 101 younger people from the generations Y and Z. We present a first exploratory analysis of the findings, focusing on knowledge and self-perception. Results show that these young generations are seriously doubting their knowledge on the metaverse and virtual worlds – regarding both the definition and the usage. It is interesting to see only a medium confidence level, considering that the participants are young and from an academic environment, which should increase their interest in and the affinity towards virtual worlds. Males from both generations perceive themselves as significantly more knowledgeable than females. Regarding a fitting definition, almost 40% agreed on the metaverse as a “universal and immersive virtual world that is made accessible using virtual reality and augmented reality technologies”. Regarding the topic in general, several participants (almost 40%) considered themselves sceptics or “just” users (38%). Interestingly, generation Y participants were more likely than the younger generation Z participants to identify themselves as early adopters or innovators. In result, the considerable amount of “mixed feelings” regarding digital, virtual environments and the metaverse shows that in-depth studies on the perception of the metaverse as well as its ethical and integrity implications are required to create more accessible, inclusive, safe, and inclusive digital, virtual environments.
Convolutional neural networks (CNN) define the state-of-the-art solution on many perceptual tasks. However, current CNN approaches largely remain vulnerable against adversarial perturbations of the input that have been crafted specifically to fool the system while being quasi-imperceptible to the human eye. In recent years, various approaches have been proposed to defend CNNs against such attacks, for example by model hardening or by adding explicit defence mechanisms. Thereby, a small “detector” is included in the network and trained on the binary classification task of distinguishing genuine data from data containing adversarial perturbations. In this work, we propose a simple and light-weight detector, which leverages recent findings on the relation between networks’ local intrinsic dimensionality (LID) and adversarial attacks. Based on a re-interpretation of the LID measure and several simple adaptations, we surpass the state-of-the-art on adversarial detection by a significant m argin and reach almost perfect results in terms of F1-score for several networks and datasets. Sources available at: https://github.com/adverML/multiLID
In this contribution, we present a novel 3D printed multi-material, electromagnetic vibration harvester. The harvester is based on a cantilever design and utilizes an embedded constantan wire within a matrix of polyethylene terephthalate glycol (PETG). A prototype has been manufactured with a combination of a fused filament fabrication (FFF) printer and a robot with a custom-made tool.
In this paper, we describe a first publicly available fine-grained product recognition dataset based on leaflet images. Using advertisement leaflets, collected over several years from different European retailers, we provide a total of 41.6k manually annotated product images in 832 classes. Further, we investigate three different approaches for this fine-grained product classification task, Classification by Image, by Text, as well as by Image and Text. The approach "Classification by Text" uses the text extracted directly from the leaflet product images. We show, that the combination of image and text as input improves the classification of visual difficult to distinguish products. The final model leads to an accuracy of 96.4% with a Top-3 score of 99.2%. We release our code at https://github.com/ladwigd/Leaflet-Product-Classification.
It is common practice to apply padding prior to convolution operations to preserve the resolution of feature-maps in Convolutional Neural Networks (CNN). While many alternatives exist, this is often achieved by adding a border of zeros around the inputs. In this work, we show that adversarial attacks often result in perturbation anomalies at the image boundaries, which are the areas where padding is used. Consequently, we aim to provide an analysis of the interplay between padding and adversarial attacks and seek an answer to the question of how different padding modes (or their absence) affect adversarial robustness in various scenarios.
Artificial Intelligence (AI) can potentially transform many aspects of modern society in various ways, including automation of tasks, personalization of products and services, diagnosis of diseases and their treatment, transportation, safety, and security in public spaces, etc. Recently, AI technology has been transforming the financial industry, offering new ways to analyse data and automate processes, reduce costs, increase efficiency, and provide more personalized services to customers. However, it also raised important ethical and regulatory questions that need to be addressed by the industry and society as a whole. The aim of the Erasmus+ project Transversal Skills in Applied Artificial Intelligence - TSAAI (KA220-HED - Cooperation Partnerships in higher education) has been to establish a training platform that will incorporate teaching guidelines based on a curriculum covering different areas of application of AI technology. In this work, we will focus on applying AI models in the financial and insurance sectors.
While most ultrafast time-resolved optical pump-probe experiments in magnetic materials reveal the spatially homogeneous magnetization dynamics of ferromagnetic resonance (FMR), here we explore the magneto-elastic generation of GHz-to-THz frequency spin waves (exchange magnons). Using analytical magnon oscillator equations, we apply time-domain and frequency-domain approaches to quantify the results of ultrafast time-resolved optical pump-probe experiments in free-standing ferromagnetic thin films. Simulations show excellent agreement with the experiment, provide acoustic and magnetic (Gilbert) damping constants and highlight the role of symmetry-based selection rules in phonon-magnon interactions. The analysis is extended to hybrid multilayer structures to explore the limits of resonant phonon-magnon interactions up to THz frequencies.
Variable refrigerant flow (VRF) and variable air volume (VAV) systems are considered among the best heating, ventilation, and air conditioning systems (HVAC) thanks to their ability to provide cooling and heating in different thermal zones of the same building. As well as their ability to recover the heat rejected from spaces requiring cooling and reuse it to heat another space. Nevertheless, at the same time, these systems are considered one of the most energy-consuming systems in the building. So, it is crucial to well size the system according to the building’s cooling and heating needs and the indoor temperature fluctuations. This study aims to compare these two energy systems by conducting an energy model simulation of a real building under a semi-arid climate for cooling and heating periods. The developed building energy model (BEM) was validated and calibrated using measured and simulated indoor air temperature and energy consumption data. The study aims to evaluate the effect of these HVAC systems on energy consumption and the indoor thermal comfort of the building. The numerical model was based on the Energy Plus simulation engine. The approach used in this paper has allowed us to reach significant quantitative energy saving along with a high level of indoor thermal comfort by using the VRF system compared to the VAV system. The findings prove that the VRF system provides 46.18% of the annual total heating energy savings and 6.14% of the annual cooling and ventilation energy savings compared to the VAV system.
Training deep neural networks using backpropagation is very memory and computationally intensive. This makes it difficult to run on-device learning or fine-tune neural networks on tiny, embedded devices such as low-power micro-controller units (MCUs). Sparse backpropagation algorithms try to reduce the computational load of on-device learning by training only a subset of the weights and biases. Existing approaches use a static number of weights to train. A poor choice of this so-called backpropagation ratio limits either the computational gain or can lead to severe accuracy losses. In this paper we present TinyProp, the first sparse backpropagation method that dynamically adapts the back-propagation ratio during on-device training for each training step. TinyProp induces a small calculation overhead to sort the elements of the gradient, which does not significantly impact the computational gains. TinyProp works particularly well on fine-tuning trained networks on MCUs, which is a typical use case for embedded applications. For typical datasets from three datasets MNIST, DCASE2020 and CIFAR10, we are 5 times faster compared to non-sparse training with an accuracy loss of on average 1%. On average, TinyProp is 2.9 times faster than existing, static sparse backpropagation algorithms and the accuracy loss is reduced on average by 6 % compared to a typical static setting of the back-propagation ratio.
Currently, immersive technologies are enjoying great popularity. This trend is reflected in technological advances and the emergence of new products for the mass market, such as augmented reality glasses. The range of applications for immersive technologies is growing with more efficient and affordable technologies and student adoption. Especially in education, the use will improve existing learning methods. Immersive application use visual, audio and haptic sensors to fully engage the user in a virtual environment. This impression is reinforced with the help of realistic visualizations and the opportunity for interaction. In particular, Augmented reality is characterized by a high degree of integration between reality and the inserted virtual objects. An augmented interactive simulation for the determination of the specific charge of an electron will be used as an example to demonstrate how such immersion can be created for users. A virtual Helmholtz coil is used to measure and calculate the e/m constant. The voltage at the cathode for generating the electron beam, but also the voltage of the homogeneous magnetic field for deflecting the electron beam, can be variably controlled by haptic user input. Based on these voltages, an immersive virtual electron beam is calculated and visualized. In this paper, the authors present the conceptual steps of this immersive application and address the challenges associated with designing and developing an augmented and interactive simulation.
Redesigning a curriculum for teaching media technology is a major challenge. Up-to-date teaching and learning concepts are necessary that meet the constant technological progress and prepare students specifically for their professional life. Teaching and studying should be characterized by a student-oriented teaching and learning culture. In order to achieve this goal, consistent evaluation is essential. The aim of the evaluation concept presented here is to generate structured information regarding the quality of content-related, didactic and organizational aspects of teaching. The exchange of opinions between students and lecturers should be encouraged in order to continuously improve the teaching and learning processes.
The paper will focus on the activities of the International Year of Light and Optical Technologies 2015 (IYL) with their impact in life, science, art, culture, education and outreach as well as the importance in promoting the objectives for sustainable development. It describes our activities carried out in the run-up to or during the IYL, as well as reports on the generic projects that led to the success of the IYL. The success of the IYL is illustrated by examples and statistics. Relating to the potential and success of the IYL, the impact and the genesis of the International Day of Light (IDL) is presented. Impressions from the opening ceremony of the IYL in Paris at UNESCO headquarters and the Inaugural Ceremony of the IDL will then be covered. A second focus is placed on the interdisciplinary media projects realized by the students of our university dedicated to these events. Finally, an analysis of the impact and legacy of IYL and IDL will be presented.
3D Bin Picking with an innovative powder filled gripper and a torque controlled collaborative robot
(2023)
A new and innovative powder filled gripper concept will be introduced to a process to pick parts out of a box without the use of a camera system which guides the robot to the part. The gripper is a combination of an inflatable skin, and a powder inside. In the unjammed condition, the powder is soft and can adjust to the geometry of the part which will be handled. By applying a vacuum to the inflatable skin, the powder gets jammed and transforms to a solid shaped form in which the gripper was brought before applying the vacuum. This physical principle is used to pick parts. The flexible skin of the gripper adjusts to all kinds of shapes, and therefore, can be used to realize 3D bin picking. With the help of a force controlled robot, the gripper can be pushed with a consistent force on flexible positions depending of the filling level of the box. A Kuka LBR iiwa with joint torque sensors in all of its seven axis’ was used to achieve a constant contact pressure. This is the basic criteria to achieve a robust picking process.
The use of artificial intelligence continues to impact a broad variety of domains, application areas, and people. However, interpretability, understandability, responsibility, accountability, and fairness of the algorithms' results - all crucial for increasing humans' trust into the systems - are still largely missing. The purpose of this seminar is to understand how these components factor into the holistic view of trust. Further, this seminar seeks to identify design guidelines and best practices for how to build interactive visualization systems to calibrate trust.
Landing heel first has been associated with elevated external knee abduction moments (KAM), thereby potentially increasing the risk of sustaining a non-contact ACL injury. Apart from the foot strike angle, knee valgus angle (VAL) and vertical center of mass velocity at initial ground contact (IC) have been associated with increased KAM in females across different sidestep cuts. While real-time biofeedback training has been proven effective for gait retraining [4], the highly dynamic, non-cyclical nature of cutting maneuvers makes real-time feedback unsuitable and alternative approaches necessary. This study aimed at assessing the efficacy of immediate software-aided feedback on cutting technique in reducing KAM during handball-specific cutting maneuvers.
Due to globalization and the resulting increase in competition on the market, products must be produced more and more cheaply, especially in series production, because buyers expect new variants or even completely new products in ever shorter cycles. Injection molding is the most important production process for manufacturing plastic components in large quantities. However, the conventional production of a mold is extremely time-consuming and costly, which creates a contradiction to the fast pace of the market. Additive tooling is an area of application of additive manufacturing, which in the field of injection molding is preferably used for the prototype production of mold inserts. This allows injection molding tools to be produced faster and more cheaply than through the subtractive manufacturing of metal tools. Material Jetting processes using polymers (MJT-UV/P), also called Polyjet Modeling (PJM), have a great potential for use in additive tooling. Due to the poorer mechanical and thermal properties compared to conventional mold insert materials, e.g. steel or aluminum, the previously used design principles cannot be applied. Accordingly, new design guidelines are necessary, which are developed in this paper. The necessary information is obtained with the help of a systematic literature research. The design guidelines are mapped in a uniform design guide, which is structured according to the design process of injection molds. The guidelines do not only refer to the constructive design of the injection mold or the polymer mold insert, but to the entire design process and describe the four phases of planning, conception, development and realization. Particular attention is paid to the special geometric designs of a polymer mold insert and the thermomechanical properties of the mold insert materials. As a result, design guidelines are available that are adapted to the special requirements of additive tooling of molds inserts made of plastics for injection molding.
The main advantage of mobile context-aware applications is to provide effective and tailored services by considering the environmental context, such as location, time, nearby objects and other data, and adapting their functionality according to the changing situations in the context information without explicit user interaction. The idea behind Location-Based Services (LBS) and Object-Based Services (OBS) is to offer fully-customizable services for user needs according to the location or the objects in a mobile user's vicinity. However, developing mobile context-aware software applications is considered as one of the most challenging application domains due to the built-in sensors as part of a mobile device. Visual Programming Languages (VPL) and hybrid visual programming languages are considered to be innovative approaches to address the inherent complexity of developing programs. The key contribution of our new development approach for location and object-based mobile applications is a use case driven development approach based on use case templates and visual code templates to enable even programming beginners to create context-aware mobile applications. An example of the use of the development approach is presented and open research challenges and perspectives for further development of our approach are formulated.
Sensors and actuators enable creation of context-aware applications in which applications can discover and take advantage of contextual information, such as user location, nearby people and objects. In this work, we use a general context definition, which can be applied to various devices, e.g., robots and mobile devices. Developing context-based software applications is considered as one of the most challenging application domains due to the sensors and actuators as part of a device. We introduce a new development approach for context-based applications by using use-case descriptions and Visual Programming Languages (VPL). The introduction of web-based VPLs, such as Scratch and Snap, has reinvigorated the usefulness of VPLs. We provide an in-depth discussion of our new VPL based method, a step by step development process to enable development of context-based applications. Two case studies illustrate how to apply our approach to different problem domains: Context-based mobile apps and context-based humanoid robot applications.
The variable refrigerant flow system is one of the best heating, ventilation, and air conditioning systems (HVAC) thanks to its ability to provide thermal comfort inside buildings. But, at the same time, these systems are considered one of the most energy-consuming systems in the building sector. Thus, it is crucial to well size the system according to the building’s cooling and heating needs and the indoor temperature fluctuations. Although many researchers have studied the optimization of the building energy performance considering heating or cooling needs, using air handling units, radiant floor heating, and direct expansion valves, few studies have considered the use of multi-objective optimization using only the thermostat setpoints of VRF systems for both cooling and heating needs. Thus, the main aim of this study is to conduct a sensitivity analysis and a multi-objective optimization strategy for a residential building containing a variable refrigerant flow system, to evaluate the effect of the building performance on energy consumption and improve the building energy efficiency. The numerical model was based on the EnergyPlus, jEPlus, and jEPlus+EA simulation engines. The approach used in this paper has allowed us to reach significant quantitative energy saving by varying the cooling and heating setpoints and scheduling scenarios. It should be stressed that this approach could be applied to several HVAC systems to reduce energy-building consumption.
Die Positionierung mobiler Systeme mit hoher Genauigkeit ist eine Voraussetzung für intelligentes autonomes Verhalten, sowohl in der Feldrobotik als auch in industriellen Umgebungen. Dieser Beitrag beschreibt den Aufbau einer Roboterplattform und ihre Verwendung für den Test und die Bewertung von Kalman-Filter-Konfigurationen. Der Aufbau wurde mit einem mobilen Roboter Husky A200 und einem LiDAR-Sensor (Light Detection and Ranging) realisiert. Zur Verifizierung des vorgeschlagenen Aufbaus wurden fünf verschiedene Szenarien ausgearbeitet. Mit denen wurden die Filter auf ihre Leistungsfähigkeit hinsichtlich der Genauigkeit der Positionsbestimmung getestet.
Evaluierung von Kalman Filter Konfigurationen zur Roboterlokaliserung mittels Sensordatenfusion
(2023)
In dieser Arbeit werden drei verschiedene Konfigurationen der von Tom Moore, für das Robot Operating System, entwickelte Kalman-Filter vorgestellt. Diese bilden die Grundlage für eine Lokalisierung mittels Sensorfusion in dem verwendeten ROS-Framework. Ziel dieser Arbeit ist der Aufbau und die Verifikation einer Lokalisierung für ein mobiles Robotersystem Husky A200 der Firma Clearpath Robotics. Hierzu wurden die Möglichkeiten des bestehenden Systems untersucht und mehrere Versionen von Lokalisierungsfiltern konfiguriert. Am an Ende, wird eine Verifikation der Ergebnisse in verschiedenen Szenarien gegeneinandergestellt. Hierzu werden die Ergebnisse einer Variante des Extended Kalman-Filters in 2D (EKF2D), eine Variante des Unscented Kalman-Filter in 2D (UKF2D) und eine Variante des Extended Kalman-Filters in 3D (EKF3D) verifiziert und verglichen. Die Untersuchungen ergaben das der EKF2D die besten und robustesten Ergebnisse für eine Lokalisierung erbringt, trotz, im Vergleich zu der UKF2D Variante, 17,3 % höhere Endpositionsabweichung aufweist. Die in diesem Projekt gewählte EKF3D Konfigurationsvariante eignet sich, wegen seinen starken Ungenauigkeiten in der Höhenbestimmung nicht für eine aussagekräftige Positionsbestimmung.
A smart energy concept was designed and implemented for a cluster of 5 existing multi-family houses, which combines heat pumps, photovoltaic (PV) modules and combined heat and power units (CHP) to achieve energy- and cost-efficient operation. Measurement results of the first year of operation show that the local power generation by PV modules and CHP unit has a positive effect on the electrical self-sufficiency by reducing electricity import from the grid. In winter, when the CHP unit operates continuously for long periods, the entire electricity for the heat pump and 91 % of the total electricity demand of the neighborhood are supplied locally. In summer, only 53 % is generated within the neighborhood. The use of a specifically developed energy management system EMS is intended to further increase this share. CO2 emissions for heating and electricity of the neighborhood are evaluated and amount to 18.4 kg/(m2a). Compared to the previous energy system consisting of gas boilers (29.1 kg/(m2a)), savings of 37 % are achieved with electricity consumption from the grid being reduced by 65 %. In the second construction stage, an additional heat pump, CHP unit and PV modules will be added. The measurement results indicate that the final district energy system is likely to achieve the ambitious CO2 reduction goal of -50% and further increase the self-sufficiency of the district.
The increasing diffusion of rapidly developing AI technologies led to the idea of the experiment to combine TRIZ-based automated idea generation with the natural language processing tool ChatGPT, using the chatbot to interpret the automatically generated elementary solution principles. The article explores the opportunities and benefits of a novel AI-enhanced approach to teaching systematic innovation, analyses the learning experience, identifies the factors that affect students' innovation and problem-solving performance, and highlights the main difficulties students face, especially in interdisciplinary problems.
Inner Congo
(2023)
This research-creation project, part of the DE\GLOBALIZE artistic research cycle presented at the #IFM2022 Conference, investigates the complexities of Congo violence, care, and colonialism. Drawing on Michel Serres' metaphor of the great estuaries, the study explores the topology of interactive documentaries, blending theory, emotion, and personal experiences. Accessible through the interactive web documentation at http://deglobalize.com, the platform offers a media-archaeological archive for speculative ethnography, enabling the forensic processing of single documents in line with actor-network theory.
The isolation measures adopted during the COVID-19 pandemic brought light to discussions related to the importance of meaningful social relationships as a basic need to human well-being. But even before the pandemic outbreak in the years 2020 and 2021, organizations and scholars were already drawing attention to the growing numbers related to lonely people in the world (World Economic Forum, 2019). Loneliness is an emotional distress caused by the lack of meaningful social connections, which affects people worldwide across all age groups, mainly young adults (Rook, 1984). The use of digital technologies has gained prominence as a means of alleviating the distress. As an example, studies have shown the benefits of using digital games both to stimulate social interactions (Steinfield, Ellison & Lampe, 2008) and to enhance the effects of digital interventions for mental health treatments, through gamification (Fleming et al., 2017). It is with these aspects in mind that the gamified app Noneliness was designed with the intention of reducing loneliness rates among young students at a German university. In addition to sharing the related works that supported the application development, this chapter also presents the aspects considered for the resource's design, its main functionalities, and the preliminary results related to the reduction of loneliness in the target audience.
We aim to debate and eventually be able to carefully judge how realistic the following statement of a young computer scientist is: “I would like to become an ethical correctly acting offensive cybersecurity expert”. The objective of this article is not to judge what is good and what is wrong behavior nor to present an overall solution to ethical dilemmas. Instead, the goal is to become aware of the various personal moral dilemmas a security expert may face during his work life. For this, a total of 14 cybersecurity students from HS Offenburg were asked to evaluate several case studies according to different ethical frameworks. The results and particularities are discussed, considering different ethical frameworks. We emphasize, that different ethical frameworks can lead to different preferred actions and that the moral understanding of the frameworks may differ even from student to student.
Sweaty has already participated several times in RoboCup soccer competitions (Adult Size). Now the work is focused on stabilizing the gait. Moreover, we would like to overcome the constraints of a ZMP-algorithm that has a horizontal footplate as precondition for the simplification of the equations. In addition we would like to switch between impedance and position control with a fuzzy-like algorithm that might help to minimize jerks when Sweaty’s feet touch the ground.
Generative machine learning models for creative purposes play an increasingly prominent role in the field of dance and technology. A particularly popular approach is the use of such models for generating synthetic motions. Such motions can either serve as source of ideation for choreographers or control an artificial dancer that acts as improvisation partner for human dancers. Several examples employ autoencoder-based deep-learning architectures that have been trained on motion capture recordings of human dancers. Synthetic motions are then generated by navigating the autoencoder's latent space. This paper proposes an alternative approach of using an autoencoder for creating synthetic motions. This approach controls the generation of synthetic motions on the level of the motion itself rather than its encoding. Two different methods are presented that follow this principle. Both methods are based on the interactive control of a single joint of an artificial dancer while the other joints remain under the control of the autoencoder. The first method combines the control of the orientation of a joint with iterative autoencoding. The second method combines the control of the target position of a joint with forward kinematics and the application of latent difference vectors. As illustrative example of an artistic application, this latter method is used for an artificial dancer that plays a digital instrument. The paper presents the implementation of these two methods and provides some preliminary results.
Over the last years, Convolutional Neural Networks (CNNs) have been the dominating neural architecture in a wide range of computer vision tasks. From an image and signal processing point of view, this success might be a bit surprising as the inherent spatial pyramid design of most CNNs is apparently violating basic signal processing laws, i.e. Sampling Theorem in their down-sampling operations. However, since poor sampling appeared not to affect model accuracy, this issue has been broadly neglected until model robustness started to receive more attention. Recent work in the context of adversarial attacks and distribution shifts, showed after all, that there is a strong correlation between the vulnerability of CNNs and aliasing artifacts induced by poor down-sampling operations. This paper builds on these findings and introduces an aliasing free down-sampling operation which can easily be plugged into any CNN architecture: FrequencyLowCut pooling. Our experiments show, that in combination with simple and Fast Gradient Sign Method (FGSM) adversarial training, our hyper-parameter free operator substantially improves model robustness and avoids catastrophic overfitting. Our code is available at https://github.com/GeJulia/flc_pooling
Despite the success of convolutional neural networks (CNNs) in many academic benchmarks for computer vision tasks, their application in the real-world is still facing fundamental challenges. One of these open problems is the inherent lack of robustness, unveiled by the striking effectiveness of adversarial attacks. Current attack methods are able to manipulate the network's prediction by adding specific but small amounts of noise to the input. In turn, adversarial training (AT) aims to achieve robustness against such attacks and ideally a better model generalization ability by including adversarial samples in the trainingset. However, an in-depth analysis of the resulting robust models beyond adversarial robustness is still pending. In this paper, we empirically analyze a variety of adversarially trained models that achieve high robust accuracies when facing state-of-the-art attacks and we show that AT has an interesting side-effect: it leads to models that are significantly less overconfident with their decisions, even on clean data than non-robust models. Further, our analysis of robust models shows that not only AT but also the model's building blocks (like activation functions and pooling) have a strong influence on the models' prediction confidences. Data & Project website: https://github.com/GeJulia/robustness_confidences_evaluation
Estimating the Robustness of Classification Models by the Structure of the Learned Feature-Space
(2022)
Over the last decade, the development of deep image classification networks has mostly been driven by the search for the best performance in terms of classification accuracy on standardized benchmarks like ImageNet. More recently, this focus has been expanded by the notion of model robustness, \ie the generalization abilities of models towards previously unseen changes in the data distribution. While new benchmarks, like ImageNet-C, have been introduced to measure robustness properties, we argue that fixed testsets are only able to capture a small portion of possible data variations and are thus limited and prone to generate new overfitted solutions. To overcome these drawbacks, we suggest to estimate the robustness of a model directly from the structure of its learned feature-space. We introduce robustness indicators which are obtained via unsupervised clustering of latent representations from a trained classifier and show very high correlations to the model performance on corrupted test data.
Many commonly well-performing convolutional neural network models have shown to be susceptible to input data perturbations, indicating a low model robustness. Adversarial attacks are thereby specifically optimized to reveal model weaknesses, by generating small, barely perceivable image perturbations that flip the model prediction. Robustness against attacks can be gained for example by using adversarial examples during training, which effectively reduces the measurable model attackability. In contrast, research on analyzing the source of a model’s vulnerability is scarce. In this paper, we analyze adversarially trained, robust models in the context of a specifically suspicious network operation, the downsampling layer, and provide evidence that robust models have learned to downsample more accurately and suffer significantly less from aliasing than baseline models.
Teaching and learning concepts that are adapted to the constantly evolving requirements due to rapid technological progress are essential for teaching in media photonics technology. After the development of a concept for research-oriented education in optics and photonics, the next step will be a conceptual restructuring and redesign of the entire curriculum for education in media photonics technology. By including typical research activities as essential components of the learning process, a broad platform for practical projects and applied research can be created, offering a variety of new development opportunities.
Recently, RobustBench (Croce et al. 2020) has become a widely recognized benchmark for the adversarial robustness of image
classification networks. In it’s most commonly reported sub-task, RobustBench evaluates and ranks the adversarial robustness of trained neural networks on CIFAR10 under AutoAttack (Croce and Hein 2020b) with l∞ perturbations limited to ϵ = 8/255. With leading scores of the currently best performing models of around 60% of the baseline, it is fair to characterize this benchmark to be quite challenging. Despite it’s general acceptance in recent literature, we aim to foster discussion about the suitability of RobustBench as a key indicator for robustness which could be generalized to practical applications. Our line of argumentation against this is two-fold and supported by excessive experiments presented in this paper: We argue that I) the alternation of data by AutoAttack with l∞, ϵ = 8/255 is unrealistically strong, resulting in close to perfect detection rates of adversarial samples even by simple detection algorithms and human observers.
We also show that other attack methods are much harder to detect while achieving similar success rates. II) That results on low resolution data sets like CIFAR10 do not generalize well to higher resolution images as gradient based attacks appear to become even more detectable with increasing resolutions.
Harnessing the overall benefits of the latest advancements in artificial intelligence (AI) requires the extensive collaboration of academia and industry. These collaborations promote innovation and growth while enforcing the practical usefulness of newer technologies in real life. The purpose of this article is to outline the challenges faced during cross-collaboration between academia and industry. These challenges are also inspected with the help of an ongoing project titled “Quality Assurance of Machine Learning Applications” (Q-AMeLiA), in which three universities cooperate with five industry partners to make the product risk of AI-based products visible. Further, we discuss the hurdles and the key challenges in machine learning (ML) technology transformation from academia to industry based on robustness, simplicity, and safety. These challenges are an outcome of the lack of common standards, metrics, and missing regulatory considerations when state-of-the-art (SOTA) technology is developed in academia. The use of biased datasets involves ethical concerns that might lead to unfair outcomes when the ML model is deployed in production. The advancement of AI in small and medium sized enterprises (SMEs) requires more in terms of common tandardization of concepts rather than algorithm breakthroughs. In this paper, in addition to the general challenges, we also discuss domain specific barriers for five different domains i.e., object detection, hardware benchmarking, continual learning, action recognition, and industrial process automation, and highlight the steps necessary for successfully managing the cross-sectoral collaborations between academia and industry.
Recent work has investigated the distributions of learned convolution filters through a large-scale study containing hundreds of heterogeneous image models. Surprisingly, on average, the distributions only show minor drifts in comparisons of various studied dimensions including the learned task, image domain, or dataset. However, among the studied image domains, medical imaging models appeared to show significant outliers through "spikey" distributions, and, therefore, learn clusters of highly specific filters different from other domains. Following this observation, we study the collected medical imaging models in more detail. We show that instead of fundamental differences, the outliers are due to specific processing in some architectures. Quite the contrary, for standardized architectures, we find that models trained on medical data do not significantly differ in their filter distributions from similar architectures trained on data from other domains. Our conclusions reinforce previous hypotheses stating that pre-training of imaging models can be done with any kind of diverse image data.
Despite the success of convolutional neural networks (CNNs) in many academic benchmarks for computer vision tasks, their application in the real-world is still facing fundamental challenges. One of these open problems is the inherent lack of robustness, unveiled by the striking effectiveness of adversarial attacks. Adversarial training (AT) is often considered as a remedy to train more robust networks. In this paper, we empirically analyze a variety of adversarially trained models that achieve high robust accuracies when facing state-of-the-art attacks and we show that AT has an interesting side-effect: it leads to models that are significantly less overconfident with their decisions even on clean data than non-robust models. Further, our analysis of robust models shows that not only AT but also the model's building blocks (like activation functions and pooling) have a strong influence on the models' prediction confidences.
In this paper, we propose a unified approach for network pruning and one-shot neural architecture search (NAS) via group sparsity. We first show that group sparsity via the recent Proximal Stochastic Gradient Descent (ProxSGD) algorithm achieves new state-of-the-art results for filter pruning. Then, we extend this approach to operation pruning, directly yielding a gradient-based NAS method based on group sparsity. Compared to existing gradient-based algorithms such as DARTS, the advantages of this new group sparsity approach are threefold. Firstly, instead of a costly bilevel optimization problem, we formulate the NAS problem as a single-level optimization problem, which can be optimally and efficiently solved using ProxSGD with convergence guarantees. Secondly, due to the operation-level sparsity, discretizing the network architecture by pruning less important operations can be safely done without any performance degradation. Thirdly, the proposed approach finds architectures that are both stable and well-performing on a variety of search spaces and datasets.
In the development of new vehicles, increasing customer comfort requirements and rising safety regulations often result in an increase in weight. Nevertheless, in order to be able to meet the demand for reduced fuel consumption, it is necessary within product development process to implement complex and filigree lightweight structures. This contribution therefore addresses the potential of generatively developed components for fiber-reinforced additive manufacturing (FRAM). Currently, several commercial systems for this application are available on the market. Therefore, a comparison of the systems is first made to determine a suitable system. Then, a highly stressed and safety-relevant chassis component of a race car is generatively designed and manufactured using FRAM. A matrix with short fiber reinforcement and additional long fiber reinforcement with carbon fibers is applied. Finally, tensile tests are carried out to check the mechanical properties. In addition, relevant properties such as weight and cost are obtained in order to be able to compare them with conventionally developed and manufactured components.
The integration of additive manufacturing processes into the teaching of students is an important prerequisite for the further dissemination of this new technology. In this context, the DfAM is of particular importance. For this reason, this paper presents an approach in which a connection is made between methodical product development and practical implementation by AM. Using a model racing car as an example, students independently develop significant improvements of particular assemblies. A final evaluation shows that the students have significantly improved their skills and competencies.
This paper presents a method for supporting the application of Additive Tooling (AT)-based validation environments in integrated product development. Based on a case study, relevant process steps, activities and possible barriers in the realisation of an injection-moulded product are identified and analysed. The aim of the method is to support the target-oriented application of Additive Tooling to obtain physical prototypes at an early stage and to shorten validation cycles.
Lithium-ion batteries show strongly nonlinear behaviour regarding the battery current and state of charge. Therefore, the modelling of lithium-ion batteries is complex. Combining physical and data-driven models in a grey-box model can simplify the modelling. Our focus is on using neural networks, especially neural ordinary differential equations, for grey-box modelling of lithium-ion batteries. A simple equivalent circuit model serves as a basis for the grey-box model. Unknown parameters and dependencies are then replaced by learnable parameters and neural networks. We use experimental full-cycle data and data from pulse tests of a lithium iron phosphate cell to train the model. Finally, we test the model against two dynamic load profiles: one consisting of half cycles and one dynamic load profile representing a home-storage system. The dynamic response of the battery is well captured by the model.
Robust scheduling problem is a major decision problem that is addressed in the literature, especially for remanufacturing systems; this problem is complex because of the high uncertainty and complex constraints involved. Generally, the existing approaches are dedicated to specific processes and do not enable the quick and efficient generation and evaluation of schedules. With the emergence of the Industry 4.0 paradigm, data availability is now considered an opportunity to facilitate the decision-making process. In this study, a data-driven decisionmaking process is proposed to treat the robust scheduling problem of remanufacturing systems in uncertain environments. In particular, this process generates simulation models based on a data-driven modeling approach. A robustness evaluation approach is proposed to answer several decision questions. An application of the decision process in an industrial case of a remanufacturing system is presented herein, illustrating the impact of robustness evaluation results on real-life decisions.
Physik durch Informatik
(2022)
Selbsttests in Lernmanagementsystemen (LMS) ermöglichen es Studierenden, den eigenen Lernfortschritt einzuschätzen. Das didaktische Konzept Physik durch Informatik (PDI) ist charakterisiert durch die Nutzung einer Programmiersprache zur Lösungseingabe bei Mathematik und Physik-Aufgaben. Im Gegensatz zur Lösungseingabe durch Zahlenwerte oder im Antwort-Auswahl-Verfahren erfordert die Implementierung einer Lösung in einer Programmiersprache eine höhere Kompetenzstufe.
An import ban of Russian energy sources to Germany is currently being increasingly discussed. We want to support the discussion by showing a way how the electricity system in Germany can manage low energy imports in the short term and which measures are necessary to still meet the climate protection targets. In this paper, we examine the impact of a complete stop of Russian fossil fuel imports on the electricity sector in Germany, and how this will affect the climate coals of an earlier coal phase-out and climate neutrality by 2045.
Following a scenario-based analysis, the results gave a point of view on how much would be needed to completely rely on the scarce non-renewable energy resources in Germany. Huge amounts of investments would be needed in order to ensure a secure supply of electricity, in both generation energy sources (RES) and energy storage systems (ESS). The key findings are that a rapid expansion of renewables and storage technologies will significantly reduce the dependence of the German electricity system on energy imports. The huge integration of renewable energy does not entail any significant imports of the energy sources natural gas, hard coal, and mineral oil, even in the long term. The results showed that a ban on fossil fuel imports from Russia outlines huge opportunities to go beyond the German government's climate targets, where the 1.5-degree-target is achieved in the electricity system.
Peer-to-peer energy trading and local electricity markets have been widely discussed as new options for the transformation of the energy system from the traditional centralized scheme to the novel decentralized one. Moreover, it has also been proposed as a more favourable alternative for already expiring feed in tariff policies that promote investment in renewable energy sources. Peer-to-peer energy trading is usually defined as the integration of several innovative technologies, that enable both prosumers and consumers to trade electricity, without intermediaries, at a consented price. Furthermore, the techno-economic aspects go hand in hand with the socio-economic aspects, which represent at the end significant barriers that need to be tackled to reach a higher impact on current power systems. Applying a qualitative analysis, two scalable peer-to-peer concepts are presented in this study and the possible participant´s entry probability into such concepts. Results show that consumers with a preference for environmental aspects have in general a higher willingness to participate in peer-to-peer energy trading. Moreover, battery storage systems are a key technology that could elevate the entry probability of prosumers into a peer-to-peer market.
In the railway technical centers, scheduling the maintenance activities is a very complex task, it consists in ordering, in the time, all the maintenance operations on the workstations, while respecting the number of resources, precedence constraints, and the workstations' availabilities. Currently, this process is not completely automatic. For improving this situation, this paper presents a mathematical model for the maintenance activities scheduling in the case of railway remanufacturing systems. The studied problem is modeled as a flexible job-shop, with the possibility for a job to be executed several times on a stage. MILP formulation is implemented with the Makespan as an objective, representing the time for remanufacturing the train. The aim is to create a generic model for optimizing the planning of the maintenance activities and improving the performance of the railway technical centers. At last, numerical results are presented, discussing the impact of the instances size on the computing time to solve the described problem.
To achieve Germany's climate targets, the industrial sector, among others, must be transformed. The decarbonization of industry through the electrification of heating processes is a promising option. In order to investigate this transformation in energy system models, high-resolution temporal demand profiles of the heat and electricity applications for different industries are required. This paper presents a method for generating synthetic electricity and heat load profiles for 14 industry types. Using this methodology, annual profiles with a 15-minute resolution can be generated for both energy demands. First, daily profiles for the electricity demand were generated for 4 different production days. These daily profiles are additionally subdivided into eight end-use application categories. Finally, white noise is applied to the profile of the mechanical drives. The heat profile is similar to the electrical but is subdivided into four temperature ranges and the two applications hot water and space heating. The space heating application is additionally adjusted to the average monthly outdoor temperature. Both time series were generated for the analysis of an electrification of industrial heat application in energy system modelling.
The energy system is changing since some years in order to achieve the climate goals from the Paris Agreement which wants to prevent an increase of the global temperature above 2 °C [1]. Decarbonisation of the energy system has become for governments a big challenge and different strategies are being stablished. Germany has set greenhouse gas reduction limits for different years and keeps track of the improvement made yearly. The expansion of renewable energy systems (RES) together with decarbonisation technologies are a key factor to accomplish this objective.
This research is done to analyse the effect of introducing biochar, a decarbonisation technology, and study how it will affect the energy system. Pyrolysis is the process from which biochar is obtained and it is modelled in an open-source energy system model. A sensibility analysis is done in order to assess the effect of changing the biomass potential and the costs for pyrolysis.
The role of pyrolysis is analysed in the form of different future scenarios for the year 2045 to evaluate the impact when the CO2 emission limit is zero. All scenarios are compared to the reference scenario, where pyrolysis is not considered.
Results show that biochar can be used to compensate the emissions from other conventional power plant and achieve an energy transition with lower costs. Furthermore, it was also found that pyrolysis can also reduce the need of flexibility. This study also shows that the biomass potential and the pyrolysis costs can strongly affect the behaviour of pyrolysis in the energy system.
The contribution of the RoofKIT student team to the SDE 21/22 competition is the extension of an existing café in Wuppertal, Germany, to create new functions and living space for the building with simultaneous energetic upgrading. A demonstration unit is built representing a small cut-out of this extension. The developed energy concept was thoroughly simulated by the student team in seminars using Modelica. The system uses mainly solar energy via PVT collectors as the heat source for a brine-water heat pump (space heating and hot water). Energy storage (thermal and electrical) is installed to decouple generation and consumption. Simulation results confirm that carbon neutrality is achieved for the building operation, consuming and generating around 60 kWh/m2a.
The purpose of this study was to describe the effects of running speed and slope on metatarsophalangeal (MTP) joint kinematics. 22 male and female runners underwent 3D motion analysis on an instrumented treadmill at three different speeds (2.5 m/s, 3.0 m/s, 3.5 m/s). At each speed, participants ran at seven slope conditions (downhill: -15%, -10%, -5%, level, and uphill: +5%, +10%, +15%). We found a significant main effect (p < 0.001) of running speed and slope on peak MTP dorsiflexion and a running speed by slope interaction effect (p < 0.001) for peak MTP dorsiflexion velocity. These findings highlight the need to consider running intensity and environmental factors like running surface inclination when considering MTP joint mechanics and technological aids to support runners.
Weitsprung mit und ohne Unterschenkelprothese – gleiche Sportart, unterschiedliche Disziplinen
(2022)
In recent years, the topic of embedded machine learning has become very popular in AI research. With the help of various compression techniques such as pruning, quantization and others compression techniques, it became possible to run neural networks on embedded devices. These techniques have opened up a whole new application area for machine learning. They range from smart products such as voice assistants to smart sensors that are needed in robotics. Despite the achievements in embedded machine learning, efficient algorithms for training neural networks in constrained domains are still lacking. Training on embedded devices will open up further fields of applications. Efficient training algorithms would enable federated learning on embedded devices, in which the data remains where it was collected, or retraining of neural networks in different domains. In this paper, we summarize techniques that make training on embedded devices possible. We first describe the need and requirements for such algorithms. Then we examine existing techniques that address training in resource-constrained environments as well as techniques that are also suitable for training on embedded devices, such as incremental learning. At the end, we also discuss which problems and open questions still need to be solved in these areas.
During the coronavirus crisis, labs had to be offered in digital form in mechanical engineering at short notice. For this purpose, digital twins of more complex test benches in the field of fluid energy machines were used in the mechanical engineering course, with which the students were able to interact remotely to obtain measurement data. The concept of the respective lab was revised with regard to its implementation as a remote laboratory. Fortunately, real-world labs were able to be fully replaced by remote labs. Student perceptions of remote labs were mostly positive. This paper explains the concept and design of the digital twins and the lab as well as the layout, procedure, and finally the results of the accompanying evaluation. However, the implementation of the digital twins to date does not yet include features that address the tactile experience of working in real-world labs.
The purpose of this study was to 1) compare knee joint kinematics and kinetics of fake-and-cut tasks of varying complexity in 51 female handball players and 2) present a case study of one athlete who ruptured her ACL three weeks post data collection. External knee joint moments and knee joint angles in all planes at the instance of the peak external knee abduction moment (KAM) as well as moment and angle time curves were analyzed. Peak KAMs and knee internal rotation moments were substantially higher than published values obtained during simple change-of-direction tasks and, along with flexion angles, differed significantly between the tasks. Introducing a ball reception and a static defender increased joint loads while they partially decreased again when anticipation was lacking. Our results suggest to use game-specific assessments of injury risk while complexity levels do not directly increase knee loading. Extreme values of several risk factors for a post-test injured athlete highlight the need and usefulness of appropriate screenings.
This study aimed to compare a simplified calculation of the knee abduction moment with the traditional inverse dynamics calculation when athletes perform fake-cut maneuvers with different complexities. In the simplified calculation, we multiply the force vector with its lever arm to the knee, projected onto the local coordinate system of the proximal thigh, hence neglecting the inertial contributions from distal segments. We found very strong ranking consistency using Spearman’s rank correlation coefficient when using the simplified method compared to the traditional calculation. Independent of the tasks, the simplified method resulted in higher moments than the inverse dynamics. This was caused by ignoring the moment caused by segment linear acceleration generating a counteracting moment by about 7%. An alternative to the complex calculations of inverse dynamics can be used to investigate the contributions of the GRF magnitude and its lever arm to the knee.
Effect of downhill running on biomechanical risk factors associated with iliotibial band syndrome
(2022)
The purpose of this study was to identify the influence of downhill running on biomechanical risk factors for iliotibial band syndrome. We conducted a 3D motion analysis of 22 females and males running on an instrumented treadmill at four different inclinations (0%, -5%, -10%, -15%) at a speed of 3.5 m/s. We found significant differences for biomechanical risk factors associated with iliotibial band syndrome. Peak knee flexion angle at initial ground contact (p < .001), peak knee adduction angle (p = .005), and iliotibial band strain (p < .001) systematically increased with increasing slope. Downhill running increases biomechanical risk factors for iliotibial band syndrome. Our results highlight the need to consider the individual running environment in assessing overuse injury risk in runners.
In this paper, we study the runtime performance of symmetric cryptographic algorithms on an embedded ARM Cortex-M4 platform. Symmetric cryptographic algorithms can serve to protect the integrity and optionally, if supported by the algorithm, the confidentiality of data. A broad range of well-established algorithms exists, where the different algorithms typically have different properties and come with different computational complexity. On deeply embedded systems, the overhead imposed by cryptographic operations may be significant. We execute the algorithms AES-GCM, ChaCha20-Poly1305, HMAC-SHA256, KMAC, and SipHash on an STM32 embedded microcontroller and benchmark the execution times of the algorithms as a function of the input lengths.
Spatially Distributed Wireless Networks (SDWN) are one of the basic technologies for the Internet of Things (IoT) and (Industrial) Internet of Things (IIoT) applications. These SDWN for many of these applications has strict requirements such as low cost, simple installation and operations, and high potential flexibility and mobility. Among the different Narrowband Wireless Wide Area Networking (NBWWAN) technologies, which are introduced to address these categories of wireless networking requirements, Narrowband Internet of Things (NB-IoT) is getting more traction due to attractive system parameters, energy-saving mode of operation with low data rates and bandwidth, and its applicability in 5G use cases. Since several technologies are available and because the underlying use cases come with various requirements, it is essential to perform a systematic comparative analysis of competing technologies to choose the right technology. It is also important to perform testing during different phases of the system development life cycle. This paper describes the systematic test environment for automated testing of radio communication and systematic measurements of the performance of NB-IoT.
Objective: Dickkopf 3 (DKK3) has been identified as a urinary biomarker. Values above 4000 pg/mg creatinine (Cr) were linked with a higher risk of short-term decline of kidney function (J Am Soc Nephrol 29: 2722–2733). However, as of today, there is little experience with DKK3 as a risk marker in everyday clinical practice. We used algorithm-based data analysis to evaluate the potential dependence of DKK3 in a cohort from a large single center in Germany.
Method: DKK3 was measured in all CKD patients in our center October 1 st 2018 till Dec. 31 2019, together with calculated GFR (eGFR) and urinary albumin/creatinine ratio (UACR). Kidney transplant patients were excluded. Until the end of follow-up Dec 31 st 2021, repeated measurements were performed for all parameters. Data analysis was performed using MD-Explorer (BioArtProducts, Rostock, Germany) and Python with multiple libraries. Linear regression models were applied in patients for DKK3, eGFR and UACR. Comparison of the models was performed with a twosided Kolmogorov-Smirnov test.
Results: 1206 DKK3 measurements were performed in 1103 patients (621 male, age 70yrs, eGFR 29,41 ml/min/1.73qm, UACR 800 mg/g). 134 patients died during follow-up. DKK3 mean was 2905 pg/mg Cr (max. 20000, 75 % percentile 3800). 121 pts had DKK3 > 4000. At the end of follow-up 7 % of patients with DKK3 < 4000 (initial eGFR 17.6) versus 39.6 % of patients with DDK3 > 4000 (initial eGFR 15.7) underwent dialysis. Compared to eGFR and UACR at baseline, DKK3 > 4000 performed best to predict eGFR loss over the next 12 months.
Conclusion: In this cohort of CKD patients, DKK3 > 4000 at baseline predicted the eGFR slope better than eGFR or UACR at baseline. DKK3 > 4000 reflected a higher risk of progression towards ESRD in patients with similar baseline eGFR levels.
Projektmanagement entwickelt sich kontinuierlich, auch in qualitativen Sprüngen und Zyklen. Planungsiterationen aus der Agilität und die coronabedingte Digitalisierung der Kommunikation sind nicht die einzigen aktuellen Entwicklungen. Nicht einmal die Wichtigsten. Es wird ein Überblick vermittelt, der nicht nur verstehen, sondern gestalten hilft.
The majority of anterior cruciate ligament (ACL) injuries in team sports are non-contact injuries, with cutting maneuvers identified as high-risk tasks. Young female handball players have been shown to be at greater risk for ACL injuries than males. One risk factor for ACL injuries is the magnitude of the knee abduction moment (KAM). Cutting technique variables on foot placement, overall approach and knee kinematics have been shown to influence the KAM. Since injury risk is believed to increase with increasing task complexity, the purpose of the study was to test the effect of task complexity on technique variables that influence the KAM in female handball players during fake-and-cut tasks.
Electrode modelling and simulation of diagnostic and pulmonary vein isolation in atrial fibrillation
(2022)
DE\GLOBALIZE
(2022)
The artistic research cycle DE\GLOBALIZE is a media ecological search movement for the terrestrial. After examining matters of fact in India (2014-18), matters of concern in Egypt (2016-2019) and matters of care in the Upper Rhine (2018-22), the focus turns toward matters of violence in the Congo (2022). From matter to mater, mother-earth, the garden to exploitation. From science, water and climate to migration, oppression and extermination.
The long-term research is accessible through interactive web documentation. The platform serves as a continuous media-archaeological archive for a speculative ethnography. The relational structure of the videographic essay is enabling the forensic processing of single documents in the sense of the actor-network theory.
The subject of the presentation at IFM is a field trip to the Congo planned for March 2022, which will focus on the ambivalence of violence and care in collaboration with local artists. The field trip is based on the postcolonial reflection luderitzcargo by the author from 1996, in which a freight container was transformed into a translocal cinema in Namibia.
Through the journey to Congo, a group of media artists, a psychotherapist, a theater dramaturg, a filmmaker and a philosopher intend to explore the political, technological and psycho-geographic borders. By artistic interventions with locals, we want to interfere with relational string figures as part of the new Earth Politics. They are focusing on the displaced consumption of resources which are hard-fought and guarantee prosperity in the global north. The so-called ghost acreages are repressed and justified as part of a civilizational mission. With this trip, we want to confront our self-lies with the ones of our hosts. We want to confront ourselves with the foreign, the dark and the displaced ghosts within ourselves. In the presentation at the #IFM2022 Conference, the platform DE\GLOBALIZE will be problematized itself as an example of epistemic violence for the ethnographic memory of (Western) knowledge.
We are not the missionaries but the perplexed travellers. In our search movement, we are dealing with psychoanalysis, video, performance and trance. As disoriented white men we try the reversal of Black Skin and White Mask by Franz Fanon without blackfacing. We will not only care about the sensitivity of our skin but that of our g/hosts and the one of mother earth.
VR-based implementation of interactive laboratory experiments in optics and photonics education
(2022)
Within the framework of a developed blended learning concept, a lot of experience has already been gained with a mixture of theoretical lectures and hands-on activities, combined with the advantages of modern digital media. Here, visualizations using videos, animations and augmented reality have proven to be effective tools to convey learning content in a sustainable way. In the next step, ideas and concepts were developed to implement hands-on laboratory experiments in a virtual environment. The main focus is on the realization of virtual experiments and environments that give the students a deep insight into selected subfields of optics and photonics.
The sharp rise in electricity and oil prices due to the war in Ukraine has caused fluctuations in the results of the previous study about the economic analysis of electric buses. This paper shows how the increase in fuel prices affects the implementation of electric buses. This publication is constructing the Total Cost of Ownership (TCO) model in the small-mid-size city, Offenburg for the transition to electric buses. The future development of costs is estimated and a projection based on learning curves will be carried out. This study intends to introduce a new future prospect by presenting the latest data based on previous research. Through the new TCO result, the cost differences between the existing diesel bus and the electric bus are updated, and also the future prospects for the economic feasibility of the electric bus in a small and midsize city are presented.
Fallstudien sollen theoretische Lerninhalte zu Konzepten von Business Intelligence und Data Warehousing veranschaulichen und in einen praxisnahen Kontext bringen. Außerdem sollen Studierende umsetzungsorientierte Kompetenzen mit praxisrelevanten Systemen erwerben. Um diese Kompetenzen abzuprüfen und um die Auseinandersetzung mit Software und Konzepten zu vertiefen, haben sich Projekte als Ergänzung zu Fallstudien und Klausuren vielfach bewährt. Der Vortrag stellt dar, welche Möglichkeiten Dozierende im Rahmen der vom UCC zur Verfügung gestellten Plattform SAP Data Warehouse Cloud (SAP DWC) haben, um studentische Projekte zu Data Warehousing und Analytics durchzuführen. Der Autor berichtet über seine Erfahrung aus der Betreuung von über 30 Projekten mit SAP DWC aus verschiedenen Studiengängen seit 2020. Neben einer Übersicht über die von Studierenden gewählten Themen werden ausgewählte Projektergebnisse vorgestellt. Außerdem wird auf den Modus der Durchführung sowie existierende systemseitige Limitationen eingegangen. Für Dozierende, die mit ihren Studierenden eigene Projekte erfolgreich durchführen möchten, werden konkrete Hinweise und Maßnahmen dargestellt.
The importance of machine learning has been increasing dramatically for years. From assistance systems to production optimisation to support the health sector, almost every area of daily life and industry comes into contact with machine learning. Besides all the benefits that ML brings, the lack of transparency and the difficulty in creating traceability pose major risks. While there are solutions that make the training of machine learning models more transparent, traceability is still a major challenge. Ensuring the identity of a model is another challenge. Unnoticed modification of a model is also a danger when using ML. One solution is to create an ML birth certificate and an ML family tree secured by blockchain technology. Important information about training and changes to the model through retraining can be stored in a blockchain and accessed by any user to create more security and traceability about an ML model.
Duplikaterkennung, -suche und -konsolidierung für Kunden- und Geschäftspartnerdaten, sog. „Identity Resolution“, ist die Voraussetzung für erfolgreiches Customer Relationship Management und Customer Experience Management, aber auch für das Risikomanagement zur Minimierung von Betrugsrisiken und Einhaltung regulatorischer Vorschriften und viele weitere Anwendungsfälle. Diese Systeme sind jedoch hochkomplex und müssen individuell an die kundenspezifischen Anforderungen angepasst werden. Der Einsatz lernbasierter Verfahren bietet großes Potenzial zur automatisierten Anpassung. In diesem Beitrag präsentieren wir für ein KMU praxisfähige, lernbasierte Verfahren zur automatischen Konfiguration von Business-Regeln in Duplikaterkennungssystemen. Dabei wurden für Fachanwender Möglichkeiten entwickelt, um beispielgetrieben das Match-System an individuelle Business-Regeln (u.a. Umzugserkennung, Sperrlistenabgleich) anzupassen und zu konfigurieren. Die entwickelten Verfahren wurden evaluiert und in einer prototypischen Lösung integriert. Wir konnten zeigen, dass unser Machine-Learning-Verfahren, die von einem Domainexperten erstellten Business-Regeln für das Duplikaterkennungssystem „identity“ verbessern konnte. Zudem konnte der hierzu erforderliche Zeitaufwand verkürzt werden.
Elektronische Türschilder zur Darstellung von Informationen sind insbesondere in öffentlichen Gebäuden zwischenzeitlich weit verbreitet. Die Varianz dieser elektronischen Türschilder reicht vom Tablet-basierten Türschild bis hin zum PC-basierten Türschild mit externem Bildschirm. Zumeist werden die Systeme mit 230 V betrieben. Bei einer großen Summe von Türschildern in öffentlichen Gebäuden kann dies zu einem signifikanten Umsatz an Energie führen. Im Rahmen dieses Papers wird die Entwicklung eines energieautarken arbeiten Türschildes vorgestellt, bei dem ein E-Paper-Display zum Einsatz kommt. Das Türschild lässt sich per Smartphone-App und NFC-Schnittstelle konfigurieren. Es wird insbesondere auf das Low-Power-Hardware-Design der Elektronik und energetische Aspekte eingegangen.