Refine
Year of publication
- 2023 (487) (remove)
Document Type
- Conference Proceeding (115)
- Bachelor Thesis (104)
- Article (reviewed) (84)
- Master's Thesis (64)
- Article (unreviewed) (38)
- Part of a Book (24)
- Other (13)
- Patent (13)
- Book (9)
- Doctoral Thesis (6)
Conference Type
- Konferenzartikel (88)
- Konferenz-Abstract (20)
- Konferenz-Poster (2)
- Konferenzband (2)
- Sonstiges (2)
Keywords
- Biomechanik (19)
- Künstliche Intelligenz (12)
- IT-Sicherheit (11)
- Deep Leaning (10)
- Marketing (10)
- Wärmepumpe (8)
- Medizintechnik (7)
- Social Media (7)
- Deep Learning (6)
- Machine Learning (6)
Institute
- Fakultät Maschinenbau und Verfahrenstechnik (M+V) (151)
- Fakultät Medien (M) (ab 22.04.2021) (136)
- Fakultät Elektrotechnik, Medizintechnik und Informatik (EMI) (ab 04/2019) (132)
- Fakultät Wirtschaft (W) (68)
- INES - Institut für nachhaltige Energiesysteme (38)
- IMLA - Institute for Machine Learning and Analytics (27)
- ivESK - Institut für verlässliche Embedded Systems und Kommunikationselektronik (23)
- IBMS - Institute for Advanced Biomechanics and Motion Studies (ab 16.11.2022) (21)
- POIM - Peter Osypka Institute of Medical Engineering (12)
- IfTI - Institute for Trade and Innovation (9)
- IUAS - Institute for Unmanned Aerial Systems (4)
- ACI - Affective and Cognitive Institute (3)
- Zentrale Einrichtungen (3)
- CRT - Campus Research & Transfer (2)
- IDEeP - Institute for Digital Engineering and Production (ab 12.10.2022) (2)
- WLRI - Work-Life Robotics Institute (2)
Open Access
- Closed (301)
- Open Access (168)
- Bronze (59)
- Diamond (47)
- Gold (37)
- Hybrid (20)
- Closed Access (18)
- Grün (5)
The interest of scientists to study motion sequences exists in the fields of sports science, clinical analysis and computer animation for quite some time. While in the last decades mainly markerbased motion capture systems have been used to evaluate movements, the interest in markerless systems is growing more and more. Nevertheless, in the field of clinical analysis, markerless methods have not yet proven their value, partly due to a lack of studies evaluating the quality of the obtained data. Therefore, this study aims to validate two markerless motion capture softwares from Simi Reality Motion Systems. The software Simi Shape, which is a mixture of traditional image-based tracking supported by an artificial intelligence net (AI net), and the software Crush, that uses a completely AI-based method. For this purpose, all motion data was recorded with two in-house motion capture systems. One system for recording the movements for a markerbased evaluation as gold standard and one system for markerless tracking. Within a laboratory environment, eight cameras per system were mounted around the area of motion. By placing two cameras in the same position and using the same calibration, deviations in the image data between those for markerbased and markerless tracking were extremely minimal. Based on this data, markerbased tracking was performed using the Simi Motion program, markerless tracking was performed using the Simi Shape software system and the latest software from Simi Reality Motion Systems, Crush. When comparing the markerless data with the markerbased data, an average root mean square error of 0,038 m was calculated for Simi Shape and a deviation of 0,037 m for Crush. In a direct comparison of the two markerless systems, a root mean square error of 0,019 m was scored. Based on these data, conclusions could be drawn about the accuracies of the two markerless systems. The obtained kinematic data of the tracking are in the range of high accuracy, which is limited to a deviation of less than 0,05 m according to the literature.
Proteine spielen eine entscheidende Rolle im Stoffwechsel aller Lebewesen. Für die Lebensmittelindustrie sind die Grundbausteine der Proteine, besonders interessant, da sie den Geschmack von Lebensmitteln stark beeinflussen können, oder sich in Form von Supplementen als Nahrungsergänzungsmittel für Sportler eignen. Zur Gewinnung dieser Aminosäuren werden unteranderem Aminopeptidasen verwendet. In dieser Arbeit wurden fünf fungale Aminopeptidasen Pep 1 aus Aspergillus turcosus, Pep 4 aus Aspergillus thermomutatus, Pep 6 aus Coccidioides posadasii C735 delta SOWgp, Pep 7 aus Trichophyton interdigitale H6 und Pep 8 aus Onygena corvina in Pichia pastoris rekombinant exprimiert, mittels Ultrafiltration aufkonzentriert und biochemisch charakterisiert. Die Aminopeptidasen aus Aspergillus turcosus (Pep 1), Aspergillus thermomutatus (Pep 4) und Trichophyton interdigitale H6 (Pep 7) hatten ihr pH-Optimum bei pH 9, während die Aminopeptidasen aus Coccidioides posadasii C735 delta SOWgp (Pep 6) und Onygena corvina (Pep 8) bei pH 8,5 die höchste Aktivität aufwiesen. Die Temperaturoptima lagen bei 60 °C (Pep 1 und Pep 6) und 50 °C (Pep 4, Pep 7 und Pep 8) und alle waren zwischen 45 und 55 °C für 15 Minuten temperaturstabil. Die Aminopeptidasen hatten nach der Deglykosilierung ein Molekulargewicht von 57 (Pep 1), 52 (Pep 4 und Pep 7), 54 (Pep 6) und 53 kDA (Pep 8). Mit Leu-pNA als Substrat zeigten die fungalen Aminopeptidasen maximale spezifische Umsatzgeschwindigkeiten von 357 (Pep 1), 238 (Pep 4), 2000 (Pep 6), 56 (Pep 7) und 476 nkat/g (Pep 8). Pep 1 wurde am stärksten positiv von Bivalenten Ionen wie Mg2+, Co2+, Mn2+ und Ca2+ beeinflusst, in Gegenwart von 0,1 mM Mg2+ konnte eine relative Enzymaktivität von 345 % im Vergleich zum Referenzwert gemessen werden. Bei den anderen Aminopeptidasen konnten inhibitorische Effekte festgestellt werden. Bie allen fungalen Aminopeptidasen bei der Hydrolyse von vorhydrolysiertem Gluten die Freisetzung von Aminogruppen festgestellt werden, somit eignen sich die Aminopeptidasen in Kombination mit Endopeptidasen zur Hydrolyse von Proteingemischen. Bei der Hydrolyse konnte zudem die Freisetzung von Glutamin durch das L-Glutamine/Ammonia Assay Kit (Rapid) von Megazyme nachgewiesen werden.
Die Bachelorarbeit wurde im Rahmen einer Äquivalenz mit der Ingenieurschule INSA Straßburg geschrieben. Zum Thema wurde eine Projektarbeit im Auftrag der Kriminalpolizei mit einer Bearbeitungszeit von drei Semestern. Aufgabenstellung war es, einen Überblick eines Tatorts ohne menschlichen Eingriff zu verschaffen. Noch heute müssen die Polizisten den Tatort betreten, um Beweisspuren zu sammeln. Das Problem hierbei ist, dass dadurch Hinweise fast systematisch verloren gehen.
Mit einem ferngesteuerten Luftschiff sollte diese Aufgabenstellung gemeistert werden. Die Arbeit beschäftigt sich spezifisch mit dem Entwurf eines integrierten Kamerasystems und mit der Implementierung der Elektronik im Gesamtsystem für einen ersten Prototypen. Auf die Konzeptionierung der Fortbewegung und des Flugobjekts wird somit nicht in Detail eingegangen. Diese Aspekte werden bei der Erstellung des Lastenhefts geschildert und bei Berührungspunkten mit dem im Fokus stehenden Themengebiet erklärt.
Auf struktureller Ebene wird wie gefolgt vorgegangen: Die technischen Lösungen werden abgewogen, getestet und progressiv zusammengeführt. Zum Ende der Bearbeitungszeit können ferngesteuert Bilder und Panoramen aufgenommen und der Tatort live mit einem Video-Feedback observiert werden. Die Elektronik wurde, nach Erarbeitung der elektronischen Architektur, mithilfe einer gedruckten Leiterplatte kompakt in der Gondel des Luftschiffs integriert. Mit zahlreichen Optimierungsperspektiven wurde der erste Prototyp fertiggestellt und das Ziel dieser Arbeit erreicht.
Diese Arbeit befasst sich mit dem Entwurf und der Herstellung einer Roboterhandprothese, die amputierten Menschen eine gewisse Mobilität und eine teilweise Nutzung der Hand ermöglichen soll.
Das Projekt konzentriert sich insbesondere auf die Erkennung der vom Benutzer ausgeführten Bewegung und wird die Schritte der Erfassung, der Bewegung der Übertragung und die Erkennung detailliert darstellen.
Im Rahmen des Projekts "Myo-Hand" soll ein Prototyp einer myoelektrischen Handprothese entwickelt werden. Diese Prothese soll für handamputierte Menschen nützlich sein, da sie ihren Komfort und ihre Unabhängigkeit im Alltag verbessert und somit zu ihrer Eingliederung in die Gesellschaft beiträgt.
In dieser Abschlussarbeit werden die gesamte Arbeit, Studien und Realisierungen vorgestellt, die in den letzten zwei Semestern (WS2022 und SS2023) der Mechanik der Prothese durchgeführt wurden.
Nach einer kurzen Vorstellung der Arbeit des letzten Jahres wird ein Überblick über die Idee und die allgemeine Funktionsweise der Prothese gegeben, die verschiedene Komponenten umfasst: Elektronik, Machine-Learning, Steuerung, Wiederverwertbarkeit und Carbon Footprint. Das Lastenheft wird ebenfalls vorgestellt.
Die technologischen Entscheidungen werden näher erläutert: das Design der Teile, die Wahl der Verbindungen und die verwendeten Motoren. In einem zweiten Teil werden die Schritte zur Herstellung des ersten und zweiten Prototyps sowie eine Kritik des Ergebnisses vorgestellt. Weiter werden die Verbesserungen und die Ziele bis Januar 2024 besprochen.
Anschließend werden der Lebenszyklus der Prothese und ihr Carbon Footprint analysiert und mithilfe der Software Ecodesign Studio berechnet.
Diese Bachelorarbeit hat sich ausführlich mit der Erstellung von Sicherheitskonzepten und der damit einhergehenden Erstellung einer Risikobeurteilung beschäftigt.
Dabei wurde festgestellt, dass eine besonders große Gefahr des Werkzeugwechselsystems durch die Gefährdung des herabfallenden Werkzeuges entsteht. Zum einen kann dies durch das Aufkommen eines Energie- bzw. Druckverlust entstehen. Dieses Risko wird durch den Einsatz einer bewährten Feder und einer selbsthemmenden Keilverbindung zwischen Verriegelungskolben und -bolzen vermindert.
Zum anderen kann durch Steuerungs- oder Programmierfehler das Werkzeugwechselsystem an nicht vorgesehener Stelle geöffnet werden und sich Fest- und Losteil trennen. Die dadurch entstehende Gefahr eines Herabfallenden Werkzeuges wird von einer steuerungsabhängigen Schutzmaßnahme verhindert, welche über eine genaue Positionsabfrage vor dem gefahrbringenden Szenario schützen sollen.
Hierbei sind verschiedene Konzepte ausgearbeitet worden, welche das von der Risikobeurteilung geforderte erforderliche PLr d erfüllen. Für die Konzepte wurden verschiedene berührungslose Sensoren und Sicherheitsschalter ausgesucht und ein Ventilschaltplan für die Erfüllung der Sicherheitsfunktion erstellt.
Auch andere Gefährdungen werden mithilfe der Anwendung von Normen, den Einsatz von inhärent sicheren Konstruktionen, technische/ergänzende Schutzmaßnahme oder Benutzerinformationen ausreichend vermindert.
Mithilfe einer weiteren Abschlussarbeit werden diese Sicherheitskonzepte und die damit einhergehende Risikobeurteilung überprüft und validiert. Hierfür wird die Norm DIN EN ISO 13849-2 verwendet.
In 2015, Google engineer Alexander Mordvintsev presented DeepDream as technique to visualise the feature analysis capabilities of deep neural networks that have been trained on image classification tasks. For a brief moment, this technique enjoyed some popularity among scientists, artists, and the general public because of its capability to create seemingly hallucinatory synthetic images. But soon after, research moved on to generative models capable of producing more diverse and more realistic synthetic images. At the same time, the means of interaction with these models have shifted away from a direct manipulation of algorithmic properties towards a predominance of high level controls that obscure the model's internal working. In this paper, we present research that returns to DeepDream to assess its suit-ability as method for sound synthesis. We consider this research to be necessary for two reasons: it tackles a perceived lack of research on musical applications of DeepDream, and it addresses DeepDream's potential to combine data driven and algorithmic approaches. Our research includes a study of how the model architecture, choice of audio data-sets, and method of audio processing influence the acoustic characteristics of the synthesised sounds. We also look into the potential application of DeepDream in a live-performance setting. For this reason, the study limits itself to models consisting of small neural networks that process time-domain representations of audio. These models are resource-friendly enough to operate in real time. We hope that the results obtained so far highlight the attractiveness of Deep-Dream for musical approaches that combine algorithmic investigation with curiosity driven and open ended exploration.
The progress in machine learning has led to advanced deep neural networks. These networks are widely used in computer vision tasks and safety-critical applications. The automotive industry, in particular, has experienced a significant transformation with the integration of deep learning techniques and neural networks. This integration contributes to the realization of autonomous driving systems. Object detection is a crucial element in autonomous driving. It contributes to vehicular safety and operational efficiency. This technology allows vehicles to perceive and identify their surroundings. It detects objects like pedestrians, vehicles, road signs, and obstacles. Object detection has evolved from being a conceptual necessity to an integral part of advanced driver assistance systems (ADAS) and the foundation of autonomous driving technologies. These advancements enable vehicles to make real-time decisions based on their understanding of the environment, improving safety and driving experiences. However, the increasing reliance on deep neural networks for object detection and autonomous driving has brought attention to potential vulnerabilities within these systems. Recent research has highlighted the susceptibility of these systems to adversarial attacks. Adversarial attacks are well-designed inputs that exploit weaknesses in the deep learning models underlying object detection. Successful attacks can cause misclassifications and critical errors, posing a significant threat to the functionality and safety of autonomous vehicles. With the rapid development of object detection systems, the vulnerability to adversarial attacks has become a major concern. These attacks manipulate inputs to deceive the target system, significantly compromising the reliability and safety of autonomous vehicles. In this study, we focus on analyzing adversarial attacks on state-of-the-art object detection models. We create adversarial examples to test the models’ robustness. We also check if the attacks work on a different object detection model meant for similar tasks. Additionally, we extensively evaluate recent defense mechanisms to see how effective they are in protecting deep neural networks (DNNs) from adversarial attacks and provide a comprehensive overview of the most commonly used defense strategies against adversarial attacks, highlighting how they can be implemented practically in real-world situations.
In Deutschland wird zur Bewertung von KWK-Anlagen die Stromgutschriftmethode angewendet. Dies ist die Standardmethode dieses Arbeitsblattes und der DIN V 18599-1. Die vorliegende Fassung der FW 309-1 ist wie seine früheren Fassungen als Anwendungs- und Auslegungshilfe zu den geltenden Normen und Gesetzen konzipiert, da diese Dokumente Regelungslücken aufweisen. Das Bilanzierungsprinzip, die meisten Primärenergiefaktoren und die KWK-Bewertungsmethode wurden beibehalten.
Änderung gegenüber der Fassung 2021:
Anhang A, Tabelle A.3: In der Neufassung des Gebäudeenergiegesetzes ab dem 1.1.2023 wird ein neuer Primärenergiefaktor für netzbezogenen Strom von 1,2 für Großwärmepumpen festgelegt. Diese Festlegung wird in dieses Dokument übernommen.
Ultra-low-power passive telemetry systems for industrial and biomedical applications have gained much popularity lately. The reduction of the power consumption and size of the circuits poses critical challenges in ultra-low-power circuit design. Biotelemetry applications like leakage detection in silicone breast implants require low-power-consuming small-size electronics. In this doctoral thesis, the design, simulation, and measurement of a programmable mixed-signal System-on-Chip (SoC) called General Application Passive Sensor Integrated Circuit (GAPSIC) is presented. Owing to the low power consumption, GAPSIC is capable of completely passive operation. Such a batteryless passive system has lower maintenance complexity and is also free from battery-related health hazards. With a die area of 4.92 mm² and a maximum analog power consumption of 592 µW, GAPSIC has one of the best figure-of-merits compared to similar state-of-the-art SoCs. Regarding possible applications, GAPSIC can read out and digitally transmit the signals of resistive sensors for pressure or temperature measurements. Additionally, GAPSIC can measure electrocardiogram (ECG) signals and conductivity.
The design of GAPSIC complies with the International Organization for Standardization (ISO) 15693/NFC (near field communication) 5 standard for radio frequency identification (RFID), corresponding to the frequency range of 13.56 MHz. A passive transponder developed with GAPSIC comprises of an external memory storage and very few other external components, like an antenna and sensors. The passive tag antenna and reader antenna use inductive coupling for communication and energy transfer, which enables passive operation. A passive tag developed with GAPSIC can communicate with an NFC compatible smart device or an ISO 15693 RFID reader. An external memory storage contains the programmable application-specific firmware.
As a mixed-signal SoC, GAPSIC includes both analog and digital circuitries. The analog block of GAPSIC includes a power management unit, an RFID/NFC communication unit, and a sensor readout unit. The digital block includes an integrated 32-bit microcontroller, developed by the Hochschule Offenburg ASIC design center, and digital peripherals. A 16-kilobyte random-access memory and a read-only 16-kilobyte memory constitute the GAPSIC internal memory. For the fabrication of GAPSIC, one poly, six-metal 0.18 µm CMOS process is used.
The design of GAPSIC includes two stages. In the first stage, a standalone RFID/NFC frontend chip with a power management unit, an RFID/NFC communication unit, a clock regenerator unit, and a field detector unit was designed. In the second stage, the rest of the functional blocks were integrated with the blocks of the RFID/NFC frontend chip for the final integration of GAPSIC. To reduce the power consumption, conventional low-power design techniques were applied extensively like multiple power supplies, and the operation of complementary metal-oxide-semiconductor (CMOS) transistors in the sub-threshold region of operation, as well as further innovative circuit designs.
An overvoltage protection circuit, a power rectifier, a bandgap reference circuit, and two low-dropout (LDO) voltage regulators constitute the power management unit of GAPSIC. The overvoltage protection circuit uses a novel method where three stacked transistor pairs shunt the extra voltage. In the power rectifier, four rectifier units are arranged in parallel, which is a unique approach. The four parallel rectifier units provide the optimal choice in terms of voltage drop and the area required.
The communication unit is responsible for RFID/NFC communication and incorporates demodulation and load modulation circuitry. The demodulator circuit comprises of an envelope detector, a high-pass filter, and a comparator. Following a new approach, the bandgap reference circuit itself acts as the load for the envelope detector circuit, which minimizes the circuit complexity and area. For the communication between the reader and the RFID/NFC tag, amplitude-shift keying (ASK) is used to modulate signals, where the smallest modulation index can be as low as 10%. A novel technique involving a comparator with a preset offset voltage effectively demodulates the ASK signal. With an effective die area of 0.7 mm² and power consumption of 107 µW, the standalone RFID/NFC frontend chip has the best figure-of-merits compared to the state-of-the-art frontend chips reported in the relevant literature. A passive RFID/NFC tag developed with the standalone frontend chip, as well as temperature and pressure sensors demonstrate the full passive operational capability of the frontend chip. An NFC reader device using a custom-built Android-based application software reads out the sensor data from the passive tag.
The sensor readout circuit consists of a channel selector with two differential and four single-ended inputs with a programmable-gain instrumentation amplifier. The entire sensor readout part remains deactivated when not in use. The internal memory stores the measured offset voltage of the instrumentation amplifier, where a firmware code removes the offset voltage from the measured sensor signal. A 12-bit successive approximation register (SAR) type analog-to-digital-converter (ADC) based on a charge redistribution architecture converts the measured sensor data to a digital value. The digital peripherals include a serial peripheral interface, four timers, RFID/NFC interfaces, sensor readout unit interfaces, and 12-bit SAR logic.
Two sets of studies with custom-made NFC tag antennas for biomedical applications were conducted to ascertain their compatibility with GAPSIC. The first study involved the link efficiency measurements of NFC tag antennas and an NFC reader antenna with porcine tissue. In a separate experiment, the effect of a ferrite compared to air core on the antenna-coupling factor was investigated. With the ferrite core, the coupling factor increased by four times.
Among the state-of-the-art SoCs published in recent scientific articles, GAPSIC is the only passive programmable SoC with a power management unit, an RFID/NFC communication interface, a sensor readout circuit, a 12-bit SAR ADC, and an integrated 32-bit microcontroller. This doctoral research includes the preliminary study of three passive RFID tags designed with discrete components for biomedical and industrial applications like measurements of temperature, pH, conductivity, and oxygen concentration, along with leakage detection in silicone breast implants. Besides its small size and low power consumption, GAPSIC is suitable for each of the biomedical and industrial applications mentioned above due to the integrated high-performance microcontroller, the robust programmable instrumentation amplifier, and the 12-bit analog-to-digital converter. Furthermore, the simulation and measurement data show that GAPSIC is well suited for the design of a passive tag to monitor arterial blood pressure in patients experiencing Peripheral Artery Disease (PAD), which is proposed in this doctoral thesis as an exemplary application of the developed system.
This study investigates the impact of global payroll outsourcing on organizational efficiency and cost reduction based on the analysis of diverse implications stemming from thirty one (31) survey results. The findings reveal multifaceted challenges and benefitsassociated with outsourcing global payroll processing.
The research also unveils the most benefits of global payroll outsourcing. Notably, there's a consensus on the reduction in time-to-process payroll, cost per payroll processed, and improved payroll accuracy rate. Outsourcing streamlines processes, enhances operational efficiency, and contributes to faster, more accurate financial reporting.
Despite these benefits and challenges, statistical analysis reveals weak correlations between outsourcing global payroll and cost reduction or improved efficiency in various parameters, indicating a lack of a significant relationship. Consequently, the results, suggest no substantial correlation between global payroll outsourcing and enhanced efficiency or cost reduction based on this study's data.
Decarbonisation Strategies in Energy Systems Modelling: APV and e-tractors as Flexibility Assets
(2023)
This work presents an analysis of the impact of introducing Agrophotovoltaic technologies and electric tractors into Germany’s energy system. Agrophotovoltaics involves installing photovoltaic systems in agricultural areas, allowing for dual usage of the land for both energy generation and food production. Electric tractors, which are agricultural machinery powered by electric motors, can also function as energy storage units, providing flexibility to the grid. The analysis includes a sensitivity study to understand how the availability of agricultural land influences Agrophotovoltaic investments, followed by the examination of various scenarios that involve converting diesel tractors to electric tractors. These scenarios are based on the current CO2 emission reduction targets set by the German Government, aiming for a 65% reduction below 1990 levels by 2030 and achieving zero emissions by 2045. The results indicate that approximately 3% of available agricultural land is necessary to establish a viable energy mix in Germany. Furthermore, the expansion of electric tractors tends to reduce the overall system costs and enhances the energy-cost-efficiency of Agrophotovoltaic investments.
KI-gestützte Cyberangriffe
(2023)
Die Fortschritte im Bereich der künstlichen Intelligenz (KI) und des Deep Learning haben in den letzten Jahren enorme Fortschritte gemacht. Insbesondere Technologien wie Large Language Models (LLMs) machen KI-Technologie innerhalb kurzer Zeit zugänglich für die Allgemeinheit. Die Generierung von Text, Bild und Sprache durch künstliche Intelligenz erzielt innerhalb kurzer Zeit gute Ergebnisse. Parallel zu dieser Entwicklung hat die Cyberkriminalität in den vergangenen Jahren im Bereich der KI zugenommen. Cyberangriffe verursachen im Zuge der Digitalisierung größeren Schaden und Angriffe entwickeln sich kontinuierlich weiter, um bestehende Schutzmaßnahmen zu umgehen.
Diese Arbeit bietet eine Einführung in das Themengebiet KI-gestützte Cyberangriffe. Sie präsentiert aktuelle KI-gestützte Cyberangriffsmodelle und analysiert, inwiefern diese für Anfänger*innen in der Cyberkriminalität zugänglich sind.
Das Softwareunternehmen HRworks implementiert eine Personalverwaltungssoftware unter Verwendung der Programmiersprache Smalltalk und des Model-View-Controller (MVC) Musters. Innerhalb des Unternehmens erfordert jede Model-Klasse des Patterns das Vorhandensein einer korrespondierenden Controllerklasse. Controller verfügen über ein wiederkehrendes Grundgerüst, das bei jeder neuen Implementierung umgesetzt werden muss. Die Unterscheidungen innerhalb dieses Grundgerüsts ergeben sich lediglich aus dem Namen und der spezifischen Struktur der korrespondierenden Model-Klasse. Die vorliegende Arbeit adressiert die Herausforderung der automatischen Generierung dieses Controllergrundgerüsts, wobei die Besonderheiten jeder Model-Klasse berücksichtigt wird. Dies wird durch den gezielten Einsatz von Metaprogrammierung in der Programmiersprache Smalltalk realisiert und durch eine Benutzeroberfläche in der Entwicklungsumgebung unterstützt. Zusätzlich wird der Controller um eine Datentypprüfung erweitert, wofür ein spezialisierter Parser implementiert wurde. Dieser extrahiert aus einem definierten Getter der Model-Klasse den entsprechenden Datentyp des Attributes. Im Ergebnis liefert die Arbeit eine Methodik zur automatisierten Generierung und Anpassung von Controllergrundgerüsten sowie dazugehörigen Teststrukturen basierend auf der jeweiligen Model-Klasse. Zusätzlich wird die Funktionalität der Controller durch eine integrierte Datentypprüfung erweitert.
Diese Arbeit befasst sich mit der Entwicklung eines umfassenden Konzepts zur Umsetzung des Lieferkettensorgfaltspflichtengesetzes (LkSG), welches ab dem 01.01.2024 für Unternehmen mit mehr als 1.000 Beschäftigten in Kraft tritt.
Die Bachelorarbeit ist in zwei Teile aufgebaut: zum einen den Theorieteil, der die Lieferkette in einem Verkehrsunternehmen und das Lieferkettensorgfaltspflichtengesetz umfasst, um eine Wissensbasis zu schaffen und zum anderen den Hauptteil, bestehend aus der Konzeptionierung. Da das Gesetz bereits im Jahr 2023 für große Unternehmen mit mehr als 3.000 Beschäftigten in Kraft trat, stehen bereits zahlreiche Systeme zur Unterstützung der Umsetzung zur Verfügung. Im Verlauf der Arbeit wurde eine Marktrecherche durchgeführt, bei der die geeignetste Softwarelösung für SWEG identifiziert wird. Um diese Untersuchung effektiv durchzuführen, wurden zwei Softwareauswahl-Tools entwickelt, die bei der Analyse und Bewertung helfen. Zusätzlich zum Software-Lösungsansatz wurde ein alternativer Lösungsansatz erarbeitet, der die eigenständige Handlung im Unternehmen vorsieht. Diese beiden Lösungsansätze werden zum Schluss verglichen, um die Geschäftsführung bei der Entscheidung, welche Version sie schließlich verwenden möchte, zu unterstützen.
Das erarbeitete Konzept bildet auch die Grundlage für die zukünftige Entwicklung und Umsetzung von Gesetzen und Richtlinien im Bereich Menschenrechte und Umweltaspekte. Für die Zukunft werden weitere Gesetze in Kraft treten, beispielsweise die kommende EU-Richtlinie.
Vorhofflimmern ist die häufigste tachykarde Herzrhythmusstörung weltweit. Dabei verliert das Herz seinen normofrequenten Sinusrhythmus und schlägt nicht mehr regelmäßig, sondern zu schnell und unregelmäßig. Vorhofflimmern ist normalerweise keine lebensbedrohliche Herzrhythmusstörung, aber es kann zu einem Schlaganfall führen. Die Ursache dieser Herzrhythmusstörung sind die Kreisende bzw. die fokalen Erregungen im linken Atrium, die hauptsächliche aus einer oder mehreren Pulmonalvenen kommen. Die übliche Therapieverfahren des Vorhofflimmerns ist die Pulmonalvenenisolation.
Diese Bachelorthesis beschäftigt sich daher mit der Modellierung unterschiedlicher linksatrialer Fokus-Modelle und intrakardialer Elektrodenkatheter für die Diagnostik und Terminierung von Vorhofflimmern mittels Pulmonalvenenisolation im Offenburger Herzrhythmusmodell nach Schalk, Krämer und Benke, welches in CST
Studio Suite realisiert wurde.
Zu Beginn wurden die verschiedenen linksatrialen fokalen Flimmerquellen modelliert und daraufhin simuliert. Hierbei wurde jeweils eine Simulation mit linksatrialen fokalen Flimmerquellen, die aus einzelnen, dualen oder allen vier Pulmonalvenen kommen, durchgeführt. Es wurde ebenfalls eine weitere Simulation mit Biosignalen (aus der Realität) erstellt. Mit diesen Simulationen konnte nun der elektrische Erregungsablauf sichtbar gemacht werden. Daraufhin wurden die Katheter für die Diagnostik und für die Pulmonalvenenisolation modelliert und in das bestehende Offenburger Herzrhythmusmodell integriert. Bei den Diagnostik-Kathetern handelte es sich um 10-polige Lasso® Katheter, zwei Varianten von PentaRay® NAV eco Katheter und 4-polige Diagnostik-Katheter „OSYPKA FINDER pure®“. Ablationskatheter sind zwei Varianten von Pentaspline Basket pose Katheter und HELIOSTAR™ Ablation Ballon. Abschließend wurden verschiedene Varianten von Isolationsverfahren der Pulmonalvenen modelliert und daraufhin die linksatrialen fokalen Flimmerquellen nach der Isolation der Pulmonalvenen simuliert.
Die Bachelorarbeit „Forensic Chain – Verwaltung digitaler Spuren in Deutschland“ untersucht die Anwendung eines Blockchain-basierten Chain of Custody Systems im deutschen rechtlichen und regulatorischen Kontext. Die digitale Forensik, die sich mit der Sicherung und Analyse digitaler Spuren befasst, gewinnt an Bedeutung, da kriminelle Aktivitäten vermehrt im digitalen Raum stattfinden. Die Blockchain-Technologie bietet transparente und unveränderliche Aufzeichnungen, die sich für die Speicherung von Informationen im Zusammenhang mit digitalen Beweismitteln eignen. Das Hautpziel der Arbeit besteht darin, die Umsetzung eines Chain of Custody Prozesses im Forensic Chain System zu untersuchen und die Eignung dieses Systems im deutschen Raum zu bewerten. Hierfür wird ein Prototyp des Forensic Chain Systems entwickelt, um das erstellte Konzept zu testen. Die Ergebnisse tragen zum Verständnis der Wichtigkeit der digitalen Forensik in Deutschland bei und bieten Einblicke in die Einführung von Blockchain-basierten Chain of Custody-Systemen in diesem Bereich. Sie leisten einen Beitrag zur Weiterentwicklung der digitalen Forensik.
Der True-Crime-Podcast erfreut sich in Deutschland einer großen Hörerschaft, die ganz unterschiedliche Nutzungsmotivationen mitbringt. Die Faszination für wahre Kriminalfälle kommt nicht von ungefähr – das Phänomen, welches dieser zugrunde liegt, ist die Morbid Curiosity. Da bislang noch nicht allzu viel Forschung in diesem Gebiet betrieben wurde, gibt diese Arbeit durch einen Vergleich verschiedener Untersuchungen Aufschluss darüber, wo diese Neugier für morbide Ereignisse herkommt. Gleichzeitig soll ein Überblick über das Genre True Crime und das Medium Podcast gegeben werden. Das Ziel ist es, eine Verbindung zwischen dem Phänomen Morbid Curiosity und dem Format True-Crime-Podcast herzustellen. Die Forschungsfrage lautet: Welche Rolle spielt Morbid Curiosity bei der Nutzungsmotivation der Konsumenten von True-Crime-Podcasts? Die Ergebnisse zeigen, dass Frauen andere Nutzungsmotivationen mitbringen als Männer und gleichzeitig die primären Konsumentinnen sind. Die Neugier für morbide Ereignisse spielt dabei durchaus eine Rolle, wenn auch scheinbar eine untergeordnete. Es werden Ansätze für weitere Forschung gegeben.
Digitalisierung ist heute allgegenwärtig. Wo im Privaten bereitwillig neue digitale Tools, Apps und Funktionen genutzt werden, tun sich Unternehmen in der Umsetzung von Digitalisierungsprojekten oft schwer. Dieser Beitrag beleuchtet die Motive für Digitalisierungsvorhaben, ihre Hürden sowie die Auswirkung auf die Arbeitsbelastung von Mitarbeitenden und versucht in der Verknüpfung mit den Grundprinzipien des Kontinuierlichen Verbesserungsprozesses, Handlungsempfehlungen für eine erfolgreiche Umsetzung dieser abzuleiten.
Projektmanagement und mit ihm die PM-Prozesse, Methoden und Werkzeuge entwickeln sich stetig weiter, in kleinen, kaum spürbaren Schritten oder in großen unübersehbaren Veränderungen. In den letzten Jahren war der Diskurs über das Pro & Contra agiler Vorgehensweisen so allgegenwärtig, dass andere Aspekte nicht immer die notwendige Aufmerksamkeit bekamen. Erkannte Notwendigkeiten der PM-Entwicklung konnten noch nicht in spürbare Fortschritte umgewandelt werden. Einflüsse der Globalisierung und der IT, aber auch die aus der zunehmenden Forderung nach Nachhaltigkeit resultierenden Veränderungen in der Projektarbeit sollen daher genauer betrachtet werden. Ist erst einmal die Sensibilität für relevante Trends beim Projektpersonal geschaffen, rücken ein aktualisiertes Kompetenzprofil und ein erweiterter Methodenkanon in greifbare Nähe.
Das Zeitalter der Digitalisierung ist geprägt durch einen erhöhten Wettbewerb. Eine Chance, bei steigendem Wettbewerb erfolgreich zu bestehen, liegt daher nur in der durchgängigen Digitalisierung von Produktionsunternehmen. Dieser Beitrag stellt eine dreistufige generische Unternehmensmodellplattform Industrie 4.0 vor, die die Durchgängigkeit von Prozessen vom Kunden bis zum Lieferanten auf allen Unternehmensebenen in den Mittelpunkt stellt. Die Schritte zur Bewertung und Gestaltung des Fortschritts auf dem Weg zum digitalisierten Produktionsunternehmen werden aufgezeigt.
Die fortschreitende Digitalisierung der Schulen macht es möglich, die Lerndaten der Schülerinnen und Schüler in einer zentralen Cloud zu speichern. Die Befürworter versprechen sich davon eine bessere individuelle Förderung und fordern eine bundesweite Lösung, um möglichst viele Daten auswerten zu können. Die Gegner befürchten eine automatisierte Steuerung des Lernens.
Social-Media-Content - Auswirkungen auf Fear of Missing Out und den Selbstwert junger Nutzer*innen
(2023)
Social-Media-Marketing ist ein wichtiger Baustein einer erfolgreichen Content-Strategie. Insbesondere jüngere Zielgruppen sind auf Social Media anzutreffen – und das oftmals über viele Stunden täglich. Neben den Vorteilen, die Social Media den Nutzer*innen bietet, gibt es aber auch Schattenseiten. Zwei negative Aspekte, die sogenannte Fear of Missing Out und ein verminderter Selbstwert, wurden im Frühjahr 2022 in einer empirischen Befragung von 1338 Personen zwischen 14 und 30 Jahren untersucht. Daneben wurden auch Daten zum grundsätzlichen Social-Media-Nutzungsverhalten erhoben. Die zentralen Erkenntnisse, die sich aus der Studie ableiten, werden in diesem Kapitel vorgestellt und mit Bezug auf ihre Relevanz für das Content-Marketing hin eingeordnet.
Die Verwendung von markenbezogenen nutzer-generierten Inhalten auf den unternehmenseigenen Social-Media-Kanälen ist ein äußerst vielversprechender Ansatz im Content-Marketing. Dabei können durch die authentischen, vom Nutzer bereitgestellten Inhalte zahlreiche Kommunikationsziele erreicht werden. Hierzu gehören etwa die Verstärkung des Nutzerengagements oder aber auch die Förderung von Verkäufen. Daneben müssen allerdings auch Risiken, wie etwa rechtliche Aspekte, beachtet werden. Damit Unternehmen die Potentiale von markenbezogenen nutzer-generierten Inhalten für sich nutzen können, wird im nachstehenden Beitrag ein Strukturierungsrahmen vorgestellt. Dieser fasst die wesentlichen Aspekte dieser durchaus komplexen Thematik strukturiert zusammen. Der hier entwickelte Strukturierungsrahmen wurde durch Experteninterviews überprüft.
In diesem Beitrag werden die psychologischen Hintergründe und Wirkungsweisen des Content-Marketing betrachtet. Nach einer kurzen Einführung in die Thematik wird zuerst das für das weitere Verständnis notwendige psychologische Basiswissen vermittelt. Darauf bezugnehmend wird die allgemeine Wirkungsweise von Content-Marketing beleuchtet. Die Sichtweise wird dann für die letzten beiden Kapitel umgedreht und die beschriebenen psychologischen Faktoren dazu genutzt, um Anwender bei der Wahl der Content-Marketing-Inhalte und zuletzt bei der konkreten Ausgestaltung zu unterstützen.
Die meisten Effekte, die durch Content-Marketing hervorgerufen werden, funktionieren im B2C- oder B2B-Bereich durch das Ansprechen von Bedürfnissen, Interessen und Emotionen sowie die recht freien Entscheidungsmöglichkeiten der Adressaten. Im B2B‑Bereich werden ebenfalls Menschen mit Bedürfnissen, Interessen und Emotionen angesprochen, jedoch vorrangig beruflicher Natur, sodass in der Ausgestaltung geringfügige Unterschiede gemacht werden müssen.
Verfassen guter Texte
(2023)
Wer Texte für seinen Internetauftritt schreibt, möchte, dass diese auch gelesen werden. Doch Lesende sind ungeduldig, insbesondere am Monitor. Fasziniert man sie nicht in den ersten Sekunden, springen sie ab. Erfahren Sie hier, welche stilistischen Regeln Journalistinnen und Journalisten nutzen, um die Aufmerksamkeit ihrer Leser- oder Hörerschaft zu gewinnen und Texte mit wenig Aufwand zu perfektionieren. Ein paar Besonderheiten gelten auch für den Aufbau. Ein Schwerpunkt des Kapitels liegt auf dem Teaser, den ersten Zeilen, die in den Text locken sollen, sowie der Headline. Häufig ist es jedoch nicht der Text, der die Aufmerksamkeit der User fesselt, sondern ein Foto, idealerweise mit einer informativen Bildunterschrift. Zahlreiche Beispiele aus dem journalistischen Alltag machen das Beschriebene anschaulich. Als Zugabe informiert die Autorin Sie über die Bedeutung des Nutzwerts und attraktive Anlässe für eine Veröffentlichung.
Vor dem Hintergrund einer zunehmenden Informations- und Reizüberlastung der Konsumenten werden aus Unternehmenssicht zielgruppenadäquate Inhalte, insbesondere zur Erreichung von kommunikationspolitischen Zielsetzungen, immer wichtiger. Um diese zu gewährleisten, bedarf es einer sinnvollen Planung, Produktion und Distribution von Inhalten. Der vorliegende Beitrag gibt einen Überblick über einen solchen Prozess und veranschaulicht die notwendigen Schritte für ein erfolgreiches Content-Marketing.
Content-Marketing
(2023)
Content-Marketing, also die Planung, Produktion und Distribution von zielgruppen-adäquaten Inhalten, hat insbesondere durch Social Media nochmals an Bedeutung gewonnen. Im Hinblick auf die enorme Menge an Inhalten, die auf Nutzer konstant einwirken, ist es für Unternehmen immer schwieriger, die Aufmerksamkeit der Nutzer zu gewinnen. Nur Inhalte, die den Wünschen der Nutzer entsprechen und diesen in irgendeiner Form einen Mehrwert bieten, haben die Chance, zur Erfüllung von Kommunikationszielen von Unternehmen beizutragen. Die Bereitstellung derartiger Inhalte setzt einen sinnvollen (Planungs-)Prozess voraus. Das vorliegende Buch bietet Praktikern und Studierenden einen Überblick über die verschiedenen Bereiche eines Content-Marketing.
Introduction: Subjects with mild to moderate hearing loss today often receive hearing aids (HA) with open-fitting (OF). In OF, direct sound reaches the eardrums with minimal damping. Due to the required processing delay in digital HA, the amplified HA sound follows some milliseconds later. This process occurs in both ears symmetrically in bilateral HA provision and is likely to have no or minor detrimental effect on binaural hearing. However, the delayed and amplified sound are only present in one ear in cases of unilateral hearing loss provided with one HA. This processing alters interaural timing differences in the resulting ear signals.
Methods: In the present study, an experiment with normal-hearing subjects to investigate speech intelligibility in noise with direct and delayed sound was performed to mimic unilateral and bilateral HA provision with OF.
Results: The outcomes reveal that these delays affect speech reception thresholds (SRT) in the unilateral OF simulation when presenting speech and noise from different spatial directions. A significant decrease in the median SRT from –18.1 to –14.7 dB SNR is observed when typical HA processing delays are applied. On the other hand, SRT was independent of the delay between direct and delayed sound in the bilateral OF simulation.
Discussion: The significant effect emphasizes the development of rapid processing algorithms for unilateral HA provision.
CNN-based deep learning models for disease detection have become popular recently. We compared the binary classification performance of eight prominent deep learning models: DenseNet 121, DenseNet 169, DenseNet 201, EffecientNet b0, EffecientNet lite4, GoogleNet, MobileNet, and ResNet18 for their binary classification performance on combined Pulmonary Chest Xrays dataset. Despite the widespread application in different fields in medical images, there remains a knowledge gap in determining their relative performance when applied to the same dataset, a gap this study aimed to address. The dataset combined Shenzhen, China (CH) and Montgomery, USA (MC) data. We trained our model for binary classification, calculated different parameters of the mentioned models, and compared them. The models were trained to keep in mind all following the same training parameters to maintain a controlled comparison environment. End of the study, we found a distinct difference in performance among the other models when applied to the pulmonary chest Xray image dataset, where DenseNet169 performed with 89.38 percent and MobileNet with 92.2 percent precision.
The COVID19 pandemic, a unique and devastating respiratory disease outbreak, has affected global populations as the disease spreads rapidly. Recent Deep Learning breakthroughs may improve COVID19 prediction and forecasting as a tool of precise and fast detection, however, current methods are still being examined to achieve higher accuracy and precision. This study analyzed the collection contained 8055 CT image samples, 5427 of which were COVID cases and 2628 non COVID. The 9544 Xray samples included 4044 COVID patients and 5500 non COVID cases. The most accurate models are MobileNet V3 (97.872 percent), DenseNet201 (97.567 percent), and GoogleNet Inception V1 (97.643 percent). High accuracy indicates that these models can make many accurate predictions, as well as others, are also high for MobileNetV3 and DenseNet201. An extensive evaluation using accuracy, precision, and recall allows a comprehensive comparison to improve predictive models by combining loss optimization with scalable batch normalization in this study. Our analysis shows that these tactics improve model performance and resilience for advancing COVID19 prediction and detection and shows how Deep Learning can improve disease handling. The methods we suggest would strengthen healthcare systems, policymakers, and researchers to make educated decisions to reduce COVID19 and other contagious diseases.
Virtual-Reality
(2023)
Die Virtual-Reality (VR) Technologie ermöglicht Unternehmen eine Produktpräsentation, die weit über traditionelle Darstellungsmethoden hinausgeht. Obgleich die Integration der VR-Technologie für Unternehmen viele Chancen eröffnet, ist deren Einsatz auch mit Risiken verbunden. Insbesondere der Mangel an empirisch gesicherten Erkenntnissen zur Kundenakzeptanz, zu den Auswirkungen der Nutzung sowie zu Kannibalisierungseffekten ist ein wesentlicher Grund, der die Verbreitung von VR in der Kundenkommunikation noch hemmt. Das Buch adressiert diese Forschungslücken und identifiziert mittels eines nutzerzentrierten, quantitativen Forschungsdesigns konkrete Chancen und Risiken, die mit dem Einsatz von VR-Produktpräsentationen verbunden sind.
This paper presents the new Deep Reinforcement Learning (DRL) library RL-X and its application to the RoboCup Soccer Simulation 3D League and classic DRL benchmarks. RL-X provides a flexible and easy-to-extend codebase with self-contained single directory algorithms. Through the fast JAX-based implementations, RL-X can reach up to 4.5x speedups compared to well-known frameworks like Stable-Baselines3.
The use of artificial intelligence continues to impact a broad variety of domains, application areas, and people. However, interpretability, understandability, responsibility, accountability, and fairness of the algorithms' results - all crucial for increasing humans' trust into the systems - are still largely missing. The purpose of this seminar is to understand how these components factor into the holistic view of trust. Further, this seminar seeks to identify design guidelines and best practices for how to build interactive visualization systems to calibrate trust.
With the rising necessity of explainable artificial intelligence (XAI), we see an increase in task-dependent XAI methods on varying abstraction levels. XAI techniques on a global level explain model behavior and on a local level explain sample predictions. We propose a visual analytics workflow to support seamless transitions between global and local explanations, focusing on attributions and counterfactuals on time series classification. In particular, we adapt local XAI techniques (attributions) that are developed for traditional datasets (images, text) to analyze time series classification, a data type that is typically less intelligible to humans. To generate a global overview, we apply local attribution methods to the data, creating explanations for the whole dataset. These explanations are projected onto two dimensions, depicting model behavior trends, strategies, and decision boundaries. To further inspect the model decision-making as well as potential data errors, a what-if analysis facilitates hypothesis generation and verification on both the global and local levels. We constantly collected and incorporated expert user feedback, as well as insights based on their domain knowledge, resulting in a tailored analysis workflow and system that tightly integrates time series transformations into explanations. Lastly, we present three use cases, verifying that our technique enables users to (1)~explore data transformations and feature relevance, (2)~identify model behavior and decision boundaries, as well as, (3)~the reason for misclassifications.
There is an ongoing debate about the use and scope of Clayton M. Christensen´s idea of disruptive innovation, including the question of whether it is a management buzz phrase or a valuable theory. This discussion considers the general question of how innovation in the field of management theories and concepts finds its way to the different target groups. This conceptual paper combines the different concepts of the creation and dissemination of management trends in a basic framework based on a short review of models for the dissemination of management ideas. This framework allows an analysis of the character of new management ideas like disruptive innovation. By measuring the impact of the theory on the academic sphere using a bibliometric statistic of the number of academic publications on Google scholar and Scopus and a meta-analysis of research papers, we show the significant influence of disruptive innovation beyond pure management fads.
Variable refrigerant flow (VRF) and variable air volume (VAV) systems are considered among the best heating, ventilation, and air conditioning systems (HVAC) thanks to their ability to provide cooling and heating in different thermal zones of the same building. As well as their ability to recover the heat rejected from spaces requiring cooling and reuse it to heat another space. Nevertheless, at the same time, these systems are considered one of the most energy-consuming systems in the building. So, it is crucial to well size the system according to the building’s cooling and heating needs and the indoor temperature fluctuations. This study aims to compare these two energy systems by conducting an energy model simulation of a real building under a semi-arid climate for cooling and heating periods. The developed building energy model (BEM) was validated and calibrated using measured and simulated indoor air temperature and energy consumption data. The study aims to evaluate the effect of these HVAC systems on energy consumption and the indoor thermal comfort of the building. The numerical model was based on the Energy Plus simulation engine. The approach used in this paper has allowed us to reach significant quantitative energy saving along with a high level of indoor thermal comfort by using the VRF system compared to the VAV system. The findings prove that the VRF system provides 46.18% of the annual total heating energy savings and 6.14% of the annual cooling and ventilation energy savings compared to the VAV system.
Modern CNNs are learning the weights of vast numbers of convolutional operators. In this paper, we raise the fundamental question if this is actually necessary. We show that even in the extreme case of only randomly initializing and never updating spatial filters, certain CNN architectures can be trained to surpass the accuracy of standard training. By reinterpreting the notion of pointwise ($1\times 1$) convolutions as an operator to learn linear combinations (LC) of frozen (random) spatial filters, we are able to analyze these effects and propose a generic LC convolution block that allows tuning of the linear combination rate. Empirically, we show that this approach not only allows us to reach high test accuracies on CIFAR and ImageNet but also has favorable properties regarding model robustness, generalization, sparsity, and the total number of necessary weights. Additionally, we propose a novel weight sharing mechanism, which allows sharing of a single weight tensor between all spatial convolution layers to massively reduce the number of weights.
Learning programming fundamentals is considered as one of the most challenging and complex learning activities. Some authors have proposed visual programming language (VPL) approaches to address part of the inherent complexity [1]. A visual programming language lets users develop programs by combining program elements, like loops graphically rather than by specifying them textually. Visual expressions, spatial arrangements of text and graphic symbols are used either as syntax elements or secondary notation. VPLs are normally used for educational multimedia, video games, system development, and data warehousing/business analytics purposes. For example, Scratch, a platform of Massachusetts Institute of Technology, is designed for kids and after school programs.
Design of mobile software applications is considered as one of the most challenging application domains due to the build in sensors as part of a mobile device, like GPS, camera or Near Field Communication (NFC). Sensors enable creation of context-aware mobile applications in which applications can discover and take advantage of contextual information, such as user location, nearby people and objects, and the current user activity. As a consequence, context-aware mobile applications can sense clues about the situational environment making mobile devices more intelligent, adaptive, and personalized. Such context aware mobile applications seem to be motivating and attractive case studies, especially for programming beginners (“my own first app”).
In this work, we introduce a use-case centered approach as well as clear separation of user interface design and sensor-based program development. We provide an in-depth discussion of a new VPL based teaching method, a step by step development process to enable programming beginners the creation of context aware mobile applications. Finally, we argue that addressing challenges for programming beginners by our teaching approach could make programming teaching more motivating, with an additional impact on the final software quality and scalability.
The key contributions of our study are the following:
- An overview of existing attempts to use VPL approaches for mobile applications
- A use case centered teaching approach based on a clear separation of user interface design and sensor-based program development
- A teaching case study enabling beginners a step by step creation of context-aware mobile applications based on the MIT App Inventor (a platform of Massachusetts Institute of Technology)
- Open research challenges and perspectives for further development of our teaching approach
References:
[1] Idrees, M., Aslam, F. (2022). A Comprehensive Survey and Analysis of Diverse Visual Programming Languages, VFAST Transactions on Software Engineering, 2022, Volume 10, Number 2, pp 47-60.
During pyrolysis, biomass is carbonised in the absence of oxygen to produce biochar with heat and/or electricity as co-products making pyrolysis one of the promising negative emission technologies to reach climate goals worldwide. This paper presents a simplified representation of pyrolysis and analyses the impact of this technology on the energy system. Results show that the use of pyrolysis can allow getting zero emissions with lower costs by making changes in the unit commitment of the power plants, e.g. conventional power plants are used differently, as the emissions will be compensated by biochar. Additionally, the process of pyrolysis can enhance the flexibility of energy systems, as it shows a correlation between the electricity generated by pyrolysis and the hydrogen installation capacity, being hydrogen used less when pyrolysis appears. The results indicate that pyrolysis, which is available on the market, integrates well into the energy system with a promising potential to sequester carbon.
TRIZ Innovationstechnologie
(2023)
3D Bin Picking with an innovative powder filled gripper and a torque controlled collaborative robot
(2023)
A new and innovative powder filled gripper concept will be introduced to a process to pick parts out of a box without the use of a camera system which guides the robot to the part. The gripper is a combination of an inflatable skin, and a powder inside. In the unjammed condition, the powder is soft and can adjust to the geometry of the part which will be handled. By applying a vacuum to the inflatable skin, the powder gets jammed and transforms to a solid shaped form in which the gripper was brought before applying the vacuum. This physical principle is used to pick parts. The flexible skin of the gripper adjusts to all kinds of shapes, and therefore, can be used to realize 3D bin picking. With the help of a force controlled robot, the gripper can be pushed with a consistent force on flexible positions depending of the filling level of the box. A Kuka LBR iiwa with joint torque sensors in all of its seven axis’ was used to achieve a constant contact pressure. This is the basic criteria to achieve a robust picking process.
Socially assistive robots (SARs) are becoming more prevalent in everyday life, emphasizing the need to make them socially acceptable and aligned with users' expectations. Robots' appearance impacts users' behaviors and attitudes towards them. Therefore, product designers choose visual qualities to give the robot a character and to imply its functionality and personality. In this work, we sought to investigate the effect of cultural differences on Israeli and German designers' perceptions and preferences regarding the suitable visual qualities of SARs in four different contexts: a service robot for an assisted living/retirement residence facility, a medical assistant robot for a hospital environment, a COVID-19 officer robot, and a personal assistant robot for domestic use. Our results indicate that Israeli and German designers share similar perceptions of visual qualities and most of the robotics roles. However, we found differences in the perception of the COVID-19 officer robot's role and, by that, its most suitable visual design. This work indicates that context and culture play a role in users' perceptions and expectations; therefore, they should be taken into account when designing new SARs for diverse contexts.
Recent advances in spiked shoe design, characterized by increased longitudinal stiffness, thicker midsole foams, and reconfigured geometry are considered to improve sprint performance. However, so far there is no empirical data on the effects of advanced spikes technology on maximal sprinting speed (MSS) published yet. Consequently, we assessed MSS via ‘flying 30m’ sprints of 44 trained male (PR: 10.32 s - 12.08 s) and female (PR: 11.56 s - 14.18 s) athletes, wearing both traditional and advanced spikes in a randomized, repeated measures design. The results revealed a statistically significant increase in MSS by 1.21% on average when using advanced spikes technology. Notably, 87% of participants showed improved MSS with the use of advanced spikes. A cluster analysis unveiled that athletes with higher MSS may benefit to a greater extent. However, individual responses varied widely, suggesting the influence of multiple factors that need detailed exploration. Therefore, coaches and athletes are advised to interpret the promising performance enhancements cautiously and evaluate the appropriateness of the advanced spike technology for their athletes critically.
High-tech running shoes and spikes ("super-footwear") are currently being debated in sports. There is direct evidence that distance running super shoes improve running economy; however, it is not well established to which extent world-class performances are affected over the range of track and road running events.
This study examined publicly available performance datasets of annual best track and road performances for evidence of potential systematic performance effects following the introduction of super footwear. The analysis was based on the 100 best performances per year for men and women in outdoor events from 2010 to 2022, provided by the world governing body of athletics (World Athletics).
We found evidence of progressing improvements in track and road running performances after the introduction of super distance running shoes in 2016 and super spike technology in 2019. This evidence is more pronounced for distances longer than 1500 m in women and longer than 5000 m in men. Women seem to benefit more from super footwear in distance running events than men.
While the observational study design limits causal inference, this study provides a database on potential systematic performance effects following the introduction of super shoes/spikes in track and road running events in world-class athletes. Further research is needed to examine the underlying mechanisms and, in particular, potential sex differences in the performance effects of super footwear.
We revisit the quantitative analysis of the ultrafast magnetoacoustic experiment in a freestanding nickel thin film by Kim and Bigot [J.-W. Kim and J.-Y. Bigot, Phys. Rev. B 95, 144422 (2017)] by applying our recently proposed approach of magnetic and acoustic eigenmode decomposition. We show that the application of our modeling to the analysis of time-resolved reflectivity measurements allows for the determination of amplitudes and lifetimes of standing perpendicular acoustic phonon resonances with unprecedented accuracy. The acoustic damping is found to scale as ∝ω2 for frequencies up to 80 GHz, and the peak amplitudes reach 10−3. The experimentally measured magnetization dynamics for different orientations of an external magnetic field agrees well with numerical solutions of magnetoelastically driven magnon harmonic oscillators. Symmetry-based selection rules for magnon-phonon interactions predicted by our modeling approach allow for the unambiguous discrimination between spatially uniform and nonuniform modes, as confirmed by comparing the resonantly enhanced magnetoelastic dynamics simultaneously measured on opposite sides of the film. Moreover, the separation of timescales for (early) rising and (late) decreasing precession amplitudes provide access to magnetic (Gilbert) and acoustic damping parameters in a single measurement.
While most ultrafast time-resolved optical pump-probe experiments in magnetic materials reveal the spatially homogeneous magnetization dynamics of ferromagnetic resonance (FMR), here we explore the magneto-elastic generation of GHz-to-THz frequency spin waves (exchange magnons). Using analytical magnon oscillator equations, we apply time-domain and frequency-domain approaches to quantify the results of ultrafast time-resolved optical pump-probe experiments in free-standing ferromagnetic thin films. Simulations show excellent agreement with the experiment, provide acoustic and magnetic (Gilbert) damping constants and highlight the role of symmetry-based selection rules in phonon-magnon interactions. The analysis is extended to hybrid multilayer structures to explore the limits of resonant phonon-magnon interactions up to THz frequencies.
The technique of laser ultrasonics perfectly meets the need for noncontact, noninvasive, nondestructive mechanical probing of nanometer- to millimeter-size samples. However, this technique is limited to the excitation of low-amplitude strains, below the threshold for optical damage of the sample. In the context of strain engineering of materials, alternative optical techniques enabling the excitation of high-amplitude strains in a nondestructive optical regime are needed. We introduce here a nondestructive method for laser-shock wave generation based on additive superposition of multiple laser-excited strain waves. This technique enables strain generation up to mechanical failure of a sample at pump laser fluences below optical ablation or melting thresholds. We demonstrate the ability to generate nonlinear surface acoustic waves (SAWs) in Nb-SrTiO3 substrates, with associated strains in the percent range and pressures up to 3 GPa at 1 kHz repetition rate and close to 10 GPa for several hundred shocks. This study paves the way for the investigation of a host of high-strain SAW-induced phenomena, including phase transitions in conventional and quantum materials, plasticity and a myriad of material failure modes, chemistry and other effects in bulk samples, thin layers, and two-dimensional materials.
The utilisation of artificial intelligence (AI) is progressively emerging as a significant mechanism for innovation in human resource management (HRM). The capacity to facilitate the transformation of employee performance across numerous responsibilities. AI development, there remains a dearth of comprehensive exploration into the potential opportunities it presents for enhancing workplace performance among employees. To bridge this gap in knowledge, the present work carried out a survey with 300 participants, utilises a fuzzy set-theoretic method that is grounded on the conceptualisation of AI, KS, and HRM. The findings of our study indicate that the exclusive adoption of AI technologies does not adequately enhance HRM engagements. In contrast, the integration of AI and KS offers a more viable HRM approach for achieving optimal performance in a dynamic digital society. This approach has the potential to enhance employees’ proficiency in executing their responsibilities and cultivate a culture of creativity inside the firm.
Purpose
Although start-ups have gained increasing scholarly attention, we lack sufficient understanding of their entrepreneurial strategic posture (ESP) in emerging economies. The purpose of this study is to examine the processes of ESP of new technology venture start-ups (NTVs) in an emerging market context.
Design/methodology/approach
In line with grounded theory guidelines and the inductive research traditions, the authors adopted a qualitative approach involving 42 in-depth semi-structured interviews with Ghanaian NTV entrepreneurs to gain a comprehensive analysis at the micro-level on the entrepreneurs' strategic posturing. A systematic procedure for data analysis was adopted.
Findings
From the authors' analysis of Ghanaian NTVs, the authors derived a three-stage model to elucidate the nature and process of ESP Phase 1 spotting and exploiting market opportunities, Phase II identifying initial advantages and Phase III ascertaining and responding to change.
Originality/value
The study contributes to advancing research on ESP by explicating the process through which informal ties and networks are utilised by NTVs and NTVs' founders to overcome extreme resource constraints and information vacuums in contexts of institutional voids. The authors depart from past studies in demonstrating how such ties can be harnessed in spotting and exploiting market opportunities by NTVs. On this basis, the paper makes original contributions to ESP theory and practice.
Purpose
Although recent literature has examined diverse measures adopted by SMEs to navigate the COVID-19 turbulence, there is a shortage of evidence on how crisis-time strategy creation behaviour and digitalization activities increase (1) sales and (2) cash flow. Thus, predicated on a novel strategy creation perspective, this inquiry aims to investigate the crisis behaviour, sales and cash flow performance of 528 SMEs in Morocco.
Design/methodology/approach
Novel links between (1) aggregate wage cuts, (2) variable operating hours, (3) deferred payment to suppliers, (4) deferred payment to tax authorities and (5) sales performance are developed and tested. A further link between sales performance and cash flow is also examined and the analysis is conducted using a non-linear structural equation modelling technique.
Findings
While there is a significant association between strategy creation behaviours and sales performance, only variable operating hours have a positive effect. Also, sales performance increases cash flow and this relationship is substantially strengthened by e-commerce digitalization and innovation.
Originality/value
Theoretically, to the best of the authors’ knowledge, this is one of the first inquiries to espouse the strategy creation view to explain SMEs' crisis-time behaviour and digitalization. For practical purposes, to supplement Moroccan SMEs' propensity to seek tax deferrals, it is argued that debt and equity support measures are also needed to boost sales performance and cash flow.
In the past ten years, applications of artificial neural networks have changed dramatically. outperforming earlier predictions in domains like robotics, computer vision, natural language processing, healthcare, and finance. Future research and advancements in CNN architectures, Algorithms and applications are expected to revolutionize various industries and daily life further. Our task is to find current products that resemble the given product image and description. Deep learning-based automatic product identification is a multi-step process that starts with data collection and continues with model training, deployment, and continuous improvement. The caliber and variety of the dataset, the design selected, and ongoing testing and improvement all affect the model's effectiveness. We achieved 81.47% training accuracy and 72.43% validation accuracy for our combined text and image classification model. Additionally, we have discussed the outcomes from the other dataset and numerous methods for creating an appropriate model.
An international study summarizes the threat situation in the OT environment under the heading "Growing security threats" [1]. According to this study, attacks on automation systems are likely to increase in the future. Accordingly, an automation system must be able to protect the integrity of the transmitted information in the future. This requirement is motivated, among other things, by the fact that the network-side isolation of industrial communication systems is no longer considered sufficient as the sole protective measure. This paper uses the example of PROFINET to show how the future requirements for a real-time communication protocol can be met and how they can be derived from the IEC 62443 standard.
As the population grows, so does the amount of biowaste. As demand for energy grows, biogas is a promising solution to the problem. Lignocellulosic materials are challenged of slow degradability due to the presence of polymers such as cellulose, lignin and hemicellulose. There are several pretreatment methods available to enhance the degradability of such materials, including enzymatic pretreatment. In this pretreatment, there are few parameters that can influence the results, the most important being the enzyme to solid ratio and the solid to liquid ratio. During this project, experiments were conducted to determine the optimal conditions for those two factors. It was discovered that a solid to liquid ratio of 31 g of buffer per 1 gram of organic dry matter produced the highest reducing sugar release in flasks when combined with 34 mg of protein per 1 gram of organic dry mass. Additionally, another experiment was carried out to investigate the impact of enzymatic pretreatment on biogas production using artificial biowaste as a substrate. Artificial biowaste produced 577,9 NL/kg oDM, while enzymatically pretreated biowaste produced 639,3 NL/kg oDM. This resulted in a 10,6% rise in cumulative biogas production compared to its use without enzymatic pretreatment. By the conclusion of the investigation, specific cumulative dry methane yields of 364,7 NL/kg oDM and 426,3 NL/kg oDM were obtained from artificial biowaste without and with enzymatic pretreatment, respectively. This resulted in a methane production boost of 16,9%. Additionally in case of the reactors with enzymatically pretreated substrate kinetic constant was lower more than double, where maximum volume of biogas increased, comparing to the reactors without enzymatic pretreatment.
Polyarticulated active prostheses constitute a promising solution for upper limb amputees. The bottleneck for their adoption though, is the lack of intuitive control. In this context, machine learning algorithms based on pattern recognition from electromyographic (EMG) signals represent a great opportunity for naturally operating prosthetic devices, but their performance is strongly affected by the selection of input features. In this study, we investigated different combinations of 13 EMG-derived features obtained from EMG signals of healthy individuals performing upper limb movements and tested their performance for movement classification using an Artificial Neural Network. We found that input data (i.e., the set of input features) can be reduced by more than 50% without any loss in accuracy, while diminishing the computing time required to train the classifier. Our results indicate that input features must be properly selected in order to optimize prosthetic control.
The main focus of this chapter is the theoretical and instrumental processes that underpin densitometric methods widely used in thin-layer chromatography (TLC). Densitometric methods include UV–vis, luminescence, and fluorescence optical measurements as well as infrared and Raman spectroscopic measurements. The chapter is divided in two general parts: a theoretical part and a practical part. The systems for direct radioactivity measurements and the combination of TLC with mass spectrometry are also discussed. All these systems allow measuring an intensity distribution directly on a TLC plate. We call this “in situ detection” because no analyte is removed from the plate.
Recently, photovoltaic (PV) with energy storage systems (ESS) have been widely adopted in buildings to overcome growing power demands and earn financial benefits. The overall energy cost can be optimized by combining a well-sized hybrid PV/ESS system with an efficient energy management system (EMS). Generally, EMS is implemented within the overall functions of the Building Automation System (BAS). However, due to its limited computing resources, BAS cannot handle complex algorithms that aim to optimize energy use in real-time under different operating conditions. Furthermore, islanding the building's local network to maximize the PV energy share represents a challenging task due to the potential technical risks. In this context, this article addresses an improved approach based on upgrading the BAS data analytics capability by means of an edge computing technology. The edge communicates with the BAS low-level controller using a serial communication protocol. Taking advantage of the high computing ability of the edge device, an optimization-based EMS of the PV/ESS hybrid system is implemented. Different testing scenarios have been carried out on a real prototype with different weather conditions, and the results show the implementation feasibility and technical performance of such advanced EMS for the management of building energy resources. It has also been proven to be feasible and advantageous to operate the local energy network in island mode while ensuring system safety. Additionally, an estimated energy saving improvement of 6.23 % has been achieved using optimization-based EMS compared to the classical rule-based EMS, with better ESS constraints fulfillment.
Following the traditional paradigm of convolutional neural networks (CNNs), modern CNNs manage to keep pace with more recent, for example transformer-based, models by not only increasing model depth and width but also the kernel size. This results in large amounts of learnable model parameters that need to be handled during training. While following the convolutional paradigm with the according spatial inductive bias, we question the significance of \emph{learned} convolution filters. In fact, our findings demonstrate that many contemporary CNN architectures can achieve high test accuracies without ever updating randomly initialized (spatial) convolution filters. Instead, simple linear combinations (implemented through efficient 1×1 convolutions) suffice to effectively recombine even random filters into expressive network operators. Furthermore, these combinations of random filters can implicitly regularize the resulting operations, mitigating overfitting and enhancing overall performance and robustness. Conversely, retaining the ability to learn filter updates can impair network performance. Lastly, although we only observe relatively small gains from learning 3×3 convolutions, the learning gains increase proportionally with kernel size, owing to the non-idealities of the independent and identically distributed (\textit{i.i.d.}) nature of default initialization techniques.
We have developed a methodology for the systematic generation of a large image dataset of macerated wood references, which we used to generate image data for nine hardwood genera. This is the basis for a substantial approach to automate, for the first time, the identification of hardwood species in microscopic images of fibrous materials by deep learning. Our methodology includes a flexible pipeline for easy annotation of vessel elements. We compare the performance of different neural network architectures and hyperparameters. Our proposed method performs similarly well to human experts. In the future, this will improve controls on global wood fiber product flows to protect forests.
State-of-the-art models for pixel-wise prediction tasks such as image restoration, image segmentation, or disparity estimation, involve several stages of data resampling, in which the resolution of feature maps is first reduced to aggregate information and then sequentially increased to generate a high-resolution output. Several previous works have investigated the effect of artifacts that are invoked during downsampling and diverse cures have been proposed that facilitate to improve prediction stability and even robustness for image classification. However, equally relevant, artifacts that arise during upsampling have been less discussed. This is significantly relevant as upsampling and downsampling approaches face fundamentally different challenges. While during downsampling, aliases and artifacts can be reduced by blurring feature maps, the emergence of fine details is crucial during upsampling. Blurring is therefore not an option and dedicated operations need to be considered. In this work, we are the first to explore the relevance of context during upsampling by employing convolutional upsampling operations with increasing kernel size while keeping the encoder unchanged. We find that increased kernel sizes can in general improve the prediction stability in tasks such as image restoration or image segmentation, while a block that allows for a combination of small-size kernels for fine details and large-size kernels for artifact removal and increased context yields the best results.
Fix your downsampling ASAP! Be natively more robust via Aliasing and Spectral Artifact free Pooling
(2023)
Convolutional neural networks encode images through a sequence of convolutions, normalizations and non-linearities as well as downsampling operations into potentially strong semantic embeddings. Yet, previous work showed that even slight mistakes during sampling, leading to aliasing, can be directly attributed to the networks' lack in robustness. To address such issues and facilitate simpler and faster adversarial training, [12] recently proposed FLC pooling, a method for provably alias-free downsampling - in theory. In this work, we conduct a further analysis through the lens of signal processing and find that such current pooling methods, which address aliasing in the frequency domain, are still prone to spectral leakage artifacts. Hence, we propose aliasing and spectral artifact-free pooling, short ASAP. While only introducing a few modifications to FLC pooling, networks using ASAP as downsampling method exhibit higher native robustness against common corruptions, a property that FLC pooling was missing. ASAP also increases native robustness against adversarial attacks on high and low resolution data while maintaining similar clean accuracy or even outperforming the baseline.
Motivated by the recent trend towards the usage of larger receptive fields for more context-aware neural networks in vision applications, we aim to investigate how large these receptive fields really need to be. To facilitate such study, several challenges need to be addressed, most importantly: (i) We need to provide an effective way for models to learn large filters (potentially as large as the input data) without increasing their memory consumption during training or inference, (ii) the study of filter sizes has to be decoupled from other effects such as the network width or number of learnable parameters, and (iii) the employed convolution operation should be a plug-and-play module that can replace any conventional convolution in a Convolutional Neural Network (CNN) and allow for an efficient implementation in current frameworks. To facilitate such models, we propose to learn not spatial but frequency representations of filter weights as neural implicit functions, such that even infinitely large filters can be parameterized by only a few learnable weights. The resulting neural implicit frequency CNNs are the first models to achieve results on par with the state-of-the-art on large image classification benchmarks while executing convolutions solely in the frequency domain and can be employed within any CNN architecture. They allow us to provide an extensive analysis of the learned receptive fields. Interestingly, our analysis shows that, although the proposed networks could learn very large convolution kernels, the learned filters practically translate into well-localized and relatively small convolution kernels in the spatial domain.
Assessing the robustness of deep neural networks against out-of-distribution inputs is crucial, especially in safety-critical domains like autonomous driving, but also in safety systems where malicious actors can digitally alter inputs to circumvent safety guards. However, designing effective out-of-distribution tests that encompass all possible scenarios while preserving accurate label information is a challenging task. Existing methodologies often entail a compromise between variety and constraint levels for attacks and sometimes even both. In a first step towards a more holistic robustness evaluation of image classification models, we introduce an attack method based on image solarization that is conceptually straightforward yet avoids jeopardizing the global structure of natural images independent of the intensity. Through comprehensive evaluations of multiple ImageNet models, we demonstrate the attack's capacity to degrade accuracy significantly, provided it is not integrated into the training augmentations. Interestingly, even then, no full immunity to accuracy deterioration is achieved. In other settings, the attack can often be simplified into a black-box attack with model-independent parameters. Defenses against other corruptions do not consistently extend to be effective against our specific attack.
Project website: https://github.com/paulgavrikov/adversarial_solarization
Entity Matching (EM) defines the task of learning to group objects by transferring semantic concepts from example groups (=entities) to unseen data. Despite the general availability of image data in the context of many EM-problems, most currently available EM-algorithms solely rely on (textual) meta data. In this paper, we introduce the first publicly available large-scale dataset for "visual entity matching", based on a production level use case in the retail domain. Using scanned advertisement leaflets, collected over several years from different European retailers, we provide a total of ~786k manually annotated, high resolution product images containing ~18k different individual retail products which are grouped into ~3k entities. The annotation of these product entities is based on a price comparison task, where each entity forms an equivalence class of comparable products. Following on a first baseline evaluation, we show that the proposed "visual entity matching" constitutes a novel learning problem which can not sufficiently be solved using standard image based classification and retrieval algorithms. Instead, novel approaches which allow to transfer example based visual equivalent classes to new data are needed to address the proposed problem. The aim of this paper is to provide a benchmark for such algorithms.
Information about the dataset, evaluation code and download instructions are provided under https://www.retail-786k.org/.
For the treatment of bone defects, biodegradable, compressive biomaterials are needed as replacements that degrade as the bone regenerates. The problem with existing materials has either been their insufficient mechanical strength or the excessive differences in their elastic modulus, leading to stress shielding and eventual failure. In this study, the compressive strength of CPC ceramics (with a layer thickness of more than 12 layers) was compared with sintered β-TCP ceramics. It was assumed that as the number of layers increased, the mechanical strength of 3D-printed scaffolds would increase toward the value of sintered ceramics. In addition, the influence of the needle inner diameter on the mechanical strength was investigated. Circular scaffolds with 20, 25, 30, and 45 layers were 3D printed using a 3D bioplotter, solidified in a water-saturated atmosphere for 3 days, and then tested for compressive strength together with a β-TCP sintered ceramic using a Zwick universal testing machine. The 3D-printed scaffolds had a compressive strength of 41.56 ± 7.12 MPa, which was significantly higher than that of the sintered ceramic (24.16 ± 4.44 MPa). The 3D-printed scaffolds with round geometry reached or exceeded the upper limit of the compressive strength of cancellous bone toward substantia compacta. In addition, CPC scaffolds exhibited more bone-like compressibility than the comparable β-TCP sintered ceramic, demonstrating that the mechanical properties of CPC scaffolds are more similar to bone than sintered β-TCP ceramics.
Differentiation between human and non-human objects can increase efficiency of human-robot collaborative applications. This paper proposes to use convolutional neural networks for classifying objects in robotic applications. The body temperature of human beings is used to classify humans and to estimate the distance to the sensor. Using image classification with convolutional neural networks it is possible to detect humans in the surroundings of a robot up to five meters distance with low-cost and low-weight thermal cameras. Using transfer learning technique we trained the GoogLeNet and MobilenetV2. Results show accuracies of 99.48 % and 99.06 % respectively.
Detecting Images Generated by Deep Diffusion Models using their Local Intrinsic Dimensionality
(2023)
Diffusion models recently have been successfully applied for the visual synthesis of strikingly realistic appearing images. This raises strong concerns about their potential for malicious purposes. In this paper, we propose using the lightweight multi Local Intrinsic Dimensionality (multiLID), which has been originally developed in context of the detection of adversarial examples, for the automatic detection of synthetic images and the identification of the according generator networks. In contrast to many existing detection approaches, which often only work for GAN-generated images, the proposed method provides close to perfect detection results in many realistic use cases. Extensive experiments on known and newly created datasets demonstrate that the proposed multiLID approach exhibits superiority in diffusion detection and model identification.Since the empirical evaluations of recent publications on the detection of generated images are often mainly focused on the "LSUN-Bedroom" dataset, we further establish a comprehensive benchmark for the detection of diffusion-generated images, including samples from several diffusion models with different image sizes.The code for our experiments is provided at https://github.com/deepfake-study/deepfake-multiLID.
Following their success in visual recognition tasks, Vision Transformers(ViTs) are being increasingly employed for image restoration. As a few recent works claim that ViTs for image classification also have better robustness properties, we investigate whether the improved adversarial robustness of ViTs extends to image restoration. We consider the recently proposed Restormer model, as well as NAFNet and the "Baseline network" which are both simplified versions of a Restormer. We use Projected Gradient Descent (PGD) and CosPGD for our robustness evaluation. Our experiments are performed on real-world images from the GoPro dataset for image deblurring. Our analysis indicates that contrary to as advocated by ViTs in image classification works, these models are highly susceptible to adversarial attacks. We attempt to find an easy fix and improve their robustness through adversarial training. While this yields a significant increase in robustness for Restormer, results on other networks are less promising. Interestingly, we find that the design choices in NAFNet and Baselines, which were based on iid performance, and not on robust generalization, seem to be at odds with the model robustness.
The identification of vulnerabilities is an important element in the software development life cycle to ensure the security of software. While vulnerability identification based on the source code is a well studied field, the identification of vulnerabilities on basis of a binary executable without the corresponding source code is more challenging. Recent research [1] has shown how such detection can generally be enabled by deep learning methods, but appears to be very limited regarding the overall amount of detected vulnerabilities. We analyse to what extent we could cover the identification of a larger variety of vulnerabilities. Therefore, a supervised deep learning approach using recurrent neural networks for the application of vulnerability detection based on binary executables is used. The underlying basis is a dataset with 50,651 samples of vulnerable code in the form of a standardised LLVM Intermediate Representation. Te vectorised features of a Word2Vec model are used to train different variations of three basic architectures of recurrent neural networks (GRU, LSTM, SRNN). A binary classification was established for detecting the presence of an arbitrary vulnerability, and a multi-class model was trained for the identification of the exact vulnerability, which achieved an out-of-sample accuracy of 88% and 77%, respectively. Differences in the detection of different vulnerabilities were also observed, with non-vulnerable samples being detected with a particularly high precision of over 98%. Thus, our proposed technical approach and methodology enables an accurate detection of 23 (compared to 4 [1]) vulnerabilities.
The importance of machine learning (ML) has been increasing dramatically for years. From assistance systems to production optimisation to healthcare support, almost every area of daily life and industry is coming into contact with machine learning. Besides all the benefits ML brings, the lack of transparency and difficulty in creating traceability pose major risks. While solutions exist to make the training of machine learning models more transparent, traceability is still a major challenge. Ensuring the identity of a model is another challenge, as unnoticed modification of a model is also a danger when using ML. This paper proposes to create an ML Birth Certificate and ML Family Tree secured by blockchain technology. Important information about training and changes to the model through retraining can be stored in a blockchain and accessed by any user to create more security and traceability about an ML model.
Grundzüge der Strömungslehre
(2023)
Dieses ausgereifte Lehrbuch stellt in prägnant kurzer und mathematisch verständlicher Darstellung die strömungstechnischen Grundlagen dar. Aufgaben mit Lösungen helfen den Lernstoff richtig anzuwenden und fördern das Verständnis. Das Buch eignet sich zur Begleitung und Vertiefung der Vorlesungen über Strömungslehre sowie zum Selbststudium. Die vorliegende Auflage geht auf die immer größer werdende Rolle des Energiehaushalts ein und trägt damit den aktuellen Entwicklungen Rechnung. Ergänzt wurden aktuelle Übungsaufgaben der Strömungsmechanik, zahlreiche Beispiele veranschaulichen den Energiesatz.
Automation devices or automation stations (AS) take on the task of controlling, regulating, monitoring and, if necessary, optimising building systems and their system components (e.g. pumps, compressors, fans) based on recorded process variables. For this purpose, a wide range of control and regulation methods are used, starting with simple on/off controllers, through classic PID controllers, to higher-order controllers such as adaptive, model-predictive, knowledge-based or adaptive controllers.
Starting with a brief introduction to automation technology (Sect. 7.1), the chapter goes into the structure and functionality of the usual compact controllers using the application examples of solar thermal systems and heat pump systems (Sect. 7.2). Finally, the integration of system automation into a higher-level building automation system and into the building management system is described using specific application examples (Sect. 7.3).
This central book chapter now details the implementation of automation of solar domestic hot water systems, solar assisted building heating, rooms, solar cooling systems, heat pump heating systems, geothermal systems and thermally activated building component systems. Hydraulic and automation diagrams are used to explain how the automation of these systems works. A detailed insight into the engineering and technical interrelationships involved in the use of these systems, as well as the use of simulation tools, enables effective control and regulation. System characteristic curves and systematic procedures support the automation engineer in his tasks.
Renewable energy sources such as solar radiation, geothermal heat and ambient heat are available for energy conversion. With the help of special converters, these resources can be put to use. These include solar collectors, geothermal probes and chillers. They collect the energy and convert it to a temperature level high enough to be suitable for heat purposes. In the case of refrigeration machines, a distinction is made between electrically and thermally driven machines.
The use of renewable energy sources for heating and cooling in buildings today offers the best opportunities to avoid the use of fossil fuels and the associated climate-damaging emissions. However, unlike fossil fuels, renewable energy sources such as solar radiation are not available at the push of a button, but occur uncontrollably depending on weather conditions, the location of the building and the time of year. Their use is free of charge. However, complex converters and systems usually have to be installed in order to use them. These must be carefully planned and operated in order to avoid unnecessary costs and to generate the maximum possible yield. The regenerative energy systems are usually integrated into existing conventional systems. When designing the control and regulation equipment, it is crucial to design the automation of the systems in such a way that primarily renewable energy sources are used and the share of fossil energy sources is minimized.
This textbook helps use regenerative systems for heating and cooling effectively. Integration and automation schemes provide a quick overview. Practical examples clearly show standard solutions for the integration of regenerative energy sources. For the 2nd edition, improvements have been made to the text and illustrations, and references to standards have been updated. Control questions at the end of the main chapters serve to consolidate the understanding of the content.
Public educational institutions are increasingly confronted with a decline in the number of applicants, which is why competition between colleges and universities is also intensifying. For this reason, it is important to position oneself as an institution in order to be perceived by the various target groups and to differentiate oneself from the competition. In this context, the brand and thus its perception and impact play a decisive role, especially in view of the desired communication of the institution's own values and its self-image, the brand identity. To this end, emotions serve as an approach to creating positive stimulation and brand loyalty.
In this study, circular economy (CE) relevance in Germany will be discussed based on LinkedIn readily available data. LinkedIn company profiles located in Germany with ‘circular economy’ in their description or any other field were selected and used as a data source to analyze their CE relation. Overall, 514 German companies were analyzed in reference to the 15 German regions they belong. Most companies are located in the federal state of Berlin (126), followed by North Rhine-Westphalia (96) and Bavaria (77). In terms of the industry sector, they are self-classified to environmental services (64), management consulting (50), renewables & environment (33), research (31), and computer software (18) etc. Regarding their employees with LinkedIn profiles, 22,621 people are affiliated with these companies, ranging from one to 7,877. All examined companies have a total of 819,632 followers on LinkedIn, ranging from none to 88,167. An increase in CE-related companies was recorded in 13 of the 16 federal states of Germany over a one-year period. This work provides essential insights into the increasing relevance and trends of the circular economy in German enterprises and will help conduct further national studies with readily available data from LinkedIn.
Human interaction frequently includes decision-making processes during which interactants call on verbal and non-verbal resources to manage the flow of interaction. In 2017, Stevanovic et al. carried out pioneering work, analyzing the unfolding of moment-by-moment dynamics by investigating the behavioral matching during search and decision-making phases. By studying the similarities in the participant's body sway during a conversation task in Finnish, the authors showed higher behavioral matching during decision phases than during search phases. The purpose of this research was to investigate the whole-body sway and its coordination during joint search and decision-making phases as a replication of the study by Stevanovic et al. (2017) but based on a German population. Overall, 12 dyads participated in this study and were asked to decide on 8 adjectives, starting with a pre-defined letter, to describe a fictional character. During this joint-decision task (duration: 206.46 ± 116.08 s), body sway of both interactants was measured using a 3D motion capture system and center of mass (COM) accelerations were computed. Matching of body sway was calculated using a windowed cross correlation (WCC) of the COM accelerations. A total of 101 search and 101 decision phases were identified for the 12 dyads. Significant higher COM accelerations (5.4*10−3 vs. 3.7*10−3 mm/s2, p < 0.001) and WCC coefficients (0.47 vs. 0.45, p = 0.043) were found during decision-making phases than during search phases. The results suggest that body sway is one of the resources humans use to communicate the arrival at a joint decision. These findings contribute to a better understanding of interpersonal coordination from a human movement science perspective.
Femtosecond (fs) time-resolved magneto-optics is applied to investigate laser-excited ultrafast dynamics of one-dimensional nickel gratings on fused silica and silicon substrates for a wide range of periodicities Λ = 400–1500 nm. Multiple surface acoustic modes with frequencies up to a few tens of GHz are generated. Nanoscale acoustic wavelengths Λ/n have been identified as nth-spatial harmonics of Rayleigh surface acoustic wave (SAW) and surface skimming longitudinal wave (SSLW), with acoustic frequencies and lifetimes being in agreement with theoretical calculations. Resonant magnetoelastic excitation of the ferromagnetic resonance (FMR) by SAW’s third spatial harmonic, and, most interestingly fingerprints of the parametric resonance at 1/2 SAW frequency have been observed. Numerical solutions of Landau–Lifshitz–Gilbert (LLG) equation magnetoelastically driven by complex polychromatic acoustic fields quantitatively reproduce all resonances at once. Thus, our results provide a solid experimental and theoretical base for a quantitative understanding of ultrafast fs-laser-driven magnetoacoustics and tailoring the magnetic-grating-based metasurfaces at the nanoscale.
Die Erfindung betrifft in einem ersten Aspekt eine Vorrichtung zur transkutanen Aufbringung eines elektrischen Stimulationsreizes auf ein Ohr. Die Vorrichtung umfasst einen Schaltungsträger, mindestens zwei Elektroden sowie eine Steuerungseinheit, wobei die Steuerungseinheit dazu konfiguriert ist, anhand von Stimulationsparametern ein elektrisches Stimulationssignal an den Elektroden zu erzeugen. Dabei ist die Vorrichtung, insbesondere eine Oberfläche des Schaltungsträgers der Vorrichtung, auf eine anatomische Form eines Ohres angepasst, sodass Elektroden auf der Oberfläche des Schaltungsträgers aufgebracht sind und ausgewählte Bereiche des Ohres kontaktieren Die Vorrichtung ist dadurch kennzeichnet, dass diese weiterhin einen Sensor zur Erkennung mindestens eines physiologischen Parameter umfasst und eine Steuerungseinheit dazu konfiguriert ist, anhand des mindestens einen physiologischen Parameters die Stimulationsparameter für den Stimulationsreiz anzupassen.In einem weiteren Aspekt betrifft die Erfindung ein Verfahren zur Herstellung der erfindungsgemäßen Vorrichtung.
Die Erfindung betrifft ein Verfahren zum Maximieren der von einer analogen Entropiequelle abgeleiteten Entropie, wobei das Verfahren folgende Schritte aufweist:- Bereitstellen von Eingabedaten für die analoge Entropiequelle (2);- Erzeugen von Rückgabewerten durch die analoge Entropiequelle basierend auf den Eingabedaten (3); und- Gruppieren der Rückgabewerte, wobei das Gruppieren der Rückgabewerte ein Anwenden von Versätzen auf Rückgabewerte aufweist (4).
Bei vielen Schulungen, Unterrichten und Weiterbildungen kommen Präsentationen zum Einsatz, um ausbildungsrelevante Inhalte zu vermitteln. Oft sind diese jedoch nicht interessant und zielführend gestaltet, was sich z. B. durch ein Übermaß an Text auszeichnet. Die Autoren stellen alternativ eine visualisierte Aufbereitung von Inhalten vor. Ziel ist es, komplexe Sachverhalte als einfache Bilder und Skizzen komprimiert darzustellen. Mit Hilfe der vorgestellten Methoden können beispielsweise Übungen effizienter vorbereitet, Einsätze übersichtlich erfasst, aber auch alltägliche Situationen vereinfacht kommuniziert werden.
Sustainable Production
(2023)
Visual programming languages (VPL) let users develop software programs by combining visual program elements, like lists of objects, loops or conditional statements rather than by specifying them textually.
Humanoid robots programming is a very attractive and motivating application domain for students, especially for programming beginners. Humanoid robots are constructed in such a way that they mimic the human body by using actuators that perform like muscles. Typically, a humanoid robot consists of sensors and actuators, i.e. torso, a head, two arms, and two legs, though some humanoid robots may replicate only part of the body, for example, from the waist up. In some cases, humanoid robots are equipped with heads designed to replicate additional human facial features such as eyes. Additional sensors are needed by a robot to gather information about the conditions of the environment to allow the robot to make necessary decisions about its position or certain actions that the situation requires, e.g. an arm movement or an open/close hand action. Other examples for sensor are reflective infrared sensors used to detect objects in proximity.
In this work, we introduce a use-case centered approach based on sensors and actors of a robot and a workflow model to visually describe the sequence of actions including conditional actions or concurrent actions. We provide an in-depth discussion of a new VPL based teaching method for programming humanoid robots based on VPLs. Open research challenges, limits and perspectives for further development of our teaching approach are discussed as well.
Sensors and actuators enable creation of context-aware applications in which applications can discover and take advantage of contextual information, such as user location, nearby people and objects. In this work, we use a general context definition, which can be applied to various devices, e.g., robots and mobile devices. Developing context-based software applications is considered as one of the most challenging application domains due to the sensors and actuators as part of a device. We introduce a new development approach for context-based applications by using use-case descriptions and Visual Programming Languages (VPL). The introduction of web-based VPLs, such as Scratch and Snap, has reinvigorated the usefulness of VPLs. We provide an in-depth discussion of our new VPL based method, a step by step development process to enable development of context-based applications. Two case studies illustrate how to apply our approach to different problem domains: Context-based mobile apps and context-based humanoid robot applications.
The main advantage of mobile context-aware applications is to provide effective and tailored services by considering the environmental context, such as location, time, nearby objects and other data, and adapting their functionality according to the changing situations in the context information without explicit user interaction. The idea behind Location-Based Services (LBS) and Object-Based Services (OBS) is to offer fully-customizable services for user needs according to the location or the objects in a mobile user's vicinity. However, developing mobile context-aware software applications is considered as one of the most challenging application domains due to the built-in sensors as part of a mobile device. Visual Programming Languages (VPL) and hybrid visual programming languages are considered to be innovative approaches to address the inherent complexity of developing programs. The key contribution of our new development approach for location and object-based mobile applications is a use case driven development approach based on use case templates and visual code templates to enable even programming beginners to create context-aware mobile applications. An example of the use of the development approach is presented and open research challenges and perspectives for further development of our approach are formulated.