Refine
Year of publication
Document Type
- Conference Proceeding (151)
- Article (reviewed) (58)
- Bachelor Thesis (43)
- Article (unreviewed) (17)
- Master's Thesis (10)
- Book (7)
- Part of a Book (5)
- Letter to Editor (2)
- Patent (2)
- Contribution to a Periodical (1)
Conference Type
- Konferenzartikel (122)
- Konferenz-Abstract (23)
- Sonstiges (5)
- Konferenz-Poster (1)
Language
- English (196)
- German (100)
- Other language (1)
- Russian (1)
Keywords
- Kommunikation (6)
- Applikation (5)
- Batterie (5)
- Brennstoffzelle (5)
- Netzwerk (4)
- Codierung (3)
- Eingebettetes System (3)
- Elektrode (3)
- Energieversorgung (3)
- Gehirn (3)
Institute
- Fakultät Elektrotechnik und Informationstechnik (E+I) (bis 03/2019) (298) (remove)
Open Access
- Closed Access (298) (remove)
This thesis deals with the creation of a cross-platform application using Xamarin.Forms. The cross-platform application will cover three different platforms android, iOS, and UWP.
The application is the first concept of a possible feature for a companion application for LS telcom. There, the user can identify cell antennas using a map-view and a camera-view making the application an augmented reality application. Thus, the user can search for a specific cell and access various information that he would not be able to see with his eyes like for example the frequency of the transmitting cells.
The cell data is generated from three different sources, Cartoradio, OpenCelliD, and the LS telcom databrowser. Eventually, the decision was taken, that the main source should be the LS telcom databrowser which has multiple advantages over the other cell sources.
The cells on the map-view are placed using the extracted coordinates from the source data. However, the cells on the camera-view are placed with complex calculations using different formulas like the Haversine formula to calculate the distance between the cell and the user and the bearing to calculate the angle between the cell and the user. Various settings will allow the user to personalize the application according to his wishes.
In previous work we [1] and other authors (e.g. [2]) have shown that agent-based systems are successful in optimizing delivery plans of single logistics companies and are meanwhile successfully productive in industry. In this paper we show that agent-based systems are particularly useful to also optimize transport across logistics companies. In intercompany optimization, privacy is of major importance between the otherwise competing companies. Some data has to be treated strictly private like the cost model or the constraint model. Other data like order information has to be shared. However, typically the amount of orders released to other companies has also to be limited. We show that our agent-based approach can be easily fine tuned to trade off privacy against the benefit of cooperation.
Mit Hilfe eines Präzisionsmessplatzes soll es ermöglicht werden, automatisierte Tests mit optischen Distanzsensoren der Firma SICK durchzuführen. Hierbei handelt es sich um applikationsbezogene Vergleichsmessungen. Für die Realisierung einer erweiterbaren, automatischen Ansteuerung wird mit LabVIEW eine Software entwickelt, die unterschiedliche Distanzsensoren für Displacement Anwendungen (kurze Reichweite) einbindet. Zur Bewertung von Sensoren werden unterschiedliche Messmodi bereitgestellt. Hierbei werden motorisierte Linearachsen angesteuert, wodurch dynamische 2D-, bzw. 3D Messungen von unterschiedlichen Materialproben ermöglicht werden. Außerdem können Messergebnisse verschiedener Materialproben visuell verglichen werden. Des Weiteren besteht die Möglichkeit, aufgenommene Messdaten zu exportieren.
Mit der realisierten Ansteuerungssoftware ist es möglich, in Zukunft Mitarbeiter der global agierenden Sales & Service Units ressourcenschonend und in einem anwendungsbezogenen Kontext in die Benutzung und Bewertung von Dis-placementsensoren einzuarbeiten. Für diese Maßnahme der betrieblichen Weiterbildung ist eine Lerneinheit konzipiert. Hier geht es hauptsächlich darum, dass Mitarbeiter die Eigenschaften und Konfigurationsmöglichkeiten von Displacementsensoren verstehen und für unterschiedliche Testobjekte anwenden. Für die Lerneinheit sind Unterrichtsmaterialien erstellt sowie ein vollständiger Unterrichtsentwurf erarbeitet. Der Unterrichtsentwurf orientiert sich an dem Perspektivenschema nach Klafki (vgl. 1994, S. 270ff.).
Entwicklung und Implementierung einer Methode zur Funktions- und Verschleißprüfung von Sägeanlagen
(2019)
Die Firma KASTO Maschinenbau produziert Lagersysteme und Sägemaschinen und möchte in Zukunft mit Hilfe von prädiktiver Wartung einen Mehrwert für den Kunden, die firmeneigenen Servicetechniker und die Inbetriebnehmer generieren. Für diesen Weg in Richtung prädiktive Wartung werden in der vorliegenden Arbeit, zunächst mittels Recherche, Grundlagen definiert. Die anschließende Analyse verschiedener Bauteile und Baugruppen in der Lagertechnik und im Sägemaschinenbau führt zur Konkretisierung der Umsetzungsmöglichkeiten.
Im weiteren Verlauf der Arbeit werden die verschleißkritischen Bauteile eines KASTOwin-Bandsägeautomats behandelt. Durch die Analyse der einzelnen Verschleißformen und -erscheinungen an dessen Komponenten können Parameter zur Verschleiß-überwachung ermittelt werden. Die Überwachung dieser Prozesse soll dabei mit bestehender Sensorik durchgeführt werden und beschränkt sich deshalb hauptsächlich auf die Messeinrichtung der Frequenzumformer. Für die Komponenten Sägebandantrieb, Sägevorschubantrieb, Materialvorschubantrieb, Bandverlaufsensor, sowie das Hydrauliksystem werden entsprechende Methoden zur Funktions- und Verschleiß-prüfung ausgearbeitet.
Auf Basis dieser Methoden wird dann die Implementierung in Form eines SPS-Programms und einer dazu passenden, grafischen Benutzeroberfläche durchgeführt. Das bestehende SPS-Programm und die Benutzeroberfläche werden durch zusätzliche Programmteile ergänzt. Sämtliche Ergänzungen sind portabel und modular ausgelegt, sodass diese problemlos auch in anderen Sägemaschinen ergänzt werden können. Um kritische Verschleißsituationen nach deren Eintritt zu analysieren und daraus Schlüsse im Sinne der prädiktiven Wartung zu ziehen, ist die Sicherung aller Langzeitdaten von enormer Bedeutung. Da der remanente Speicher der SPS jedoch begrenzt ist, werden sämtliche Daten über die Visualisierung gesichert. Alle aufgenommen Daten werden sowohl grafisch als auch in Form von Werten visualisiert und können dem Kunden und Servicetechniker, sowie dem Inbetriebnehmer Aufschluss über den Zustand der Komponenten geben.
Letztlich werden die implementierten Abläufe an einer Vorführmaschine getestet. Es kann dabei die Funktion sämtlicher Prüfungen bestätigt werden
This thesis deals with the implementation of the SUBSCALE algorithm in the Python programming language. First, the current state of research and the needs of the target group are considered. Then, the choice of language is decided based on the findings. On the basis of self-generated requirements, the implementation is carried out.
Finally, the code is evaluated for accuracy, consistency, and execution time, as well as its applicability in practice.
Since the implementation of the current work proved to be unconvincing, an approach is tested in which Python is used only as a front-end.
Bei der Produktion von Solarzellen aus multikristallinem Silizium haben Defekte aus der Kristallisationsphase starken Einfluss auf die Materialqualität der Wafer und damit auf den Wirkungsgrad der späteren Solarzelle. Ein Verständnis des Kornwachstums in multikristallinem Silizium während des Kristallisationsprozesses kann zur Optimierung desselben beitragen. In dieser Arbeit werden Methoden untersucht, optische Flüsse zwischen Korngrenzenbildern multikristalliner Si-Wafer mittels neuronaler Netze zu berechnen. Hierfür wird die Architektur eines ausgereiften faltungsbasierten neuronalen Netzes zur optischen Fluss-Berechnung genutzt und durch angepasstes Training auf Waferstrukturen zugeschnitten. Dies umfasst die Synthese eigener, auf Waferbilder basierender Trainingsdaten und das Training mit einer angepassten Fehlerfunktion zur Bewertung der Zuordnungsgenauigkeit von Körnern zwischen Wafern durch den optischen Fluss. Beide Maßnahmen zusammen führen zu einer Reduktion des Zuordnungsfehlers von Körnern zwischen Waferbildern um 45 % gegenüber einem hochoptimierten, auf allgemeine optische Flüsse trainierten Modell basierend auf demselben Netzwerk. Die geschätzte Zuordnungsgenauigkeit des besten Modells beträgt 92,4 % der Pixel der Korngrenzenbilder eines Wafers. Weiteres Verbesserungspotenzial ist vorhanden.
Am Buffet des Lebens
(2017)
In letzter Zeit sind einige neue, hochintegrierte Einchip-Radarsensoren auf den Markt gekommen. Die enormen Fortschritte im Bereich des autonomen Fahrens hat diese Sensoren hervorgebracht. Mit ihnen lassen sich diverse Anwendungen, wie zum Beispiel eine Abstandsmessung, Kollisionserkennung oder Geschwindigkeitserfassung realisieren.
Für die Nutzung eines solchen modernen Radarsensors spricht viel, jedoch besitzen alle eine differenzielle Ausgangsschnittstelle, die nicht mit den üblichen Mikrocontrollersystemen eingelesen werden kann. Darum war das Ziel der Arbeit, die Entwicklung eines Schnittstellenwandlers auf einem Low-Power-FPGA, zur Anbindung eines Radarchips an einen klassischen Mikrocontroller.
Der Lösungsweg war demnach schon vorgegeben, es folgte die konkrete Umsetzung mit der Modellierung der Hardware in VHDL. Der FPGA liest die differenzielle Schnittstelle ein, parallelisiert die Daten und speichert sie zwischen. Sobald die Messdaten vollständig sind, können sie über die serielle SPI-Schnittstelle angefordert werden. Als Gegenstelle kommt ein Mikrocontroller zum Einsatz, der die Messdaten wiederum gemäß eines definierten Protokolls zur Auswertung an einen Computer weiterleitet.
Die Machbarkeit dieser Anwendung wurde kontrolliert, indem die Messdaten vom Radarchip, übermittelt durch den FPGA und Mikrocontroller, auf dem Computer mithilfe eines Analyseprogramms bewertet wurden. Die Auswertung der Messergebnisse entspricht in vollem Umfang den Erwartungen. Der Ressourcenverbrauch im FPGA wurde hierbei ebenfalls als kritisch betrachtet, was sich im Nachhinein jedoch nicht bestätigte. Es ist sogar das Gegenteil der Fall, mit den übrigen freien Ressourcen steht einer möglichen Signalverarbeitung nichts im Wege.
Introduction: Patient selection for cardiac resynchronization therapy (CRT) requires quantification of left ventricular conduction delay (LVCD). After implantation of biventricular pacing systems, individual AV delay (AVD) programming is essential to ensure hemodynamic response. To exclude adverse effects, AVD should exceed individual implant-related interatrial conduction times (IACT). As result of a pilot study, we proposed the development of a programmer-based transoesophageal left heart electrogram (LHE) recording to simplify both, LVCD and IACT measurement. This feature was implemented into the Biotronik ICS3000 programmer simultaneously with 3-channel surface ECG.
Methods: A 5F oesophageal electrode was perorally applied in 44 heart failure CRT-D patients (34m, 10f, 65±8 yrs., QRS=162±21ms). In position of maximum left ventricular deflection, oesophageal LVCD was measured between onsets of QRS in surface ECG and oesophageal left ventricular deflection. Then, in position of maximum left atrial deflection (LA), IACT in VDD operation (As-LA) was calculated by difference between programmed AV delay and the measured interval from onset of left atrial deflection to ventricular stimulus in the oesophageal electrogram. IACT in DDD operation (Ap-LA) was measured between atrial stimulus and LA..
Results: LVCD of the CRT patients was characterized by a minimum of 47ms with mean of 69±23ms. As-LA and Ap-LA were found to be 41±23ms and 125±25ms, resp., at mean. In 7 patients (15,9%), IACT measurement in DDD operation uncovered adverse AVD if left in factory settings. In this cases, Ap-LA exceeded the factory AVD. In 6 patients (13,6%), IACT in VDD operation was less than or equal 10ms indicating the need for short AVD.
Conclusion: Response to CRT requires distinct LVCD and AVD optimization. The ICS3000 oesophageal LHE feature can be utilized to measure LVCD in order to justify selection for CRT. IACT measurement simplifies AV delay optimization in patients with CRT systems irrespective of their make and model.
In-vivo and in-vitro comparison of implant-based CRT optimization - What provide new algorithms?
(2011)
Introduction: In cardiac resynchronization therapy (CRT), individual AV delay (AVD) optimization can effectively increase hemodynamics and reduce non-responder rate. Accurate, automatic and easily comprehensible algorithms for the follow-up are desirable. QuickOpt is the first attempt of a semi-automatic intracardiac electrogram (IEGM) based AVD algorithm. We aimed to compare its accuracy and usefulness by in-vitro and in-vivo studies.
Methods: Using the programmable ARSI-4 four-chamber heart rhythm and IEGM simulator (HKP, Germany), the QuickOpt feature of an Epic HF system (St. Jude, USA) was tested in-vitro by simulated atrial IEGM amplitudes between 0.3 and 3.5mV during both, manual and automatic atrial sensing between 0.2 and 1.0mV. Subsequently, in 21 heart failure patients with implanted biventricular defibrillators, QuickOpt was performed in-vivo. Results of the algorithm for VDD and DDD stimulation were compared with echo AV delay optimization.
Results: In-vitro simulations demonstrated a QuickOpt measuring accuracy of ± 8ms. Depending on atrial IEGM amplitude, the algorithm proposed optimal AVD between 90 and 150ms for VDD and between 140 and 200ms for DDD operation, respectively. In-vivo, QuickOpt difference between individual AVD in DDD and VDD mode was either 50ms (20pts) or 40ms (1pt). QuickOpt and echo AVD differed by 41 ± 25ms (7 – 90ms) in VDD and by 18 ± 24ms (17-50ms) in DDD operation. Individual echo AVD difference between both modes was 73 ± 20ms (30-100ms).
Conclusion: The study demonstrates the value of in-vitro studies. It predicted QuickOpt deficiencies regarding IEGM amplitude dependent AVD proposals constrained to fixed individual differences between DDD and VDD mode. Consequently, in-vivo, the algorithm provided AVD of predominantly longer duration than echo in both modes. Accepting echo individualization as gold standard, QuickOpt should not be used alone to optimize AVD in CRT patients.
Introduction: To simplify AV delay (AVD) optimization in cardiac resynchronization therapy (CRT), we reported that the hemodynamically optimal AVD for VDD and DDD mode CRT pacing can be approximated by individually measuring implant-related interatrial conduction intervals (IACT) in oesophageal electrogram (LAE) and adding about 50ms. The programmer-based St Jude QuickOpt algorithm is utilizing this finding. By automatically measuring IACT in VDD operation, it predicts the sensed AVD by adding either 30ms or 60ms. Paced AVD is strictly 50ms longer than sensed AVD. As consequence of those variations, several studies identified distinct inaccuracies of QuickOpt. Therefore, we aimed to seek for better approaches to automate AVD optimization.
Methods: In a study of 35 heart failure patients (27m, 8f, age: 67±8y) with Insync III Marquis CRT-D systems we recorded telemetric electrograms between left ventricular electrode and superior vena cava shock coil (LVtip/SVC = LVCE) simultaneously with LAE. By LVCE we measured intervals As-Pe in VDD and Ap-Pe in DDD operation between right atrial sense-event (As) or atrial stimulus (Ap), resp., and end of the atrial activity (Pe). As-Pe and Ap-Pe were compared with As-LA an Ap-LA in LAE, respectively.
Results: End of the left atrial activity in LVCE could clearly be recognized in 35/35 patients in VDD and 29/35 patients in DDD operation. We found mean intervals As-LA of 40.2±24.5ms and Ap-LA of 124.3±20.6ms. As-Pe was 94.8±24.1ms and Ap-Pe was 181.1±17.8ms. Analyzing the sums of As-LA + 50ms with duration of As-Pe and Ap-LA + 50ms with duration of Ap-Pe, the differences were 4.7±9.2ms and 4.2±8.6ms, resp., only. Thus, hemodynamically optimal timing of the ventricular stimulus can be triggered by automatically detecting Pe in LVCE.
Conclusion: Based on minimal deviations between LAE and LVCE approach, we proposed companies to utilize the LVCE in order to automate individual AVD optimization in CRT pacing.
Das Projekt PHOTOPUR soll die Reduzierung von Pestiziden in Oberflächengewässern ermöglichen. In dieser Arbeit wird eine Automatisierung eines ersten Demosystems entwickelt, welches den gesamten Reinigungsprozess abbildet. Eine Projektierung der Automatisierung des Systems wird mit den dafür vorgesehenen Fließschemas und Gerätelisten durchgeführt. Darauf aufbauend wird die Ablaufsteuerung des Demosystems durch einen Ablauf-Funktionsplan umgesetzt. Um eine Systemüberwachung der Anlage zu gewährleisten wurde dazu eine Visualisierung ausgearbeitet. Zusätzlich wurden die Regelstrecken der Durchflussregelungen in den zwei Teilprozessen des Reinigungsprozesses bestimmt und durch unterschiedliche Einstellregeln der optimale Regler der Regelkreise ermittelt.
Die in dieser Arbeit entwickelte Software, beinhaltet die drei folgenden Umsetzungen: Realisierung der Ablaufsteuerung, Implementierung der Reglerparameter durch einen vorhandenen Regelalgorithmus und die Visualisierung des Demosystems.
Message co chairmen
(2017)
In this study, a high-performance controller is proposed for single-phase grid-tied energy storage systems (ESSs). To control power factor and current harmonics and manage time-shifting of energy, the ESS is required to have low steady-state error and fast transient response. It is well known that fast controllers often lack the required steady-state accuracy and trade-off is inevitable. A hybrid control system is therefore presented that combines a simple yet fast proportional derivative controller with a repetitive controller which is a type of learning controller with small steady-state error, suitable for applications with periodic grid current harmonic waveforms. This results in an improved system with distortion-free, high power factor grid current. The proposed controller model is developed and design parameters are presented. The stability analysis for the proposed system is provided and the theoretical analysis is verified through stability, transient and steady-state simulations.
Exploiting Dissent: Towards Fuzzing-based Differential Black Box Testing of TLS Implementations
(2017)
The Transport Layer Security (TLS) protocol is one of the most widely used security protocols on the internet. Yet do implementations of TLS keep on suffering from bugs and security vulnerabilities. In large part is this due to the protocol's complexity which makes implementing and testing TLS notoriously difficult. In this paper, we present our work on using differential testing as effective means to detect issues in black-box implementations of the TLS handshake protocol. We introduce a novel fuzzing algorithm for generating large and diverse corpuses of mostly-valid TLS handshake messages. Stimulating TLS servers when expecting a ClientHello message, we find messages generated with our algorithm to induce more response discrepancies and to achieve a higher code coverage than those generated with American Fuzzy Lop, TLS-Attacker, or NEZHA. In particular, we apply our approach to OpenssL, BoringSSL, WolfSSL, mbedTLS, and MatrixSSL, and find several real implementation bugs; among them a serious vulnerability in MatrixSSL 3.8.4. Besides do our findings point to imprecision in the TLS specification. We see our approach as present in this paper as the first step towards fully interactive differential testing of black-box TLS protocol implementations. Our software tools are publicly available as open source projects.
The instability of ultra-thin films of an electrolyte bordering a dielectric gas in an external tangential electric field is scrutinized. The solid wall is assumed to be either a conducting or charged dielectric surface. The problem has a steady one-dimensional solution. The theoretical results for a plug-like velocity profile are successfully compared with available experimental data. The linear stability of the steady-state flow is investigated analytically and numerically. Asymptotic long-wave expansion has a triple-zero singularity for a dielectric wall and a quadruple-zero singularity for a conducting wall, and four (for a conducting wall) or three (for a charged dielectric wall) different eigenfunctions. For infinitely small wave numbers, these eigenfunctions have a clear physical meaning: perturbations of the film thickness, of the surface charge, of the bulk conductivity, and of the bulk charge. The numerical analysis provides an important result: the appearance of a strong short-wave instability. At increasing Debye numbers, the short-wave instability region becomes isolated and eventually disappears. For infinitely large Weber numbers, the long-wave instability disappears, while the short-wave instability persists. The linear stability analysis is complemented by a nonlinear direct numerical simulation. The perturbations evolve into coherent structures; for a relatively small external electric field, these are large-amplitude surface solitary pulses, while for a sufficiently strong electric field, these are short-wave inner coherent structures, which do not disturb the surface.
A wide range catalyst screening with noble metal and oxide catalysts for a metal–air battery with an aqueous alkaline electrolyte was carried out. Suitable catalysts reduce overpotentials during the charge and discharge process, and therefore improve the round-trip efficiency of the battery. In this case, the electrodes will be used as optimized cathodes for a future lithium–air battery with an aqueous alkaline electrolyte. Oxide catalysts were synthesized via atmospheric plasma spraying. The screening showed that IrO2, RuO2, La0.6Ca0.4Co3, Mn3O4, and Co3O4 are promising bi-functional catalysts. Considering the high price for the noble metal catalysts further investigations of the oxide catalysts were carried out to analyze their electrochemical behavior at varied temperatures, molarities, and in case of La1−x Ca x CoO3 a varying calcium content. Additionally all catalysts were tested in a longterm test to proof cyclability at varied molarities. Further investigations showed that Co3O4 seems to be the most promising bi-functional catalyst of the tested oxide catalysts. Furthermore, it was shown that a calcium content of x = 0.4 in LCCO has the best performance.
Private households constitute a considerable share of Europe's electricity consumption. The current electricity distribution system treats them as effectively passive individual units. In the future, however, users of the electricity grid will be involved more actively in the grid operation and can become part of intelligent networked collaborations. They can then contribute the demand and supply flexibility that they dispose of and, as a result, help to better integrate renewable energy in-feed into the distribution grids.
High-precision signal processing algorithm to evaluate SAW properties as a function of temperature
(2013)
This paper presents a signal processing algorithm which accurately evaluates the SAW properties of a substrate as functions of temperature. The investigated acoustic properties are group velocity, phase velocity, propagation loss, and coupling coefficient. With several measurements carried out at different temperatures, we obtain the temperature dependency of the SAW properties. The analysis algorithm starts by reading the transfer functions of short and long delay lines. The analysis algorithm determines the center frequency of the delay lines and obtains the delay time difference between the short and long delay lines. The extracted parameters are then used to calculate the acoustic properties of the SAW material. To validate the algorithm, its accuracy is studied by determining the error in the calculating delay time difference, center frequency, and group velocity.
Background: Increasing awareness of the importance of evidence-based medicine is demonstrated not only by an increasing number of articles addressing it but also by a specialty-wide evidence-based medicine initiative. The authors critically analyzed the quality of reporting of randomized controlled trials published in this Journal over a 21-year period (1990 to 2010).
Methods: A hand search was conducted, including all issues of Plastic and Reconstructive Surgery from January of 1990 to December of 2010. All randomized controlled trials published during this time period were identified with the Cochrane decision tree for identification of randomized controlled trials. To assess the quality of reporting, a modification of the checklist of the Consolidated Standard of Reporting Trials Statement was used.
Results: Of 7121 original articles published from 1990 to 2010 in the Journal, 159 (2.23 percent) met the Cochrane criteria. A significant increase in the absolute number of randomized controlled trials was seen over the study period (p < 0.0001). The median quality of these trials from 1990 to 2010 was "fair," with a trend toward improved quality of reporting over time (p = 0.127).
Conclusions: A favorable trend is seen with respect to an increased number of published randomized controlled trials in Plastic and Reconstructive Surgery. Adherence to standard reporting guidelines is recommended, however, to further improve the quality of reporting. Consideration may be given to providing information regarding the quality of reporting in addition to the "level of evidence pyramid," thus facilitating critical appraisal.
There is an increasing demand by an ever-growing number of mobile customers for transfer of rich media content. This requires very high bandwidth which either cannot be provided by the current cellular systems or puts pressure on the wireless networks, affecting customer service quality. This study introduces COARSE – a novel cluster-based quality-oriented adaptive radio resource allocation scheme, which dynamically and adaptively manages the radio resources in a cluster-based two-hop multi-cellular network, having a frequency reuse of one. COARSE is a cross-layer approach across physical layer, link layer and the application layer. COARSE gathers data delivery-related information from both physical and link layers and uses it to adjust bandwidth resources among the video streaming end-users. Extensive analysis and simulations show that COARSE enables a controlled trade-off between the physical layer data rate per user and the number of users communicating using a given resource. Significantly, COARSE provides 25–75% improvement in the computed user-perceived video quality compared with that obtained from an equivalent single-hop network.
Multi-phase management is crucial for performance and durability of electrochemical cells such as batteries and fuel cells. In this paper we present a generic framework for describing the two-dimensional spatiotemporal evolution of gaseous, liquid and solid phases, as well as their interdependence with interfacial (electro-)chemistry and microstructure in a continuum description. The modeling domain consists of up to seven layers (current collectors, channels, electrodes, separator/membrane), each of which can consist of an arbitrary number of bulk phases (gas, liquid, solid) and connecting interfaces (two-phase or multi-phase boundaries). Bulk and interfacial chemistry is described using global or elementary kinetic reactions. Multi-phase management is coupled to chemistry and to mass and charge transport within bulk phases. The functionality and flexibility of this framework is demonstrated using four application areas in the context of post-lithium-ion batteries and fuel cells, that is, lithium-sulfur (Li-S) cells, lithium-oxygen (Li-O) cells, solid oxide fuel cells (SOFC) and polymer electrolyte membrane fuel cells (PEFC). The results are compared to models available in literature and properties of the generic framework are discussed.
The separation of nitrogen and methane from hydrogen-rich mixtures is systematically investigated on a recently developed binder-free zeolite 5A. For this adsorbent, the present work provides a series of experimental data on adsorption isotherms and breakthrough curves of nitrogen and methane, as well as their mixtures in hydrogen. Isotherms were measured at temperatures of 283–313 K and pressures of up to 1.0 MPa. Breakthrough curves of CH4, N2, and CH4/N2 in H2 were obtained at temperatures of 300–305 K and pressures ranging from 0.1 to 6.05 MPa with different feed concentrations. An LDF-based model was developed to predict breakthrough curves using measured and calculated data as inputs. The number of parameters and the use of correlations were restricted to focus on the importance of measured values. For the given assumptions, the results show that the model predictions agree satisfactorily with the experiments under the different operating conditions applied.
Das Ingenieurbüro Evomotiv GmbH und die University of Applied Sciences Offenburg entwickeln seit Ende 2008 ein Antriebskonzept für leichte Stadtfahrzeuge. Ziel des Elektroantriebs ist die Serientauglichkeit der hochintegrierten, getriebelosen und eisenfreien Radnabenmotoren. Das Bundesministerium für Wirtschaft und Technologie (BMWi) unterstützt das Projekt. Das Konzept des Radnabenmotors erhielt 2006 den Bosch—Innovationspreis und gewann 2008 den Shell-Eco-Marathon. Im Jahr 2011 wird Evomotiv mit seinen Partnern erste Prototypen auf der Straße testen.
We present a novel fabrication route yielding Raney-platinum film electrodes intended as glucose oxidation anodes for potentially implantable fuel cells. Fabrication roots on thermal alloying of an extractable metal with bulk platinum at 200 °C for 48 h. In contrast to earlier works using carcinogenic nickel, we employ zinc as potentially biocompatible alloying partner. Microstructure analysis indicates that after removal of extractable zinc the porous Raney-platinum film (roughness factor ∼2700) consists predominantly of the Pt3Zn phase. Release of zinc during electrode operation can be expected to have no significant effect on physiological normal levels in blood and serum, which promises good biocompatibility. In contrast to previous anodes based on hydrogel-bound catalyst particles the novel anodes exhibit excellent resistance against hydrolytic and oxidative attack. Furthermore, they exhibit significantly lower polarization with up to approximately 100 mV more negative electrode potentials in the current density range relevant for fuel cell operation. The anodes’ amenability to surface modification with protective polymers is demonstrated by the exemplary application of an approximately 300 nm thin Nafion coating. This had only a marginal effect on the anode long-term stability and amino acid tolerance. While in physiological glucose solution after approximately 100 h of operation gradually increasing performance degradation occurs, rapid electrode polarization within 24 h is observed in artificial tissue fluid. Optimization approaches may include catalyst enhancement by adatom surface modification and the application of specifically designed protective polymers with controlled charge and mesh size.
We report the fabrication and characterization of glucose-tolerant Raney-platinum cathodes for oxygen reduction in potentially implantable glucose fuel. Fabricated by extraction of aluminum from 1 μm thin platinum–aluminum bi-layers annealed at 300 °C, the novel cathodes show excellent resistance against hydrolytic and oxidative attack. This renders them superior over previous cathodes fabricated from hydrogel-bound catalyst particles. Annealing times of 60, 120, and 240 min result in approximately 400–550 nm thin porous films (roughness factors ∼100–150), which contain platinum and aluminum in a ratio of ∼9:1. Aluminum release during electrode operation can be expected to have no significant effect on physiological normal levels, which promises good biocompatibility. Annealing time has a distinct influence on the density of trenches formed in the cathode. Higher trench densities lead to lower electrode potentials in the presence of glucose. This suggests that glucose sensitivity is governed by mixed potential formation resulting from oxygen depletion within the trenches. During performance characterization the diffusion resistance to be expected from tissue capsule formation upon electrode implantation was taken into account by placing a membrane in front of the cathode. Despite the resulting limited oxygen supply, cathodes prepared by annealing for 60 min show more positive electrode potentials than previous cathodes fabricated from hydrogel-bound activated carbon. Compared to operation in phosphate buffered saline containing 3.0 mM glucose, a potential loss of approximately 120 mV occurs in artificial tissue fluid. This can be reduced to approximately 90 mV with a protective Nafion layer that is easily electro-coated onto the Raney-platinum film.
A theoretical description is given for the propagation of surface acoustic wave pulses in anisotropic elastic media subject to the influence of nonlinearity. On the basis of nonlinear elasticity theory, an evolution equation is presented for the surface slope or the longitudinal surface velocity associated with an acoustic pulse. It contains a non-local nonlinearity, characterized by a kernel that strongly varies from one propagation geometry to another due to the anisotropy of the substrate. It governs pulse shape evolution in homogeneous halfspaces and the shapes of solitary surface pulses that exist in coated substrates. The theory describing nonlinear Rayleigh-type surface acoustic waves is extended in a straightforward way to surface waves that are localized at a one-dimensional acoustic waveguide like elastic wedges.
This paper presents a multicarrier code-division multiple-access (CDMA) system architecture that is based on complete complementary orthogonal spreading codes. The architecture has several advantages as compared to conventional CDMA systems. Specifically, it offers multiple-access interference-free operation in additive white Gaussian noise channels, reduces cochannel interference significantly, and has the potential of higher capacity and spectral efficiency than conventional CDMA systems. This is accomplished by using an ldquooffset stackedrdquo spreading modulation technique. To maintain good performance in the presence of fading, the offset stacked modulator is followed by a quadrature-amplitude modulation map, which is designed to optimize performance in a fading environment. This new modulation scheme also simplifies the rate-matching algorithms that are relevant for multimedia services and Internet Protocol-based applications.
Design of next-generation cdma using orthogonal complementary codes and offset stacked spreading
(2007)
This article presents an innovative code-division multiple access system architecture that is based on orthogonal complementary spreading codes and time-frequency domain spreading. The architecture has several advantages compared to conventional CDMA systems. Specifically, it offers multiple-access-interference-free operation in AWGN channels, reduces co-channel interference significantly, and has the potential for higher capacity and spectral efficiency than conventional CDMA systems. This is accomplished by using an "offset stacked" spreading modulation technique followed by quadrature amplitude modulation, which optimizes performance in a fading environment. This new spreading modulation scheme also simplifies the rate matching algorithms relevant for multimedia services and IP-based applications.
Experimental and theoretical investigations of the time of equalization of the concentration of an impurity in a rectangular flow‐type chamber have been carried out. It has been shown that the process of equalization of the concentration with time is exponential in character. The characteristic equalization time has been computed using the theory of turbulent diffusion. Theoretical results describe experimental regularities with an accuracy of about 10%. The value of the coefficient of turbulent diffusion for different configurations of flows in the chamber has been obtained from a comparison of experimental and calculated results.
The importance of obtaining simultaneous particle size and concentration values has grown up with continuing discussion of the health effects, of internal combustion engine generated particulate emissions and in particular of Diesel soot emissions. In the present work an aerosol measurement system is described that delivers information about particle size and concentration directly from the undiluted exhaust gas.
Using three laser diodes of different wavelengths which form one parallel light beam, each spectral attenuation is analysed by a single detector and the particle diameter and concentration is evaluated by the use of the Mie theory and shown on-line at a frequency of 1 Hz. The system includes an optical long-path-cell (White principle) with an adjustable path length from 2.5 to 15 m, which allows the analysis within a broad concentration range.
On-line measurements of the particulate emissions in the hot, undiluted exhaust of Diesel engines are presented under stationary and transient engine’s load conditions. Mean particle diameters well below 100 nm are detected for modern Diesel engines. The measured particle concentration corresponds excellently with the traditional gravimetrical measurements of the diluted exhaust. Additionally, measurements of particle emissions (mostly condensed hydricarbons) from a two-stroke engine are presented and discussed.
The flow field-flow fractionation (FIFFF) technique is a promising method for separating and analysing particles and large size macromolecules from a few nanometers to approximately 50 μm. A new fractionation channel is described featuring well defined flow conditions even for low channel heights with convenient assembling and operations features. The application of the new flow field-flow fractionation channel is proved by the analysis of pigments and other small particles of technical interest in the submicrometer range. The experimental results including multimodal size distributions are presented and discussed.
This paper treats the interaction between acoustic modes and light (Brillouin scattering) in a single mode optical fibre. Different observed spectra of the Brillouin backscattering in several fibres have been already reported. In order to have a clear idea of the process, we made a simulation to be able to `draw' the theoretical Brillouin spectrum of an optical fibre and to identify the origin of the observed backscattered lines.
First, the model and the computation method used in our simulation are described. Second, the experimentally observed spectra of two real fibres are compared with their computed spectra. Real spectra and simulated spectra are in good agreement.
Our work provides an interesting tool to investigate the changes in the Brillouin spectrum when the input parameters (characteristics of an optical fibre) vary. This should give useful indications to people working on systems which use Brillouin backscattering.
Der Beitrag beschreibt wichtige Eckdaten und Ergebnisse der Kraftschlußregelung, die in der Lokomotive 12X auf internationalen Strecken erprobt wurde, und mit der auch zukünftige Projekte ausgestattet werden. Diese werden nicht nur von weiteren technischen Verbesserungen profitieren, sondern auch von geringerem Aufwand für die Inbetriebsetzung.
Im Rahmen des Kontinentalen Tiefbohrprogramms der Bundesrepublik Deutschland (KTB) wurde im September 1987 in der Oberpfalz bei Windischeschenbach mit der Vorbohrung begonnen, die im Frühjahr des letzten Jahres bei einer Tiefe von 4000m erfolgreich abgeschlossen wurde. Mit der in diesem Jahr beginnenden Hauptbohrung will man 10 bis 12km tief in das Erdinnere vorstoßen. Nicht nur anhand von Gesteins- und Flüssigkeitsproben, sondern auch mit Hilfe von Meßsonden werden umfangreiche geophysikalische Daten gewonnen. Der Aufsatz beschreibt, wie die magnetische Suszeptibilität von Gesteinen bis ca. 300°C Umgebungstemperatur und einem Druck von 2kbar gemessen wird und die Daten über ein 14km langes Bohrlochkabel von der Sonde zum Steuerrechner übertragen werden.
Nach einer Darstellung des thermodynamischen Verhaltens von Thermoelementen und dem Vergleich mit realen Thermo-elementen der DIN-Reihe 43710 wird der prinzipielle Aufbau von Meßsystemen mit Thermoelementen und Meßverstärkern behandelt. Ein Verfahren zur Linearisierung der Ausgangskennlinie mit einem mehrstufigen Summenverstärker wird beschrieben. Der Einfluß des Linearisierungsverfahrens auf das Meßergebnis wird allgemein untersucht und anhand eines praxisnahen Beispiels für eine Fe-CuNi-Thermoelement quantitativ berechnet.
Anhand der formelmäßigen Beschreibung des Linearisierungsverfahrens wird der Einfluß der Vergleichsstellen-Temperatur auf das Meßergebnis quantitativ untersucht und an Beispielen erläutert. Daraus werden die Randbedingungen für die Vergleichsstellen-Temperatur hergeleitet und schaltungstechnische Realisierungen für die Kompensation der Temperaturdrift von Vergleichsstellen beschrieben.
Die vorliegende Bachelorarbeit beschreibt die Erstellung einer Anwendung, die es ermöglicht, einen Roboterarm mithilfe eines haptischen Geräts zu steuern. Zudem wird die Kraftrückkopplung des haptischen Geräts, abhängig von der wirkenden Kraft des Roboterarms, implementiert.
Zur Realisierung dieser Anforderungen wurden zwei Programme erstellt; diese beiden Programme sind über eine TCP Verbindung miteinander verbunden. Programm 1 stellt die haptische Umgebung bereit, Programm 2 die Robot Operating System Node, mit welcher die Daten zum Roboter gesendet und Daten vom Roboter empfangen werden können.
Die in dieser Arbeit erstellten Programme wurden mithilfe von Test Nodes überprüft und leiten die Position der haptischen Umgebung weiter. Die Kraftrückkopplung des haptischen Geräts ist ebenfalls vorhanden und integriert.
Ein Testlauf mithilfe eines simulierten Roboterarms konnte erfolgreich durchgeführt werden.
The Baroque composer Johann Sebastian Bach (1685–1750) has left us with many puzzles. The well-known oil painting by Elias Gottlob Haußmann is the only painting for which Bach actually posed in person. According to this portrait, Bach must have been quite obese. The cheeks and nose are flushed – possibly as signs of hypertension – and the eye lids are narrow – a sign of myopia. Furthermore, there is a thinning of the lateral third of the right eyebrow, which is known as Hertoghe’s sign, and indicated periorbital edema. Both signs are compatible with hypothyroidism. Bach might have been suffering from type-2 diabetes as the origin of his final illness, and the obituary reports two cataract surgeries by oculist John Taylor in March/April 1750, and, four months later, “apoplexy” followed by a high fever, of which Bach died. It may be speculated, however, that Bach’s entire illness was the result of his presumed obesity, possibly in combination with hypothyroidism.
Das Team "magmaOffenburg" nimmt seit 2009 an der 3D-Simulationsliga des RoboCups teil. Wie gut das Laufen auf zwei Beinen in der verwendeten Simulationsumgebung "SimSpark" funktioniert ist einer der zentralen Faktoren für ein erfolgreiches Abschneiden. Bisher wird für das Laufen ausschließlich eine inverskinematische Walk-Engine verwendet.
In dieser Arbeit wird mit genetischen Algorithmen und einer modellfreien Parametrisierung Laufen komplett "from scratch" gelernt. Derselbe Ansatz wurde zuvor bereits erfolgreich für das Lernen von Kicks eingesetzt. Gegenüber der alten Walk-Engine zeichnet diesen Ansatz besonders aus, dass zum ersten Mal die Zehengelenke genutzt werden, welche eines der verfügbaren Robotermodelle besitzt. Dies ermöglicht einen natürlicheren und schnelleren Gang. Zwar ist die Stabilität des gelernten Laufens noch nicht vergleichbar mit dem bisherigen (der Spieler fällt etwa 26% häufiger hin), aber beispielsweise beim Vorwärtslaufen konnten mit einer Geschwindigkeit von 1.3 m/s statt 1 m/s Steigerungen von 30% erzielt werden.
Darüber hinaus wurde untersucht, wie dem Agenten im Anschluss an das gelernte Laufen wieder ein Übergang zur alten Walk-Engine gelingen kann. Der erfolgreichste Ansatz basiert ebenfalls wieder auf einer mit genetischen Algorithmen trainierten Bewegung. Er ist in etwa der Hälfte der Fälle erfolgreich.
Auf Grundlage der Computer-Aided-Design (CAD)-rekonstruierten ersten „Eisernen Hand“ des Götz von Berlichingen wird ein umgebautes, controllergesteuertes sensomotorisches Fingersystem auf seine Funktionalität beim Greifen von unterschiedlichen Gegenständen beschrieben und geprüft. Die elektronischen Finger, die den „Pinzettengriff“ nachahmen und automatisch bei dem zuvor eingestellten Anpressdruck abschalten, bewiesen eine bemerkenswerte Alltagstauglichkeit. Das vorgestellte Grundkonzept könnte eine Alternative bei der Entwicklung einfacher und kostengünstiger, aber dennoch gut einsatzfähiger bionischer Hände sein und zeigt einmal mehr, wie historische Ideen in die Gegenwart transferiert werden können.
Diese Bachelorthesis befasst sich mit der Entwicklung eines Auslegungstools für eine PV-basierte, autarke Energieversorgung eines Wasserreinigungssystems mit Hilfe eines Batteriespeichers.
Dafür wurde auf Grundlage theoretischer Einstrahlungsmodelle die Berechnung der Bestrahlungsstärke auf eine geneigte Moduloberfläche in Microsoft Excel 2010 implementiert. Die Berechnung wurde dabei so umgesetzt, dass Parameter wie z.B. Ausrichtung und Größe der PV-Module variiert werden können. Darauf aufbauend wurde die Berechnung der Leistung eines PV-Generators mit Hilfe des Wirkungsgrades der PV-Module umgesetzt.
Des Weiteren wurde für diese Arbeit ein Bilanzierungsmodell entwickelt, welches das Verhalten des Batteriespeichers im autarken System simuliert. Dieses Modell berechnet dabei sowohl die im Batteriespeicher gespeicherte Energie, als auch die über ein Bilanzierungsintervall gemittelte Leistung des Batteriespeichers. Das Batteriemodell wurde mit Hilfe von VBA-Code umgesetzt und in das Auslegungstool eingebunden.
Ein weiteres Thema dieser Bachelorthesis ist die Energieversorgung eines speziellen Wasserreinigungssystems, welches im Rahmen des Forschungsprojekts „PHOTOPUR“ entwickelt wird. Dieses mehrstufige System hat einen charakteristischen Energieverbrauch, der möglichst präzise simuliert werden soll. Um dies zu gewährleisten, wurde ein dynamisches Lastprofil erstellt. Das entwickelte Lastprofil ist spezifisch für das Wasserreinigungssystem, enthält aber gleichzeitig viele Freiheitsgrade um es bei der weiteren Entwicklung des Systems möglichst präzise anpassen zu können.
Das entwickelte Auslegungstool kann folglich die erzeugte PV-Leistung, einen Batteriespeicher und das Lastprofil des Wasserreinigungssystems simulieren. Die Darstellung der Simulationsergebnisse erfolgt mit dem Ziel, Aussagen zu der Qualität der Energieversorgung des Wasserreinigungssystems treffen zu können.
Die Funkschnittstelle nach dem Bluetooth Standard ist heutzutage nicht mehr aus dem Consumer-Markt wegzudenken. Auch durch die zukünftige Weiterentwicklung des Prinzips von IoT (Internet of Things) wird vor allem die Bluetooth Low Energy Technologie für Unternehmen bei der Entwicklung von Geräten mehr an Bedeutung gewinnen. Um die Studierenden im Bereich der Kommunikationstechnik auf das Thema Bluetooth vorzubereiten, wird der Versuch „Kurzstreckenkommunikation nach dem Bluetooth Standard“ im Telekommunikations-Labor an der Hochschule Offenburg angeboten. Der Versuch wurde im Rahmen dieser Bachelor-Thesis an die heutigen Anforderungen angepasst.
Die nachfolgende Bachelor-Thesis beinhaltet die gesamte Laboranleitung des Versuchs, welche in die drei Teile: Theoretische Grundlagen, vorbereitende Aufgaben und der Versuchsdurchführung aufgeteilt wird. Der Versuchsanleitung folgt ein Teil mit weiteren Anmerkungen für die akademischen Mitarbeiter, welche den Versuch betreuen. Hier wird sichergestellt, dass auch in Zukunft der Versuch problemlos weiter durchgeführt werden kann. Im Anhang befinden sich die Musterlösungen für die vorbereitentenden Aufgaben und der Versuchsdurchführung.
Aufbau/Ansteuerung/Kalibrierung/Pilotierung eines Lautsprecher-Halbkreises für Lokalisationstests
(2019)
Seit September 2018 ist die Hochschule Offenburg im Besitz einer Hörkabine, in der ein Schallquellenlokalisationstest im Rahmen dieser Bachelorarbeit aufgebaut wurde. Eine Schallquellenlokalisation ist im Alltag für Normalhörende in den meisten Fällen keine Schwierigkeit. Wir Menschen sind in der Lage mit einer Genauigkeit von 1° bis 2° eine Schallquelle zu lokalisieren (Feigenspan 2017,S.624). Entfällt jedoch das Hörvermögen auf einer Seite, so verschlechtert sich die Lokalisationsfähigkeit erheb-lich. Um Schallquellen präzise lokalisieren zu können, werden optimaler Weise beide Ohren benötigt. Heutzutage gibt es viele Hörsysteme, um diverse Hörstörungen zu therapieren. Den-noch bleibt die Lokalisationsfähigkeit für Menschen mit einseitiger Taubheit, bezie hungsweise asymmetrischen Hörverlust eine Herausforderung. Ein Schallquellenlo kalisationstest findet Anwendung in der Klinik, um eine bestmögliche Versor gung/Therapie durch Hörsysteme zu gewährleisten. Ziel dieser Bachelorarbeit war der Aufbau eines Lokalisationstestes nach Leitlinien einer wissenschaftlichen Publikation, sowie die Durchführung an fünf freiwilligen Normalhörenden. Die Realisierung umfasste den Aufbau der Lautsprecher, die An-steuerung mithilfe einer externen Soundkarte und MATLAB, eine Kalibrierung der Lautsprecher und die Durchführung der Tests.
Hochgenaue Systeme zur Positionsbestimmung spielen im heutigen Zeitalter für den privaten, kommerziellen aber auch militärischen Bereich eine immer wichtigere Rolle. Anwendungsgebiete wie Wandern, Geocaching, Mauterhebungssysteme, autonome Luft-, Land- und Seefahrt, sowie etliche militärische Verwendungen sind nur ein kleiner Überblick dieser Anwendungsmöglichkeiten.
Die durch die Anwendungsgebiete auftretende Notwendigkeit der Satellitennavigation, stellt meine Motivation dar, den Laborversuch Messtechnische Analyse eines Systems zur Satellitennavigation zu erneuern. Dieser Laborversuch soll den Studierenden des Studiengangs Elektrotechnik/Informationstechnik mit dem Schwerpunkt Kommunikationstechnik die Satellitennavigation anhand moderner Laboreinrichtung näherbringen.
Die nachfolgende Arbeit beschäftigt sich zu Beginn mit den allgemeinen theoretischen Grundlagen der Satellitennavigation, gefolgt von systemspezifischen Beschreibungen. Die weiteren Kapitel beinhalten die Versuchsbeschreibung, eine Musterlösung und Informationen für den betreuenden akademischen Mitarbeiter.
Des Weiteren wurden Messreihen aufgezeichnet, welche die einzelnen
Satellitennavigationssysteme qualitativ gegenüberstellen. Auf die Messreihen wird in dieser Thesis nicht eingegangen. Hierfür darf an dieser Stelle auf einen Beitrag in der Hochschulzeitschrift Forschung im Fokus verwiesen werden.
Zum aktuellen Zeitpunkt sind alle Sensoren einer Herz-Lungen-Maschine des Unternehmens Getinge kabelgebunden. Dadurch besteht das Risiko, dass sich während einer OP Keime auf den Kabeln sammeln, die Kabel steigern den Zeitaufwand beim Reinigen und Konnektieren und lassen das Gesamtsystem unübersichtlich wirken.
Das Ziel dieser Arbeit ist es, ein Konzept zu entwickeln, wie die Datenübertragung von einem Sensor zu einer Herz-Lungen-Maschine drahtlos erfolgen kann. Dazu wurden elf gewichtete Anforderungen an einen Funkstandard für eine drahtlose Datenübertragung definiert. Bezüglich dieser Anforderungen wurden 15 Funkstandards anhand einer Literaturrecherche analysiert. Es stellte sich heraus, dass sich ZigBee, IEEE 802.15.6, Z-Wave und EnOcean für diese Anwendung eignen, wobei ZigBee im Vergleich zu den anderen Funkstandards als am besten geeignet bewertet wurde.
Um die Eigenschaften von ZigBee bezüglich der definierten Anforderungen zu verifizieren, wurde ein ZigBee-Prototyp implementiert und getestet. Für die Bereitstellung der Mess-Funktionalität für die Tests wurde ein Programm in C++ geschrieben. Mit Hilfe dieser Tests konnte die Erfüllung der Anforderungen von ZigBee bestätigt werden.
Auf dieser Grundlage ist es möglich, die drahtlose Datenübertragung zwischen Sensor und Herz-Lungen-Maschine auf Basis des ZigBee-Funkstandards zu entwickeln. Ausblickend ist es für die Entwicklung von drahtlosen Sensoren einer Herz-Lungen-Maschine ratsam, die Akzeptanz der Anwender zu erfragen und die regulatorischen Anforderungen für Funk bei Medizinprodukten zu erläutern.
A simple measuring method for acquiring the radiation pattern of an ultrawide band Vivaldi antenna is presented. The measuring is performed by combining two identical Vivaldi antennas and some of the intrinsic properties of a stepped-frequency continue wave radar (SFCW radar) in the
range from 1.0 GHz to 6.0 GHz. A stepper-motor provided the azimuthal rotation for one of the antennas from 0 ◦ to 360 ◦. The tests have been performed within the conventional environment (laboratory / office) without using an anechoic chamber or absorbing materials. Special measuring devices have not been used either. This method has been tested with different pairs of Vivaldi antennas and it can be also used for different ones (with little or no change in the system), as long as their operational
bandwidth is within the frequency range of the SFCW radar.
Keywords — SFCW Radar, Antenna Gain Characterization,
Azimuthal Radiation Pattern
Oesophageal Electrode Probe and Device for Cardiological Treatment and/or Diagnosis (EP3706626A1)
(2020)
The invention relates to an oesophageal electrode probe (10) for bioimpedance measurement and/or for neurostimulation; a device (100) for transoesophageal cardiological treatment and/or cardiological diagnosis; and a method for the open-loop or closed-loop control of a cardiac catheter ablation device and/or a cardiac, circulatory and/or respiratory support device. The oesophageal electrode probe comprises a bioimpedance measuring device for measuring the bioimpedance of at least one part of the tissue surrounding the oesophageal electrode probe. The bioimpedance device comprises at least one first and one second electrode, wherein the at least one first electrode (12A) is arranged on a side (14) of the oesophageal electrode probe facing towards the heart and the at least one second electrode (12B) is arranged on a side (16) of the oesophageal electrode probe facing away from the heart. The device (100) comprises the oesophageal electrode probe (10) and a control and/or evaluation device (30), which is configured for receiving a first bioimpedance measurement signal from the at least one first electrode (12A) and a second bioimpedance measurement signal from the at least one second electrode (12B), and comparing same, and generating a control signal on the basis of the comparison. The control signal can be a signal for the open-loop or closed-loop control of a cardiac catheter ablation device and/or a cardiac, circulatory and/or respiratory support device.
In this paper, we establish a simple model for the exchange of messages in a vehicular network and we consider fundamental limits on the achievable data rate. For a vehicular network, the exchange of data with other nearby vehicles is particularly important for traffic safety, e.g. for collision avoidance, but also for cooperative applications like platooning. These use cases are currently addressed by standards building on IEEE 802.11p, namely ITS-G5 and DSRC (dedicated short range communication), which encounter saturation problems at high vehicle densities. For this reason, we take a step back and ask for the fundamental limits for the common data rate in a vehicular network. After defining a simple single-lane model and the corresponding capacity limits for some basic multiple- access schemes, we present results for a more realistic setting. For both scenarios, non-orthogonal multiple-access (NOMA) yields the best results.
This paper evaluates the implementation of Medium Access Control (MAC) protocols suitable for massive access connectivity in 5G multi-service networks. The access protocol extends multi-packet detection receivers based on Physical Layer Network Coding (PLNC) decoding and Coded Random Access protocols considering practical aspects to implement one-stage MAC protocols for short packet communications in mMTC services. Extensions to enhance data delivery phase in two- stage protocols are also proposed. The assessment of the access protocols is extended under system level simulations where a suitable link to system interface characterization has been taken into account.
This paper is discussing the development of a wireless Indoor Smart Gardening System with the focus on energy autonomous working. The Smart Gardening System, which is presented in this paper consists of a network of energy autonomous wireless sensor nodes which are used for monitoring important plant parameters like air temperature, soil moisture, pressure or humidity and in future to control an actuator for the plant irrigation and to measure further parameter as light and fertilizer level. Solar energy harvesting is used for powering the wireless nodes without the usage of a battery. Comparable Smart Gardening Systems are usually battery-powered. Furthermore, the overall Smart Gardening System consists of a battery powered gateway based on a Raspberry Pi 3 system, which controls the wireless nodes and collects their sensor data. The gateway is able to send the information to an internet server application and via Wi-Fi to mobile devices. Particularly the architecture of the energy autonomous wireless nodes will be considered because fully energy autonomous wireless networks could not be implemented without special concepts for the energy supply and architecture of the wireless nodes.
Nowadays, robotic systems are an integral part of many orthopedic interventions. Stationary robots improve the accuracy but also require adapted surgical workflows. Handheld robotic devices (HHRDs), however, are easily integrated into existing workflows and represent a more economical solution. Their limited range of motion is compensated by the dexterity of the surgeon. This work presents control algorithms for HHRDs with multiple degrees of freedom (DOF). These algorithms protect pre- or intraoperatively defined regions from being penetrated by the end effector (e.g., a burr) by controlling the joints as well as the device’s power. Accuracy tests on a stationary prototype with three DOF show that the presented control algorithms produce results similar to those of stationary robots and much better results than conventional techniques. This work presents novel and innovative algorithms, which work robustly, accurately, and open up new opportunities for orthopedic interventions.
Uncontrollable manufacturing variations in electrical hardware circuits can be exploited as Physical Unclonable Functions (PUFs). Herein, we present a Printed Electronics (PE)-based PUF system architecture. Our proposed Differential Circuit PUF (DiffC-PUF) is a hybrid system, combining silicon-based and PE-based electronic circuits. The novel approach of the DiffC-PUF architecture is to provide a specially designed real hardware system architecture, that enables the automatic readout of interchangeable printed DiffC-PUF core circuits. The silicon-based addressing and evaluation circuit supplies and controls the printed PUF core and ensures seamless integration into silicon-based smart systems. Major objectives of our work are interconnected applications for the Internet of Things (IoT).
Vehicle-to-Everything (V2X) communication promises improvements in road safety and efficiency by enabling low-latency and reliable communication services for vehicles. Besides using Mobile Broadband (MBB), there is a need to develop Ultra Reliable Low Latency Communications (URLLC) applications with cellular networks especially when safety-related driving applications are concerned. Future cellular networks are expected to support novel latencysensitive use cases. Many applications of V2X communication, like collaborative autonomous driving requires very low latency and high reliability in order to support real-time communication between vehicles and other network elements. In this paper, we classify V2X use-cases and their requirements in order to identify cellular network technologies able to support them. The bottleneck problem of the medium access in 4G Long Term Evolution(LTE) networks is random access procedure. It is evaluated through simulations to further detail the future limitations and requirements. Limitations and improvement possibilities for next generation of cellular networks are finally detailed. Moreover, the results presented in this paper provide the limits of different parameter sets with regard to the requirements of V2X-based applications. In doing this, a starting point to migrate to Narrowband IoT (NB-IoT) or 5G - solutions is given.
The next generation cellular networks are expected to improve reliability, energy efficiency, data rate, capacity and latency. Originally, Machine Type Communication (MTC) was designed for low-bandwidth high-latency applications such as, environmental sensing, smart dustbin, etc., but there is additional demand around applications with low latency requirements, like industrial automation, driver-less cars, and so on. Improvements are required in 4G Long Term Evolution (LTE) networks towards the development of next generation cellular networks for providing very low latency and high reliability. To this end, we present an in-depth analysis of parameters that contribute to the latency in 4G networks along with a description of latency reduction techniques. We implement and validate these latency reduction techniques in the open-source network simulator (NS3) for narrowband user equipment category Cat-Ml (LTE-M) to analyze the improvements. The results presented are a step towards enabling narrowband Ultra Reliable Low Latency Communication (URLLC) networks.
The excessive control signaling in Long Term Evolution networks required for dynamic scheduling impedes the deployment of ultra-reliable low latency applications. Semi-persistent scheduling was originally designed for constant bit-rate voice applications, however, very low control overhead makes it a potential latency reduction technique in Long Term Evolution. In this paper, we investigate resource scheduling in narrowband fourth generation Long Term Evolution networks through Network Simulator (NS3) simulations. The current release of NS3 does not include a semi-persistent scheduler for Long Term Evolution module. Therefore, we developed the semi-persistent scheduling feature in NS3 to evaluate and compare the performance in terms of uplink latency. We evaluate dynamic scheduling and semi-persistent scheduling in order to analyze the impact of resource scheduling methods on up-link latency.
The Datagram Transport Layer Security (DTLS) protocol has been designed to provide end-to-end security over unreliable communication links. Where its connection establishment is concerned, DTLS copes with potential loss of protocol messages by implementing its own loss detection and retransmission scheme. However, the default scheme turns out to be suboptimal for links with high transmission error rates and low data rates, such as wireless links in electromagnetically harsh industrial environments. Therefore, in this paper, as a first step we provide an analysis of the standard DTLS handshake's performance under such adverse transmission conditions. Our studies are based on simulations that model message loss as the result of bit transmission errors. We consider several handshake variants, including endpoint authentication via pre-shared keys or certificates. As a second step, we propose and evaluate modifications to the way message loss is dealt with during the handshake, making DTLS deployable in situations which are prohibitive for default DTLS.
The Transport Layer Security (TLS) protocol is a cornerstone of secure network communication, not only for online banking, e-commerce, and social media, but also for industrial communication and cyber-physical systems. Unfortunately, implementing TLS correctly is very challenging, as becomes evident by considering the high frequency of bugfixes filed for many TLS implementations. Given the high significance of TLS, advancing the quality of implementations is a sustained pursuit. We strive to support these efforts by presenting a novel, response-distribution guided fuzzing algorithm for differential testing of black-box TLS implementations. Our algorithm generates highly diverse and mostly-valid TLS stimulation messages, which evoke more behavioral discrepancies in TLS server implementations than other algorithms. We evaluate our algorithm using 37 different TLS implementations and discuss―by means of a case study―how the resulting data allows to assess and improve not only implementations of TLS but also to identify underspecified corner cases. We introduce suspiciousness as a per-implementation metric of anomalous implementation behavior and find that more recent or bug-fixed implementations tend to have a lower suspiciousness score. Our contribution is complementary to existing tools and approaches in the area, and can help reveal implementation flaws and avoid regression. While being presented for TLS, we expect our algorithm's guidance scheme to be applicable and useful also in other contexts. Source code and data is made available for fellow researchers in order to stimulate discussions and invite others to benefit from and advance our work.
Recently, the demand for scalable, efficient and accurate Indoor Positioning Systems (IPS) has seen a rising trend due to their utility in providing Location Based Services (LBS). Visible Light Communication (VLC) based IPS designs, VLC-IPS, leverage Light Emitting Diodes (LEDs) in indoor environments for localization. Among VLC-based designs, Time Difference of Arrival (TDOA) based techniques are shown to provide very low errors in the relative position of receivers. Our considered system consists of five LEDs that act as transmitters and a single receiver (photodiode or image sensor in smart phone) whose position coordinates in an indoor environment are to be determined. As a performance criterion, Cramer Rao Lower Bound (CRLB) is derived for range estimations and the impact of various factors, such as, LED transmission frequency, position of reference LED light, and the number of LED lights, on localization accuracy has been studied. Simulation results show that depending on the optimal values of these factors, location estimation on the order of few centimeters can be realistically achieved.
Real-Time Ethernet has become the major communication technology for modern automation and industrial control systems. On the one hand, this trend increases the need for an automation-friendly security solution, as such networks can no longer be considered sufficiently isolated. On the other hand, it shows that, despite diverging requirements, the domain of Operational Technology (OT) can derive advantage from high-volume technology of the Information Technology (IT) domain. Based on these two sides of the same coin, we study the challenges and prospects of approaches to communication security in real-time Ethernet automation systems. In order to capitalize the expertise aggregated in decades of research and development, we put a special focus on the reuse of well-established security technology from the IT domain. We argue that enhancing such technology to become automation-friendly is likely to result in more robust and secure designs than greenfield designs. Because of its widespread deployment and the (to this date) nonexistence of a consistent security architecture, we use PROFINET as a showcase of our considerations. Security requirements for this technology are defined and different well-known solutions are examined according their suitability for PROFINET. Based on these findings, we elaborate the necessary adaptions for the deployment on PROFINET.
Solar irradiance prediction is vital for the power management and the cost reduction when integrating solar energy. The study is towards a ground image based solar irradiance prediction which is highly dependent on the cloud coverage. The sky images are collected by using ground based sky imager (fisheye lens). In this work, different algorithms for cloud detection being a preparation step for their segmentation are compared.
The fisheye camera has been widely studied in the field of ground based sky imagery and robot vision since it can capture a wide view of the scene at one time. However, serious image distortion is a major drawback hindering its wider use. To remedy this, this paperproposes a lens calibration and distortion correction method for detecting clouds and forecasting solar radiation. Finally, the radial distortion of the fisheye image can be corrected by incorporating the estimated calibration parameters. Experimental results validate the effectiveness of the proposed method.
This paper deals with the detection and segmentation of clouds on high-dynamic-range (HDR) images of the sky as well as the calculation of the position of the sun at any time of the year. In order to predict the movement of clouds and the radiation of the sun for a short period of time, the clouds thickness and position have to be known as precisely as possible. Consequently, the segmentation algorithm has to provide satisfactory results regardless of different weather, illumination and climatic conditions. The principle of the segmentation is based on the classification of each pixel as a cloud or as a sky. This classification is usually based on threshold methods, since these are relatively fast to implement and show a low computational burden. In order to predict if and when the sun will be covered by clouds, the position of the sun on the images has to be determined. For this purpose, the zenith and azimuth angles of the sun are determined and converted into XY coordinates.
A printed electronics technology has the advantage of additive and extremely low-cost fabrication compared with the conventional silicon technology. Specifically, printed electrolyte-gated field-effect transistors (EGFETs) are attractive for low-cost applications in the Internet-of-Things domain as they can operate at low supply voltages. In this paper, we propose an empirical dc model for EGFETs, which can describe the behavior of the EGFETs smoothly and accurately over all regimes. The proposed model, built by extending the Enz-Krummenacher-Vittoz model, can also be used to model process variations, which was not possible previously due to fixed parameters for near threshold regime. It offers a single model for all the operating regions of the transistors with only one equation for the drain current. Additionally, it models the transistors with a less number of parameters but higher accuracy compared with existing techniques. Measurement results from several fabricated EGFETs confirm that the proposed model can predict the I-V more accurately compared with the state-of-the-art models in all operating regions. Additionally, the measurements on the frequency of a fabricated ring oscillator are only 4.7% different from the simulation results based on the proposed model using values for the switching capacitances extracted from measurement data, which shows more than 2× improvement compared with the state-of-the-art model.
Oxide semiconductors are highly promising candidates for the most awaited, next-generation electronics, namely, printed electronics. As a fabrication route for the solution-processed/printed oxide semiconductors, photonic curing is becoming increasingly popular, as compared to the conventional thermal curing method; the former offers numerous advantages over the latter, such as low process temperatures and short exposure time and thereby, high throughput compatibility. Here, using dissimilar photonic curing concepts (UV–visible light and UV-laser), we demonstrate facile fabrication of high performance In2O3 field-effect transistors (FETs). Beside the processing related issues (temperature, time etc.), the other known limitation of oxide electronics is the lack of high performance p-type semiconductors, which can be bypassed using unipolar logics from high mobility n-type semiconductors alone. Interestingly, here we have found that our chosen distinct photonic curing methods can offer a large variation in threshold voltage, when they are fabricated from the same precursor ink. Consequently, both depletion and enhancement-mode devices have been achieved which can be used as the pull-up and pull-down transistors in unipolar inverters. The present device fabrication recipe demonstrates fast processing of low operation voltage, high performance FETs with large threshold voltage tunability.
Printed Electronics (PE) is a promising technology that provides mechanical flexibility and low-cost fabrication. These features make PE the key enabler for emerging applications, such as smart sensors, wearables, and Internet of Things (IoTs). Since these applications need secure communication and/or authentication, it is vital to utilize security primitives for cryptographic key and identification. Physical Unclonable Functions (PUF) have been adopted widely to provide the secure keys. In this work, we present a weak PUF based on Electrolyte-gated FETs using inorganic inkjet printed electronics. A comprehensive analysis framework including Monte Carlo simulations based on real device measurements is developed to evaluate the proposed PE-PUF. Moreover, a multi-bit PE-PUF design is proposed to optimize area usage. The analysis results show that the PE-PUF has ideal uniqueness, good reliability, and can operates at low voltage which is critical for low-power PE applications. In addition, the proposed multi-bit PE-PUF reduces the area usage around 30%.
Printed electronics offers certain technological advantages over its silicon based counterparts, such as mechanical flexibility, low process temperatures, maskless and additive manufacturing process, leading to extremely low cost manufacturing. However, to be exploited in applications such as smart sensors, Internet of Things and wearables, it is essential that the printed devices operate at low supply voltages. Electrolyte gated field effect transistors (EGFETs) using solution-processed inorganic materials which are fully printed using inkjet printers at low temperatures are very promising candidates to provide such solutions. In this paper, we discuss the technology, process, modeling, fabrication, and design aspect of circuits based on EGFETs. We show how the measurements performed in the lab can accurately be modeled in order to be integrated in the design automation tool flow in the form of a Process Design Kit (PDK). We also review some of the remaining challenges in this technology and discuss our future directions to address them.
Data Science gilt als eine der wichtigsten Entwicklungen der letzten
Jahre und viele Unternehmen sehen in Data Science die Möglichkeit,
ihre Daten zusätzlich wertschöpfend zu nutzen. Dabei kann es sich um
die Optimierung von Maintenance-Prozessen handeln, um eine bessere
Steuerung der eigenen Preis- und Lagerhaltungsstrategie oder auch
um völlig neue Services und Produkte, die durch Data Science möglich
werden. Die im Unternehmen vorliegenden Daten, an die so hohe Erwartungen
geknüpft wurden, sollen dazu genutzt werden, um Services
und Prozesse effizienter und passgenauer gestalten zu können. Vielfach
gilt Data Science dabei als Allheilmittel: Daten, die über Jahre hinweg
gesammelt wurden und mit zunehmender Geschwindigkeit und Heterogenität
anfallen, sollen endlich nutzbar gemacht werden. Zwar sind die
eingesetzten Techniken und Algorithmen teilweise schon zehn Jahre und
mehr alt, doch erst jetzt entfalten sie im Zusammenspiel mit Big Data
ihr Potenzial im Unternehmensumfeld. Die Erwartungen sind hoch, doch
der Weg zu den neuen Erkenntnissen ist mit hohem Aufwand verbunden
und wird von einigen Unternehmen noch immer unterschätzt.
Für Unternehmen mit einem traditionellen BI-Ansatz stellt Data Science
ein ergänzendes Set von Methoden und Werkzeugen dar, mit deren Hilfe
die Informationsversorgung der Entscheider auf den verschiedenen
hierarchischen Ebenen noch besser gestaltet werden kann. So zum Beispiel,
wenn man mit Data Science feststellt, dass die Wahrscheinlichkeit
für einen Versicherungsabschluss steigt, wenn bei der Auswahl der
anzusprechenden Kunden zusätzliche Daten herangezogen werden, die
zwar bereits vorliegen, aber noch nicht berücksichtigt worden sind. Im
Extremfall werden auch Entscheidungen vollständig automatisiert, die
bisher von Mitarbeiterinnen und Mitarbeitern getroffen wurden. Ein Algorithmus
legt dann fest, wann Ware nachbestellt oder welcher Preis für
den Endkunden festgesetzt wird.
Im vorliegenden E-Book soll ein Überblick über das Gebiet Data Science
gegeben werden. Dabei wird ein besonderes Augenmerk auf das Zusammenspiel
sowie das Mit- und Nebeneinander von Data Science und vorhandenen
BI-Systemen gelegt.
Targeting complex fractionated atrial electrocardiograms by automated algorithms during ablation of persistent atrial fibrillation has produced conflicting outcomes in previous electrophysiological studies and catheter ablation of atrial fibrillation and ventricular tachycardia. The aim of the investigation was to evaluate atrial and ventricular high frequency fractionated electrical signals with signal averaging technique.
Methods: Signal averaging electrocardigraphy allows high resolution ECG technique to eliminate interference noise signals in the recorded ECG. The algorithm use automatic ECG trigger function for signal averaged transthoracic, transesophageal and intra-cardiac ECG signals with novel LabVIEW software.
Results: The analysis in the time domain evaluated fractionated atrial signals at the end of the signal averaged P-wave and fractionated ventricular signals at the end of the QRS complex. We evaluated atrial flutter in the time domain with two-to-one atrioventricular conduction, 212.0 ± 4.1 ms atrial cycle length, 426.0 ± 8.2 ms ventricular cycle length, 58.2 ± 1.8 ms P-wave duration, 119.6 ± 6.4 ms PQ duration, 103.0 ± 2.4 ms QRS duration and 296.4 ± 6.8 ms QT duration. The analysis in the frequency domain evaluated high frequency fractionated atrial signals during the P-wave and high frequency fractionated ventricular signals during QRS complex.
Conclusions: Spectral analysis of signal averaging electrocardiography with novel LabVIEW software can be utilized to evaluate atrial and ventricular conduction delays in patients with atrial fibrillation and ventricular tachycardia. Complex fractionated atrial and ventricular electrocardiograms may be useful parameters to evaluate electrical cardiac bradycardia and tachycardia signals in atrial fibrillation and ventricular tachycardia ablation.
Cardiac resynchronization therapy (CRT) with hemodynamic optimized biventricular pacing is an established therapy for heart failure patients with sinus rhythm, reduced left ventricular ejection fraction and wide QRS complex. The aim of the study was to evaluate electrical right and left cardiac atrioventricular delay and left atrial delay in CRT responder and non-responder with sinus rhythm.
Methods: Heart failure patients with New York Heart Association class 3.0 ± 0.3, sinus rhythm and 27.7 ± 6.1% left ventricular ejection fraction were measured by surface ECG and transesophageal bipolar left atrial and left ventricular ECG before implantation of CRT devices. Electrical right cardiac atrioventricular delay was measured between onset of P wave and onset of QRS complex in the surface ECG, left cardiac atrioventricular delay between onset of left atrial signal and onset of left ventricular signal in the transesophageal ECG and left atrial delay between onset and offset of left atrial signal in the transesophageal ECG.
Results: Electrical atrioventricular and left atrial delay were 196.9 ± 38.7 ms right and 194.5 ± 44.9 ms left cardiac atrioventricular delay, and 47.7 ± 13.9 ms left atrial delay. There were positive correlation between right and left cardiac atrioventricular delay (r = 0.803 P < 0.001) and negative correlation between left atrial delay and left ventricular ejection fraction (r = −0.694 P = 0.026) with 67% CRT responder.
Conclusions: Transesophageal electrical left cardiac atrioventricular delay and left atrial delay may be useful preoperative atrial desynchronization parameters to improve CRT optimization.
Diese Arbeit befasst sich mit der Konzeption und Implementierung eines Dateiimportfilters für Mess- und Kalibrierwerkzeuge. Dieser ermöglicht das Importieren von INCA-Experimenten nach CANape, indem aus Experiment-Exportdateien Konfigurationsdateien generiert werden. Im Zuge der Umsetzung wurde dazu das aus binären Daten bestehende Exportdateiformat von INCA analysiert. Das Ziel dieser Analyse bestand darin, herauszufinden, wo und wie die Informationen in den Exportdateien gespeichert sind. Die so erlangten Erkenntnisse wurden verwendet, um den Dateiimportfilter zu entwickeln. Dieser liest aus Exportdateien die notwendigen Informationen aus, überführt sie in eine CANape interne Repräsentation und generiert daraus eine entsprechende Konfigurationsdatei. Die korrekte Funktionsweise wurde anhand von selbst erstellten Experimenten sowie mithilfe von Kundenexperimenten validiert.
Wireless sensor networks have recently found their way into a wide range of applications among which environmental monitoring system has attracted increasing interests of researchers. Such monitoring applications, in general, don way into a wide range of applications among which environmental monitoring system has attracted increasing interests of researc latency requirements regarding to the energy efficiency. Also a challenge of this application is the network topology as the application should be able to be deployed in very large scale. Nevertheless low power consumption of the devices making up the network must be on focus in order to maximize the lifetime of the whole system. These devices are usually battery-powered and spend most of their energy budget on radio transceiver module. A so-called Wake-On-Radio (WoR) technology can be used to achieve a reasonable balance among power consumption, range, complexity and response time. In this paper, some designs for integration of WOR into IEEE 802.1.5.4 are to be discussed, providing an overview of trade-offs in energy consumption while deploying the WoR schemes in a monitoring system.
Energy consumption for cooling is growing dramatically. In the last years, electricity peak consumption grew significantly, switching from winter to summer in many EU countries. This is endangering the stability of electricity grids. This article outlines a comprehensive analysis of an office building performances in terms of energy consumption and thermal comfort (in accordance with static – ISO 7730:2005 – and adaptive thermal comfort criteria – EN 15251:2007 –) related to different cooling concepts in six different European climate zones. The work is based on a series of dynamic simulations carried out in the Trnsys 17 environment for a typical office building. The simulation study was accomplished for five cooling technologies: natural ventilation (NV), mechanical night ventilation (MV), fan-coils (FC), suspended ceiling panels (SCP), and concrete core conditioning (CCC) applied in Stockholm, Hamburg, Stuttgart, Milan, Rome, and Palermo. Under this premise, the authors propose a methodology for the evaluation of the cooling concepts taking into account both, thermal comfort and energy consumption.
Chronic insomnia is defined by difficulties in falling asleep, maintaining sleep, and early morning awakening, and is coupled with daytime consequences such as fatigue, attention deficits, and mood instability. These symptoms persist over a period of at least 3 months (Diagnostic and Statistical Manual 5 criteria). Chronic insomnia can be a symptom of many medical, neurological, and mental disorders. As a disorder, it incurs substantial health-care and occupational costs, and poses substantial risks for the development of cardiovascular and mental disorders, including cognitive deficits. Family and twin studies confirm that chronic insomnia can have a genetic component (heritability coefficients between 42% and 57%), whereas the investigation of autonomous and central nervous system parameters has identified hyperarousal as a final common pathway of the pathophysiology, implicating an imbalance of sleep–wake regulation consisting of either overactivity of the arousal systems, hypoactivity of the sleep-inducing systems, or both. Insomnia treatments include benzodiazepines, benzodiazepine-receptor agonists, and cognitive behavioural therapy. Treatments currently under investigation include transcranial magnetic or electrical brain stimulation, and novel methods to deliver psychological interventions.
The lifetime and performance of solid-oxide fuel cells (SOFC) and electrolyzer cells (SOEC) can be significantly degraded by oxidation of nickel within the electrode and support structures. This paper documents a detailed computational model describing nickel oxide (NiO) formation as a growing film layer on top of the nickel phase in Ni/YSZ composite electrodes. The model assumes that the oxidation rate is controlled by transport of ions across the film (Wagner's theory). The computational model, which is implemented in a two-dimensional continuum framework, facilitates the investigation of alternative chemical reaction and transport mechanisms. Model predictions agree well with a literature experimental measurement of oxidation-layer growth. In addition to providing insight in interpreting experimental observations, the model provides a quantitative predictive capability for improving electrode design and controlling operating conditions.
Im Rahmen dieser Arbeit wird ein dezentrales Usermanagement entwickelt, das ein vorhandenes aber veraltetes Usermanagement ersetzen soll.
Zuerst werden verschiedene Technologien für die Umsetzung evaluiert. Dabei wird begründet, warum z. B. dem Relationalen Datenbanksystem von Oracle der Vorzug gegenüber einem NoSQL-System wie MongoDB gegeben wird.
Im Rahmen einer ausführlichen Konzeption werden zuerst die Anforderungen detailliert formu-liert. In einem Grob-Entwurf werden die Beziehungen des alten und des neuen Usermanagements zur Außenwelt dargestellt. Datenmodellierung und Skizzierung der Paket- bzw. Klassen-Struktur innerhalb der Anwendung bilden die Grundlage für die spätere Implementierung. Ein besonderer Schwerpunkt bei der Konzeptionierung liegt auf dem Entwurf der REST-API.
Bei der Implementierung werden die Schritte von der Einrichtung des Spring-Frameworks bis zur lauffähigen Anwendung beschrieben. Der Umgang mit der Datenbank und dem OR-Mapper sind Schwerpunkte der Implementierung, ebenso wie die Umsetzung der REST-Schnittstellen. Die fachlichen Prozesse der Daten-Migration, der Verarbeitung von Requests aus einem zentralen System und des GLN-Mappings werden in eigenen Kapiteln behandelt.
Am Ende werden die erzielten Ergebnisse dokumentiert, offene Punkte angesprochen und ein Fazit gezogen.
Verschiedenste Hersteller von Prothesen treiben die Entwicklungen in der Neuroprothetik immer weiter voran. Jedoch steigen dadurch nicht nur die Komplexität und die Funktionen einer solchen Prothese, sondern auch die Kosten. Oft wird vernachlässigt, dass ein einfaches Greifen meist schon ausreicht und damit oftmals viel mehr Personen geholfen werden kann als mit einer teuren Highend-Prothese.
Die vorliegende Masterthesis soll zeigen, dass es möglich ist, mit einfachen Mitteln einen funktionsfähigen bionischen Prototyp zu entwickeln. Die Steuerung funktioniert per Knopfdruck, Bewegungen werden automatisch ausgeführt.
Hierfür wurde auf der Rekonstruktion der ersten eisernen Hand des Götz von Berlichin-gen aufgebaut. Diese wurde mit günstiger und einfach beschaffbarer Elektronik verändert, damit die Prothese durch elektrische Motoren aktiv ansteuerbar ist. Das Modell wurde mit SolidWorks 2018 verändert. Die elektronischen Bauteile wurden über ein Arduino Board UNO R3 angesteuert, welcher die Schnittstelle zum Computer bildet.
In der vorliegenden Bachelorarbeit geht es um den Entwurf eines Dauerfestigkeitsprüfstands für Kraftsensoren, deren Messprinzip auf Dehnmessstreifen basiert. Ziel des Dauerfestigkeitsprüfstands ist es, die in der Konstruktionsphase simulierten Ergebnisse für die Dauerfestigkeit der Sensoren nachzuweisen. Erst dadurch sind optimale Verbesserungen der Messkörpergeometrien möglich.
The Raman spectra from the chemical compounds toluene and cyclohexane obtained using a Fourier Transform (FT)-Raman spectrometer prototype have been contrasted with the Raman spectra of these same materials collected with two different commercial FT-Raman devices. The FT-Raman spectrometer consist of a Michelson interferometer, a self-designed photon counter and a reference photo-detector. The evaluation methodology of the spectral information, contrary to the commercial devices that commonly use the zero-crossing method, is carried out by re-sampling the Raman scattering and by accurately extracting the optical path information of the Michelson interferometer. The FTRaman arrangement has been built using conventional parts without disregarding the spectral frequency precision that usually such a FTRaman instruments deliver. No additional complex hardware components or costly software modules have been included in this FT-Raman device. The main Raman lines from the spectra obtained with the three FT-Raman devices have been compared with the Raman lines from the standard Raman spectra of these two materials. The values obtained using the FT-Raman spectrometer prototype have shown a frequency accuracy comparable to that obtained with the commercial devices without facing the need for a large investment. Although the proposed FT-Raman prototype cannot be directly compared to the last generation of FT-Raman spectrometers from the commercial manufacturers, such a device could give an opportunity to users that require high frequency precision in their spectral analysis and are provided with rather scarce resources.
We report the use of the Raman spectral information of the chemical compound toluene C7H8 as a reference on the analysis of laboratory-prepared and commercially acquired gasoline-ethanol blends. The rate behavior of the characteristic Raman lines of toluene and gasoline has enabled the approximated quantification of this additive in commercial gasoline-ethanol mixtures. This rate behavior has been obtained from the Raman spectra of gasoline-ethanol blends with different proportions of toluene.
All these Raman spectra have been collected by using a self-designed, frequency precise and low-cost Fourier-transform Raman spectrometer (FT-Raman spectrometer) prototype. This FT-Raman prototype has helped to accurately confirm the frequency position of the main characteristic Raman lines of toluene present on the different gasoline-ethanol samples analyzed at smaller proportions than those commonly found in commercial gasoline-ethanol blends. The frequency accuracy validation has been performed by analyzing the same set of toluene samples with two additional state-of-the-art commercial FT-Raman devices. Additionally, the spectral information has been contrasted, with highly-correlated coefficients as a result, with the values of the standard Raman spectrum of toluene.
Energy management in distribution grids is one of the key challenges that needs to be overcome to increase the share of fluctuating renewable energies. Current control systems for energy management mainly demonstrate centralized- or decentralized-hierarchical control structures. Very few systems manifest a fully decentralized multiagent-based control structure. Multiagent-based control systems promise to be an advantageous approach for the future distributed energy supply system because no central control entity is necessary, which eases parameterization in case of grid topology changes, and the agents are more stable against failures and changes of control topologies. Research is necessary to prove these benefits. In this study, we introduce a design of a multiagent-based voltage control system for low-voltage grids. In detail we introduce cooperative decision-making processes and software solutions that allow the agents to perceive and control their environment, the agent-discovery and localization in different types of communication networks, agent-to-agent communication, and the integration of the multiagent system in existing grid-control infrastructures. Furthermore, the study proposes how different existing technologies can be combined into an applicable multiagent-based voltage control system: the Java/OSGi-based OpenMUC framework allows a generic field–device interaction; peer-to-peer discovery and session establishment functionalities are combined with the agent communication defined by the Foundation for Intelligent Physical Agents (FIPA). The ripple control-signal technology is applied as a fallback communication between the agent and a central grid-control center.
Ziel der vorliegenden Arbeit ist es, einen Fräsroboter zu konstruieren, der kompakter und leichter sein soll als ein bereits im Labor CompAssMed existierender Fräsroboter. Für den neuen Fräsroboter sollte zusätzlich eine Ansteuerung programmiert werden. Mit der Ansteuerung soll es möglich sein, die Fräserspitze im Arbeitsraum des Fräsroboters programmgesteuert zu positionieren. Eine Augmented-Reality-Brille soll evaluiert werden, um ihre Integrationsfähigkeit in das Labor zu untersuchen. Alle Ziele wurden erreicht. Das Gewicht des neuen Fräsroboters ist 75 % geringer und er hat eine 85 % kleinere Grundfläche als der vorhandene Fräsroboter. Die Ansteuerung des neuen Fräsroboters, wurde als MATLAB-Skript realisiert und ist funktional. Die Evaluierung der Augmented-Reality-Brille wurde mit dem Ergebnis durchgeführt, dass sie nicht in die Softwareinfrastruktur des Labors CompAssMed integriert werden kann.