Refine
Year of publication
- 2021 (13) (remove)
Document Type
- Bachelor Thesis (6)
- Master's Thesis (6)
- Other (1)
Language
- English (13) (remove)
Has Fulltext
- yes (13)
Is part of the Bibliography
- no (13) (remove)
Keywords
- Automatisierungstechnik (1)
- Clustering (1)
- Deep learning (1)
- Direct Aeroacoustic Simulation (1)
- Energiewende (1)
- Energy System Modelling (1)
- Functional safety (1)
- Funktionale Sicherheit (1)
- Funktionssicherheit (1)
- Future Energy System (1)
Institute
Open Access
- Closed Access (12)
- Closed (1)
Threat Modeling is a vital approach to implementing ”Security by Design” because it enables the discovery of vulnerabilities and mitigation of threats during the early stage of the Software Development Life Cycle as opposed to later on when they will be more expensive to fix. This thesis makes a review of the current threat Modeling approaches, methods, and tools. It then creates a meta-model adaptation of a fictitious cloud-based shop application which is tested using STRIDE and PASTA to check for vulnerabilities, weaknesses, and impact risk. The Analysis is done using Microsoft Threat Modeling Tool and IriusRisk. Finally, an evaluation of the results is made to ascertain the effectiveness of the processes involved with highlights of the challenges in threat modeling and recommendations on how security developers can make improvements.
Global energy demand is still on an increase during the last decade, with a lot of impact on the climate change due to the intensive use of conventional fossil-based fuels power plants to cover this demand. Most recently, leaders of the globe met in 2015 to come out with the Paris Agreement, stating that the countries will start to take a more responsible and effective behaviour toward the global warming and climate change issues. Many studies have discussed how the future energy system will look like with respecting the countries’ targets and limits of greenhouse gases and their CO2 emissions. However, these studies rarely discussed the industry sector in detail even though it is one of the major role players in the energy sector. Moreover, many studies have simulated and modelled the energy system with huge jumps of intervals in terms of years and environmental goals. In the first part of this study, a model will be developed for the German electrical grid with high spatial and temporal resolutions and different scenarios of it will be analysed meticulously on shorter periods (annual optimization), with different flexibilities and used technologies and degrees of innovations within each scenario. Moreover, the challenge in this research is to adequately map the diverse and different characteristics of the medium-sized industrial sector. In order to be able to take a first step in assessing the relevance of the industrial sector in Germany for climate protection goals, the industrial sector will be mapped in PyPSA-Eur (an open-source model data set of the European energy system at the level of the transmission network) by detailing the demand for different types of industry and assigning flexibilities to the industrial types. Synthetically generated load profiles of various industrial types are available. Flexibilities in the industrial sector are described by the project partner Fraunhofer IPA in the GaIN project and can be used. Using a scenario analysis, the development of the industrial sector and the use of flexibilities are then to be assessed quantitatively.
The Lattice Boltzmann Method is a useful tool to calculate fluid flow and acoustic effects at the same time. Although the acoustic perturbation is much smaller than normal pressure differences in fluid flow, this direct calculation is a great advantage of the Lattice Boltzmann Method (LBM). But each border used in calculation produces a multitude of reflections with the acoustic waves, which lead to an unusable result. Therefore, it is worked on different absorbing techniques.
In this thesis three absorbing layer techniques are described, explained and reviewed with different simulations. The absorbing layers are implemented in a basic LBM code in C++, and with this umpteen simulations within a box were performed to compare the different absorbing layers. The Doppler effect and a cylinder flow are also examined to compare the damping efficiencies.
The three studied absorbing techniques are the sponge layer, the perfectly matched layer and a force based Term II absorbing layer. The sponge layer is easy to implement but gives worse results than a calculation without any absorbing layer. The perfectly matched layer and a force based absorbing term provide very good results but the perfectly matched layer has problems with instability. The force based absorbing layer represents the best compromise between the additional computation time due the absorbing layer and the achieved damping efficiency.
The identification of vulnerabilities is an important element of the software development process to ensure the security of software. Vulnerability identification based on the source code is a well studied field. To find vulnerabilities on the basis of a binary executable without the corresponding source code is more challenging. Recent research has shown how such detection can be performed statically and thus runtime efficiently by using deep learning methods for certain types of vulnerabilities.
This thesis aims to examine to what extent this identification can be applied sufficiently for a variety of vulnerabilities. Therefore, a supervised deep learning approach using recurrent neural networks for the application of vulnerability detection based on binary executables is used. For this purpose, a dataset with 50,651 samples of 23 different vulnerabilities in the form of a standardised LLVM Intermediate Representation was prepared. The vectorised features of a Word2Vec model were then used to train different variations of three basic architectures of recurrent neural networks (GRU, LSTM, SRNN). For this purpose, a binary classification was trained for the presence of an arbitrary vulnerability, and a multi-class model was trained for the identification of the exact vulnerability, which achieved an out-of-sample accuracy of 88% and 77%, respectively. Differences in the detection of different vulnerabilities were also observed, with non-vulnerable samples being detected with a particularly high precision of over 98%. Thus, the methodology presented allows an accurate detection of vulnerabilities, as well as a strong limitation of the analysis scope for further analysis steps.
The Projekt-Sweaty is a project of the University of Applied Sciences of Offenburg, an autonomous robot is being developed that competes against a set of several international colleges and universities in the RoboCup.
"Sweaty" is a soccer-playing humanoid robot who participated in the RoboCup World Cup in Brazil for the first time in 2014.
RoboCup is a competition aimed at developing a robot soccer team that surpasses the human world champion team. The competition started in 1997 the first official RoboCup games and conferences were held with great success. More than 40 teams took part and more than 5,000 spectators attended. RoboCup’s rules change to promote advances in robot science and technology and to bring the league’s challenges closer to the real world.
Building a robot that plays football will not in itself generate a significant social and economic impact, but the realization will certainly be considered an important success for the field of robotics.
Thanks to the interaction of all the faculties, the team consists of professors and students from the fields of mechanical and process engineering, electrical engineering, information technology, and information and media technology. Students can use the project during their studies and use the knowledge acquired in practice to implement and through their own creative ideas complement.
The Project "Schluckspecht" of the University of Offenburg consists of participating in the European marathon called "Shell Eco-Marathon"(SEM) which consists of designing and building from the beginning a vehicle with the greatest possible energy efficiency. The University of Offenburg has participated in this project since 1998.
The team that forms the Schluckspecht project is made up of around 30 students from the faculties of mechanical engineering, process engineering, electrical engineering, medical technology and computer science, as well as the degree in Audiovisual Communication. The team was founded in 1998 and since then students have been developing and building high efficiency vehicles to participate in the European marathon Shell Eco.
In this project, students can put into practice all the theoretical knowledge obtained during their studies. Also can be learned how to work interdisciplinarity as a team, a skill that for now, many companies or require or seek.
The following topics are discussed in the Schluckspecht project, which are also ideal for the work of students:
-Conception construction and production of high efficiency vehicles.
-Computational design and manufacture of lightweight components and sets.
-Development of lightweight components and sets from renewable raw materials.
-Construction and development of special test benches, for example: motor test bench.
-Implementation and optimization of control strategies for autonomous driving
-Mechanical and electrical integration of sensors for autonomous driving
-Ergonomic studies and optimization of the driver's cabin.
The objective of the project is to develop and manufacture research vehicles that make individual mobility as efficient as possible from an energy point of view. To achieve this, current and future issues of the industry are discussed. In this project, both the theoretical and practical part of the light construction of vehicles and the reduction of friction, the variety of propulsion concepts (electric thrusters, fuel cells, diesel/petrol engines, Stirling engines) and autonomous driving are investigated. The services of the University of Offenburg together with some external partners are grouped together to make this wonderful project work.
The Timed-Up-and-Go (TUG) test aims to assess mobility, balance, walking ability, and fall risk during walking. The instrumentalization of the TUG is already described in the literature and is beginning to be implemented in the industry. The products proposed by Zhortech and Digitsole, namely connected insoles, as well as additional sensors placed on the sternum and the right and eventually left femur allow the instrumentalization of the test.
An algorithm of detection and evaluation of the TUG has been developed in two versions. The first one (V1) aiming simply to calculate the total duration of the test. A second version is an improvement of V1, allowing to segment the TUG in three sub-phases: Sit-Stand, walking, Stand-Sit. These algorithms have been declined in a variant with the five sensors mentioned, and one without the sensor of the left femur.
The performance of the algorithms was compared to manual labeling performed on video. The comparison includes a bland-Altman plot and a correlation for the total test duration, but also for the sub-phase’s duration according to the two variants.
The TUG duration shows very good results regarding the limits of agreements (lLoA = -0.33 s and uLoA+0.6 s). The bias of 0.13 s indicated that the algorithm overrates the duration of the TUG. The results of the TUG subphases are less accurate. Although the correlation coefficient is between 0.76 and 0.96 for the different subphases, the limits of agreements are still very high, between -0.71 s and -0.5 s for the lLoA and +0.39 s and +0.58 s for the uLoA. These limits of agreements indicated that the Sit-Stand and Stand-Sit transition are not accurate enough yet. The dispersion is high for a transition that could last between about one and six seconds. The two variants, with and without a sensor on the left femur, present similar results.
In this work, an implementation of the somewhat homomorphic BV encryption scheme is presented. During the implementation, care was taken to ensure that the resulting program will be as efficient as possible i.e. fast and resource-saving. The basis for this is the work of Arndt Bieberstein, who implemented the BV scheme with respect to functionality. The presented implementation supports the basics of the BV scheme, namely (symmetric and asymmetric) encryption, decryption and evaluation of addition as well as multiplication. Additionally, it supports the encoding of positive and negative numbers, various gaussian sampling methods, basically infinitely large polynomial coefficients, the generation of suitable parameters for a use case, threading and relinearization to reduce the size of a ciphertext after multiplications. After presenting the techniques used in the implementation, it’s actual efficiency is determined by measuring the timings of the operations for various parameters.
Among the billions of smartphone users in the world, Android still holds more than 80% of the market share. The applications which the users install have a specific set of features that need access to some device functionalities and sensors that may hold sensitive information about the user. Therefore, Android releases have set permission standards to let the user know what information is being disclosed to the application. Along with other security and privacy improvements, significant changes to the permission scheme are introduced with the Android 6.0 version (API level 23). In this master thesis, the Android permission scheme is tested on two devices from different eras. The evolution of Android over the years is examined in terms of confidentiality. For each device, two applications are built; one focused on extracting every piece of information within the confidentiality scope with every permission declared and/or requested, and the other app focused on getting this type of information without user notification. The resulting analysis illustrates whether how and in what way the Android permission scheme declined or improved over time.