Refine
Year of publication
Document Type
- Master's Thesis (61) (remove)
Language
- English (61) (remove)
Has Fulltext
- yes (61)
Is part of the Bibliography
- no (61)
Keywords
- IT-Sicherheit (6)
- Maschinelles Lernen (4)
- Deep learning (3)
- security (3)
- Cloud Computing (2)
- Computersicherheit (2)
- Energiemanagement (2)
- Energiewende (2)
- Homomorphic Encryption (2)
- Identitätsverwaltung (2)
- Internet der Dinge (2)
- JavaScript (2)
- Künstliche Intelligenz (2)
- Landwirtschaft (2)
- Machine Learning (2)
- Organischer Abfall (2)
- Pflanzenkohle (2)
- Risikoanalyse (2)
- Risk Analysis (2)
- Security (2)
- biowaste (2)
- optimization (2)
- 3D printing (1)
- 3D-Druck (1)
- 4-Oxocrotonic acid (1)
- 5G (1)
- 6LoWPAN (1)
- AI (1)
- Access Management (1)
- Accounting (1)
- Adversarial Attacks (1)
- Agent <Künstliche Intelligenz> (1)
- Agent Fitness (1)
- Agrophotovoltaics (1)
- Algorithmus (1)
- Amazon Echo (1)
- Angriff (1)
- Anomalieerkennung (1)
- Anomaly Detection (1)
- Artificial Intelligence (1)
- Automation (1)
- Bellman Ford algorithm (1)
- Biochar (1)
- Biogas (1)
- Biologika (1)
- Biosimilars (1)
- Bodenradar (1)
- Bot Classification (1)
- Buchführung (1)
- Business Intelligence (1)
- CCS (1)
- CNN (1)
- CO2-Bilanz (1)
- Carbon Footprint (1)
- Carbon nanotube membran (1)
- Cellulasen (1)
- Cement Industry (1)
- Chiffrierung (1)
- Client-side (1)
- Convolutional neural networks (1)
- Cyber Maturity Assessment Model (1)
- CyberDeception (1)
- Cybersecurity (1)
- Cytochrome (1)
- DTLS (1)
- DVRIP (1)
- Data Science (1)
- Data analysis (1)
- Datenanalyse (1)
- Decarbonization (1)
- Decentralized applications (1)
- Dekarbonisierung (1)
- Digital Twins (1)
- Digitaler Zwilling (1)
- Direct Aeroacoustic Simulation (1)
- Distributed Sensor Network (1)
- Distributed computing (1)
- E-Commerce (1)
- ETL Data Pipeline (1)
- EU 933/2019 (1)
- Edge Computing (1)
- Effizienz (1)
- Eingruppierung (1)
- Eisen- und Stahlindustrie (1)
- Electric Tractors (1)
- Electrochemical membrane (1)
- Electrochemical membrane reactor (1)
- Electronic Commerce (1)
- Elektrofahrzeug (1)
- Elektromagnetische Reflexionsmethode (1)
- Energie (1)
- Energiespeicherung (1)
- Energy System Analysis (1)
- Energy System Model (1)
- Energy System Modelling (1)
- Energy Transition (1)
- Energy management (1)
- Enzym (1)
- Enzyme assay (1)
- Express.js (1)
- Fermentation (1)
- Foreign Direct Investment (1)
- Fotovoltaik (1)
- Friedländer-Bakterium (1)
- Future Energy System (1)
- GPR (1)
- Gebäudeleittechnik (1)
- German Electrical Grid (1)
- Glasfaser (1)
- Global Payroll (1)
- Green House effect (1)
- Ground penetrating radar (1)
- High Performance Computing (HPC) (1)
- High content measurement (1)
- Hochleistungsrechnen , Informatik (1)
- Homomorphe Kryptosysteme (1)
- Homomorphe Verschlüsselung (1)
- Homomorphic Encryption Schemes (1)
- Human Firewewall (1)
- Hybrid approach (1)
- Hydrokultur (1)
- Hydroponic Farming (1)
- Hydroxylierung (1)
- IPv6 (1)
- ISMS (1)
- Identity Management (1)
- Industrial Sector (1)
- Information Security (1)
- Informationssicherheit (1)
- International Business Consulting (1)
- Internet of Things (1)
- IoT (1)
- Javascript (1)
- Klebsiella pneumoniae (1)
- Konfigurationen (1)
- Kosten (1)
- Kostenrechnung (1)
- Kostensenkung (1)
- Kryptologie (1)
- LINUX (1)
- Lattice Boltzmann Method (1)
- Lean Accounting (1)
- Lean Accounting Transformation (1)
- Lean Management (1)
- Leistungsbedarf (1)
- LoRaWAN (1)
- Lohnbuchführung (1)
- Low Carbon network (1)
- Machine Firewall (1)
- Mass spectrometry (1)
- Massenspektrometrie (1)
- Maturity Model (1)
- Mikrofon (1)
- MinION-sequencing (1)
- Mobilfunk (1)
- Multidrug-Resistenz (1)
- NADPH (1)
- Nachhaltigkeit (1)
- Negative Emission Technology (1)
- Netzwerk (1)
- Neuro-Evolution (1)
- Neuronales Netz (1)
- Node.js (1)
- Node.js security (1)
- Non-Reflecting Boundary Conditions (1)
- OSINT (1)
- Open Source (1)
- Open Source Intelligence (1)
- Optical Fiber Network (1)
- Optical Network (1)
- Outsourcing (1)
- Oxford Nanopore Technology (1)
- Oxyfuel (1)
- P450 Cytochrome (1)
- PROFINET (1)
- Peer-to-peer trading (1)
- Phagetherapy (1)
- Phising detection (1)
- Power Consumption (1)
- Power Quality (1)
- Privacy and Security (1)
- Proteine (1)
- Python Programming (1)
- QCM-D Array (1)
- Ransomware (1)
- Rechnungswesen (1)
- Risikomanagement (1)
- Risk Management (1)
- SOC (1)
- SPC waiver (1)
- Scientific workflow (1)
- Security Operation Center (1)
- Security Operation Centers (1)
- Security Operations Center (1)
- Sequenzanalyse (1)
- Sicherheit (1)
- Smart Home (1)
- Smart Load Management (1)
- Sound (1)
- Stand Alone (1)
- Start-up Companies (1)
- Steroid hormone micropollutant (1)
- Steroidhormon (1)
- Strom (1)
- Substrate and product specificity (1)
- Supervised Learning (1)
- Supply Chain Risk Management (1)
- Surveillance (1)
- Sustainability (1)
- TCO (1)
- TLV (1)
- TQM (1)
- Temperature control (1)
- Temperaturregelung (1)
- Terminal (1)
- Test-driven development (1)
- Thread (1)
- Timing Attacks (1)
- Tor (1)
- Total Cost of Ownership (1)
- Total Quality Management (1)
- Trichoderma reesei (1)
- Ultrafiltration (1)
- Unsupervised Learning (1)
- Unternehmensgründung (1)
- Unüberwachtes Lernen (1)
- Verfahrenstechnik (1)
- Volkswirtschaft (1)
- Vulnerabilities (1)
- Vulnerability Management (1)
- Vulnerability Tracking (1)
- Vulnerability identification (1)
- Wars (1)
- Wasseraufbereitung (1)
- Water treatment (1)
- Web development (1)
- Webassembly (1)
- Webentwicklung (1)
- Wireless Technology (1)
- Zement (1)
- Zugriffskontrolle (1)
- Zugriffsverwaltung (1)
- access management (1)
- analysis (1)
- android-security (1)
- artificial neural networks (1)
- audio (1)
- binary (1)
- biochar (1)
- biogas (1)
- cellulase (1)
- classification (1)
- climate-neutral (1)
- cloud business impacts (1)
- community (1)
- confidentiality (1)
- cyber attack (1)
- dahua (1)
- determinats of FDI (1)
- economic (1)
- enzymatic pretreatment (1)
- fermentation (1)
- financial impacts (1)
- fingerprinting (1)
- identity management (1)
- industrial communication (1)
- lignocellulosic material (1)
- machine learning (1)
- microphone (1)
- mitigations (1)
- moving to cloud (1)
- network layer (1)
- ns-3 (1)
- permissions (1)
- pharmaceuticals (1)
- privacy (1)
- protocol (1)
- routing (1)
- routing locator (1)
- sound (1)
- steel industry (1)
- thickness estimation (1)
- water treatment (1)
- web application security (1)
- Überwachtes Lernen (1)
Institute
- Fakultät Medien (M) (ab 22.04.2021) (25)
- Fakultät Maschinenbau und Verfahrenstechnik (M+V) (14)
- Fakultät Elektrotechnik, Medizintechnik und Informatik (EMI) (ab 04/2019) (12)
- Fakultät Wirtschaft (W) (10)
- ivESK - Institut für verlässliche Embedded Systems und Kommunikationselektronik (3)
- Fakultät Elektrotechnik und Informationstechnik (E+I) (bis 03/2019) (2)
- Fakultät Medien und Informationswesen (M+I) (bis 21.04.2021) (2)
- INES - Institut für nachhaltige Energiesysteme (2)
- IUAS - Institute for Unmanned Aerial Systems (1)
Open Access
- Closed (31)
- Closed Access (24)
- Open Access (6)
- Diamond (2)
The primary objective of this thesis is to examine the lean accounting transformation, which involves applying lean management principles to the accounting domain. In recent years, various sectors, including manufacturing, healthcare, and services, have experienced success with lean management practices. Nevertheless, the implementation of lean accounting within financial management has not been as extensively explored. This research aims to bridge that gap by scrutinizing the benefits and potential drawbacks of adopting lean accounting practices in business operations.
This research uses a combination of qualitative techniques and an extensive literature review to better understand the present subject matter. By describing the ideas of lean management and standard accounting and highlighting the fundamental distinctions between the two systems, the literature study lays a theoretical framework. The case studies illustrate the benefits of adopting lean accounting processes with real-world examples of firms that have made the transition effectively.
In the quantitative analysis of lean accounting's impact, both financial and operational factors are examined extensively. The results indicate that companies embracing lean accounting practices experience significant improvements in productivity, cost reduction, and decisionmaking quality. By highlighting the potential gains to be made by incorporating lean techniques into accounting procedures, this study adds to the current body of information on lean management. The findings offer practical implications for accounting professionals, business leaders, and policymakers interested in leveraging lean accounting to drive organizational performance improvement. The thesis finishes with suggestions for further study in this area, lean accounting.
Risk-based Cybermaturity Assessment Model - Protecting the company against ransomware attacks
(2023)
Ransomware has become one of the most catastrophic attacks in the previous decade, hurting businesses of all sorts worldwide. So, no organization is safe, and most companies are reviewing their ransomware defensive solutions to avoid business and operational hazards. IT departments are using cybersecurity maturity assessment frameworks like CMMC, C2M2, CMMI, NIST, CIS, CPP, and others to analyze organization security capabilities. In addition to maturity assessment models for the process layer and human pillar, there are much research on the analysis, identification, and defense of cyber threats in product/software layers that propose state-of-the-art approaches.
This motivates a comprehensive ransomware cyber security solution. Then, a crucial question arises: “How companies can measure the security maturity of controls in a specific danger for example for Ransomware attack?” Several studies and frameworks addressed this subject.
Complexity of understanding the ransomware attack, Lack of comprehensive ransomware defense solutions and Lack of cybermaturity assessment model for ransomware defense solutions are different aspects of problem statement in this study. By considering the most important limitations to developing a ransomware defense cybermaturity assessment method, this study developed a cybermaturity assessment methodology and implemented a Toolkit to conduct cyber security self-assessment specifically for ransomware attack to provide a clearer vision for enterprises to analyze the security maturity of controls regardless of industry or size.
Die Vision vom "Internet der Dinge" prägt seit Jahren Forschung und Entwicklung, wenn es um smarte Technologien und die Vernetzung von Geräten geht. In der Zukunft wird die reale Welt zunehmend mit dem Internet verknüpft, wodurch zahlreiche Gegenstände (Dinge) des normalen Alltags dazu befähigt werden, zu interagieren und sowohl online als auch autark zu kommunizieren. Viele Branchen wie Medizin, Automobilbau, Energieversorgung und Unterhaltungselektronik sind gleichermaßen betroffen, wodurch trotz Risiken auch neues wirtschaftliches Potential entsteht. Im Bereich "Connected Home" sind bereits Lösungen vorhanden, mittels intelligenter Vernetzung von Haushaltsgeräten und Sensoren, die Lebensqualität in den eigenen vier Wänden zu erhöhen. Diese Arbeit beschäftigt sich mit dem Thread Protokoll; einer neuen Technologie zur Integration mehrerer Kommunikationsschnittstellen innerhalb eines Netzwerks. Darüber hinaus wird die Implementierung auf Netzwerkebene (Network Layer) vorgestellt, sowie aufbereitete Informationen bezüglich verwendeter Technologien dargestellt.
The cellulase-producing Trichoderma reesei strain RL-P37 exhibits significant potential, yielding 7.3 g/L of cellulase in 241 hours. Microscopic investigations reveal a link between spore formation and enzyme production, suggesting the need for research into the intricate relationship between enzyme production, stress responses, and the nutritional prerequisites of fungi. Comparatively, the use of sodium hydroxide (NaOH) treatment, as opposed to water treatment, results in the reduction of micronutrient content and carbon source extraction as filtrate. Despite these challenges, research by He et al. (2021) highlights NaOH's efficiency in cellulose extraction from plant-based sources. Using NaOH pretreatment can be proven as effective by designing a proper cultivation method. The selection of inducers for enzyme induction gains importance, with soluble inducers, as emphasized by Zhang et al. (2022), exhibiting superior effectiveness. Hence, adopting soluble inducers in designing cultivation methods for improved enzyme production in shaking flasks is recommended. Enzymatic treatment of bio-waste, as outlined by Hu et al. (2021), shows promise in augmenting essential component content by breaking down plant cell walls and intercellular compartments. However, the feasibility of using an artificial bio-waste medium for cultivating Trichoderma reesei is questioned. Investigating the impact of micronutrient levels, particularly the inhibitory role of zinc, on fungal growth becomes essential. These findings underscore the necessity for ongoing research and optimization in cellulase production, emphasizing both strain productivity and cultivation methodologies.
Communication protocols enable information exchange between different information systems. If protocol descriptions for these systems are not available, they can be reverse-engineered for interoperability or security reasons. This master thesis describes the analysis of such a proprietary binary protocol, named the DVRIP or Dahua private protocol from Dahua Technology. The analysis contains the identification of the DVRIP protocol header format, security mechanisms and vulnerabilities inside the protocol implementation. With the revealing insights of the protocol, an increase of the overall security is achieved. This thesis builds the foundation for further targeted security analyses.
Webassembly is a new technology to create application in a new way. Webassembly is being developed since 2017 by the worldwide web consortium (w3c). The primary task of webassembly is to improve web applications.
Today, more and more applications are being created as web applications. Web applications have some advantages - they are platform independent and even mobile platforms can run them, and no installation is needed apart from a modern web browser.
Currently, web applications are being developed in JavaScript (JS), hypertext mark-up language 5 (HTML 5), and cascading style sheets (CSS).
These technologies are not made for huge web applications, but they should not be replaced by webassembly; rather, webassembly is an extension to the currently existing technology.
The purpose of webassembly is to fix or improve the problems in web application development.
This master’s thesis reviews all of the aspects and checks whether the promises of webassembly are kept and where problems still exist.
The growing threat posed by multidrug-resistant (MDR) pathogens, such as Klebsiella pneumoniae (Kp), represents a significant challenge in modern medicine. Traditional antibiotic therapies are often ineffective against these pathogens, leading to high mortality rates. MDR Kp infections pose a novel challenge in military medical contexts, particularly in Medical Biodefense, as they can be deliberately spread, leading to resource-intensive care in military centres. Recognizing this issue, the European Defence Agency initiated a prioritised research project in 2023 (EDF Resilience PHAGE- SGA 2023). To address this challenge, the Bundeswehr Institute of Microbiology (IMB) leads BMBF- (Federal Ministry of Education and Research) and EU-funded projects on the use of bacteriophages as adjuvant therapy alongside antibiotics. Since 2017, the IMB has isolated and characterised Kp phages, collecting over 600 isolates and optimizing their production for therapy, in compliance with the EMA (European Medicine Agency) guidelines. This involves in vitro phage genome packaging to minimize endotoxin load, reduce manufacturing costs, and shorten production times. The goal of this work was to establish MinION sequencing (Oxford Nanopore Technology) as a quick and reliable way for initial identification and characterisation of phage genomes. Especially as a quick screening method for isolated on Kp, prior to more precise but also more expensive and time consuming sequencing methods like Illumina. This characterisation is crucial for developing a personalized pipeline aimed at producing magistral or Good Manufacturing Practice (GMP) quality medicinal phage solutions tailored individually for each patient. DNA extraction methods were compared to identify suitable input DNA for sequencing purposes. Additionally, the quality of this DNA was as- sessed to determine its suitability for in vitro phage packaging, which was successfully done achieving a phage titer of 103, confirming that the DNA used for MinION sequencing could indeed be used for acellular packaging. The created genomes were annotated and compared with Illumina sequencing, revealing high similarity in all five individually tested cases. Between the generated sequences only a 4% maximal percentual difference in genome size was observed, while simultaneously showing high similarity in the actual sequence. Throughout the course of this study, a total of 645.15 GB of sequencing data were generated. In total, 38 phages were successfully characterised, with 21 phage genomes assembled and annotated, and saved in the IMB database.
The effects of climate change, including severe storms, heat waves, and melting glaciers, are highlighted as an urgent concern, emphasising the need to decrease carbon emissions to restrict global warming to 1.5°C. To accomplish this goal, it is vital to substitute fossil fuel-based power plants with renewable energy sources like solar, wind, hydro, and biofuels. Despite some progress being made, the proportion of renewables used in generating electricity is still lower than the levels needed for 2030 and 2050. Decarbonising the power grid is also critical in lowering the energy consumption of buildings, which is responsible for a substantial percentage of worldwide electricity usage. Even though there has been substantial expansion in the worldwide renewable energy market in the past 15 years, the transition to renewable energy sources also requires taking into account the importance of energy trading.
Peer-to-peer (P2P) electricity trading is an emerging type of energy exchange that can revolutionise the energy sector by providing a more decentralised and efficient way of trading energy. This research deals about P2P electricity trading in a carbon-neutral scenario. 'Python for Power System Analysis' (PyPSA) was used to develop models through which the P2P effect was tested. Data for the entire state of Baden-Württemberg (BW) was collected. Three scenarios were taken into consideration while developing models: 2019 (base), 2030 (coal phase-out), and 2040(climate neutral). Alongside this, another model with no P2P trading was developed to make a comparison. In addition, the use case of community storage in a P2P trading network is also presented.
The research concludes that P2P has a significant positive effect on a pathway to achieve climate neutrality. The findings show that the share of renewables in electricity generation is increasing compared to conventional sources in BW, which can be traded to meet the demand. From the storage analysis, it can be concluded that community storage can be effectively utilised in P2P trading. While the emissions are reduced, the operating costs are also reduced when the grid has P2P trading available. By highlighting the benefits of P2P trading, this research contributed to the growing body of research on the effectiveness of P2P trading in an electricity network grid.
The Lattice Boltzmann Method is a useful tool to calculate fluid flow and acoustic effects at the same time. Although the acoustic perturbation is much smaller than normal pressure differences in fluid flow, this direct calculation is a great advantage of the Lattice Boltzmann Method (LBM). But each border used in calculation produces a multitude of reflections with the acoustic waves, which lead to an unusable result. Therefore, it is worked on different absorbing techniques.
In this thesis three absorbing layer techniques are described, explained and reviewed with different simulations. The absorbing layers are implemented in a basic LBM code in C++, and with this umpteen simulations within a box were performed to compare the different absorbing layers. The Doppler effect and a cylinder flow are also examined to compare the damping efficiencies.
The three studied absorbing techniques are the sponge layer, the perfectly matched layer and a force based Term II absorbing layer. The sponge layer is easy to implement but gives worse results than a calculation without any absorbing layer. The perfectly matched layer and a force based absorbing term provide very good results but the perfectly matched layer has problems with instability. The force based absorbing layer represents the best compromise between the additional computation time due the absorbing layer and the achieved damping efficiency.
As e-commerce platforms have grown in popularity, new difficulties have emerged, such as the growing use of bots—automated programs—to engage with e-commerce websites. Even though some algorithms are helpful, others are malicious and can seriously hurt e-commerce platforms by making fictitious purchases, posting fictitious evaluations, and gaining control of user accounts. Therefore, the development of more effective and precise bot identification systems is urgently needed to stop such actions. This thesis proposes a methodology for detecting bots in E-commerce using machine learning algorithms such as K-nearest neighbors, Decision Tree, Random Forest, Support Vector Machine, and Neural Network. The purpose of the research is to assess and contrast the output of these machine learning methods. The suggested approach will be based on data that is readily accessible to the public, and the study’s focus will be on the research of bots in e-commerce.
The purpose of the study is to provide an overview of bots in e-commerce, as well as information on the different kinds and traits of bots, as well as current research on bots in e-commerce and associated work on bot detection in e-commerce. The research also seeks to create a more precise and effective bot detection system as well as find critical factors in detecting bots in e-commerce.
This research is significant because it sheds light on the increasing issue of bots in e-commerce and the requirement for more effective bot detection systems. The suggested approach for using machine learning algorithms to identify bots in ecommerce can give e-commerce platforms a more precise and effective bot detection system to stop malicious bot activities. The study’s results can also be used to create a more effective bot detection system and pinpoint key elements in detecting bots in e-commerce.
Singapore’s success in transforming itself from a poor, vulnerable economy to one of the richest countries in the world (IMF, 2016) is nothing short of inspirational to many small economies around the globe. Given its lack of resources, Singapore relied upon foreign investors to fuel its growth not only through cash injection into the economy in the form of Foreign Direct Investments (FDI) but also to help upgrade its skills and technological stock. This study looks at how Singapore inspired many Multi-National Corporations (MNCs) into pouring a large sum of investments into this small ailing citystate and if this idea can be generalized to apply it in other economies, especially in Oman.
In a bid to explain the large flow of Capital into an economy, this study moves on further to review most prominent literature in the field since Macdougall (1958) first laid the groundwork for the subsequent theories on FDI. Based on the review of several previous studies, the most significant determinants of FDI were found to be government policy and political stability, inflation rate as a proxy for economic stability, quality of infrastructure and institutions, market size of the host country, openness to trade, tax policies and access to low cost factors of production.
Through a case study method with the inductive approach, this study finds that Singapore excels in all of the determinants of FDI except for the market size of the host country and access to low-cost factors of production. However, it more than compensates for these shortcomings with its strategic geographical location and numerous bilateral and regional trade agreements that give it access to markets around the region. Oman like Singapore ranks well in many of these determinants that make it a potential destination for investment. However, the sultanate could gain more interest from the MNC’s to help its growth by optimizing its policies to lower existing barriers, easing immigration laws to meet the short term skill shortage, allowing for 100 percent foreign ownership, allowing for more liberal property rights, working to improve corruption perception and opting for more trade agreements to give it easy access to larger markets. Moreover, the economy’s heavy reliance on hydrocarbon exports is seen as a major risk by investors as it creates an economic vulnerability which could potentially overshadow many other benefits of investing in the sultanate. Besides the aforementioned determinants, a lot also depends on the success of Oman’s diversification plans.
Study of impact of change in market economics of Biosimilars due to SPC waiver on EU 469/2009
(2023)
This research was conducted to understand and investigate the impact of SPC waiver EU 933/2019 made as an amendment to EU 469/2019. The research was conducted for analysis and extraction of the data to compile the exact number of biological products impacted with the SPC waiver. The highest sale top-5 products were identified according to the expert’s opinion. The sales revenue opportunity valuable to the top-5 products in the top-5 non-EU markets for early exports is investigated. Additionally, a survey was conducted to assess the readiness of the industry for these changes. The information from this study will be very useful to students of the biopharmaceutical market research and to the stakeholders from the biopharmaceutical industry.
In recent years, the demand for reliable power, driven by sensitive electronic equipment, has surged. Even minor deviations from the nominal supply can lead to malfunctions or failure. Despite technological advancements, power quality issues persist due to various factors like short circuits, overloads, voltage fluctuations, unbalanced loads, and non-linear loads.
This thesis extensively explores power quality anomalies in industrial and commercial sectors, using power system data as the primary analytical resource. It addresses the critical need for power supply reliability in today's evolving power grid industry, affected by non-linear loads, renewable energy integration, and electric vehicles. This field of study is paramount for ensuring power supply reliability and stability in the evolving power grid industry.
The core of this thesis involves a comprehensive investigation of power quality, with a focus on frequency, power, and harmonics in voltage and current signals. The research employs Python programming for advanced data analysis, utilizing techniques such as advanced Fast Fourier Transformation (FFT) analysis. The primary objective is to provide valuable insights aimed at elevating power supply quality and enhancing reliability in both industrial and commercial environments.
Truth is the first causality of war”, is a very often used statement. What rather intrigues the mind is what causes the causality of truth. If one dives deeper, one may also wonder why is this so-called truth the first target in a war. Who all see the truth before it dies. These questions rarely get answered as the media and general public tends to focus more on the human and economic losses in a war or war like situation. What many fail to realize is that these truthful pieces of information are critical to how a situation further develops. One correct information may change the course of the whole war saving millions and one mis-information may do the opposite.
Since its inception, some studies have been conducted to propose and develop new applications for OSINT in various fields. In addition to OSINT, Artificial Intelligence is a worldwide trend that is being used in conjunction witThe question here is, what is this information. Who transmits this and how? What is the source. Although, there has been an extensive use of the information provided by the secret services of any nation, which have come handy to many, another kind of information system is using the one that is publicly available, but in different pieces. This kind of information may come from people posting on social media, some publicly available records and much more. The key part in this publicly available information is that these are just pieces of information available across the globe from various different sources. This could be seen as small pieces of a puzzle that need to be put together to see the bigger picture. This is where OSINT comes in place.
h other areas (AI). AI is the branch of computer science that is in charge of developing intelligent systems. In terms of contribution, this work presents a 9-step systematic literature review as well as consolidated data to support future OSINT studies. It was possible to understand where the greatest concentration of publications was, which countries and continents developed the most research, and the characteristics of these publications using this information. What are the trends for the next OSINT with AI studies? What AI subfields are used with OSINT? What are the most popular keywords, and how do they relate to others over time?A timeline describing the application of OSINT is also provided. It was also clear how OSINT was used in conjunction with AI to solve problems in various areas with varying objectives. Private investigators and journalists are no longer the primary users of open-source intelligence gathering and analysis (OSINT) techniques. Approximately 80-90 percent of data analysed by intelligence agencies is now derived from publicly available sources. Furthermore, the massive expansion of the internet, particularly social media platforms, has made OSINT more accessible to civilians who simply want to trawl the Web for information on a specific individual, organisation, or product. The General Data Protection Regulation (GDPR) of the European Union was implemented in the United Kingdom in May 2018 through the new Data Protection Act, with the goal of protecting personal data from unauthorised collection, storage, and exploitation. This document presents a preliminary review of the literature on GDPR-related work.
The reviewed literature is divided into six sections: ’What is OSINT?’, ’What are the risks?’ and benefits of OSINT?’, ’What is the rationale for data protection legislation?’, ’What are the current legislative frameworks in the UK and Europe?’, ’What is the potential impact of the GDPR on OSINT?’, and ’Have the views of civilian and commercial stakeholders been sought and why is this important?’. Because OSINT tools and techniques are available to anyone, they have the unique ability to be used to hold power accountable. As a result, it is critical that new data protection legislation does not impede civilian OSINT capabilities.
In this paper we see how OSINT has played an important role in the wars across the globe in the past. We also see how OSINT is used in our everyday life. We also gain insights on how OSINT is playing a role in the current war going on between Russia and Ukraine. Furthermore, we look into some of these OSINT tools and how they work. We also consider a use case where OSINT is used as an anti terrorism tool. At the end, we also see how OSINT has evolved over the years, and what we can expect in the future as to what OSINT may look like.
In each company Top Managers have the responsibility to take major decisions that supports the success of their company, Adopting TQM is one of these decisions, the decision to carry out companies’ operations and procedures within TQM frameworks. (ASQ , n.d.). Applying TQM, involves implementing practices that needs putting extra efforts, otherwise there will be no use of the practices and the execution. (Nicca Jirah F Campos1, 2022).
Specifically in service sector, where the key to success and increased profit, comes directly through a satisfied customer. Therefor there is a need for both management and staff to have big tolerance and willingness to achieve the needed satisfaction, in order to attain the results that every company wants. (Charantimath, 2013)
In Germany in terms of customer care practices there is a famous stereotype ‘Customer is not the king’ A reputation That after DW investigated it, DW expressed it as a phenomenon where both expats and Germans tend to believe that service companies in Germany should do a better job of treating their consumers. (DW, 2016)
New concepts of business have emerged in the late century, for example strategy, leadership, marketing, entrepreneurship and others, these concepts spread internationally among most of the companies around the world. Many studies have been done reviewing these new business structures, some of them addressed the cultural differences within countries upon the applying them. But not many studies concentrated on taking into consideration how cultural differences affects the Implementation of TQM. (Lagrosen, 2002). It was concluded in general that although the comprehensive fundamentals of quality management are applicable and similar worldwide in all nations, but when coming to real practice accurate tunning must be made, it must be taken into account aligning different standards, due to different work cultures and traditions in Europe. (Krueger, 1999)
Encryption techniques allow storing and transferring of sensitive information securely by using encryption at rest and encryption in transit, respectively. However, when computation is performed on these sensitive data, the data needs to be decrypted first and encrypted again after performing the computations. During the computations, the sensitive data becomes vulnerable to attackers as it's in decrypted form. Homomorphic encryption, a special type of encryption technique that allows computation on encrypted data can be used to solve the above-mentioned problem. The best way to achieve maximum security with homomorphic encryption is to perform at least the homomorphic encryption and decryption on the client side (browser) of a web application by not trusting the server. At present time there are many libraries with different homomorphic schemes available for homomorphic encryption. However, there are very few to no JavaScript libraries available to perform homomorphic encryption on the client side of any web application. This thesis mainly focuses on the JavaScript implementation of client-side homomorphic encryption. The fully homomorphic encryption scheme BFV is selected for the implementation. After implementing the fully homomorphic encryption scheme based on the “py-fhe” library, tests are also carried out in order to determine the applicability (in terms of time consumption, security and correctness) of this implementation in a web application by comparing the performance and security for different test cases and different settings.
Distributed Flow Control and Intelligent Data Transfer in High Performance Computing Networks
(2015)
This document contains my master thesis report, including problem definition, requirements, problem analysis, review of current state of the art, proposed solution,
designed prototype, discussions and conclusion.
During this work we propose a collaborative solution to run different types of operations in a broker-less network without relying on a central orchestrator.
Based on our requirements, we define and analyze a number of scenarios. Then we design a solution to address those scenarios using a distributed workflow management approach. We explain how we break a complicated operation into simpler parts and how we manage it in a non-blocking and distributed way. Then we show how we asynchronously launch them on the network and how we collect and aggregate results. Later on we introduce our prototype which demonstrates the proposed design.
The rapid pace of innovation and technological advancements has led to the emergence of start-up companies in various sectors. To remain competitive and sustainable, start-ups need to make informed business decisions that can enhance their operations and profitability. Business Intelligence (BI) has become an essential tool for businesses of all sizes in managing their operations and gaining a competitive edge.
This master thesis explores the role of Business Intelligence in start-up companies. The study aims to investigate the use of BI in start-up companies, the drivers and the inhibitors for its adoption and their relationship with price. The research conducted for this thesis involves a review of relevant literature on Business Intelligence, start-up companies, and related topics. The study also includes structured survey with entrepreneurs, start-up company executives, and BI experts to gain data for a quantitative analysis of the topic.
The thesis aims to contribute to the existing body of knowledge on Business Intelligence and its role in start-up companies. The research conducted for this thesis can be of value to start-up entrepreneurs, investors, and other stakeholders who seek to improve their understanding of the benefits and challenges of implementing BI in start-up companies.
This Master's Thesis discusses intelligent sensor networks considering autonomous sensor placement strategies and system health management. Sensor networks for an intelligent system design process have been researched recently. These networks consist of a distributed collective of sensing units, each with the abilities of individual sensing and computation. Such systems can be capable of self-deployment and must be scalable, long-lived and robust. With distributed sensor networks, intelligent sensor placement for system design and online system health management are attractive areas of research. Distributed sensor networks also cause optimization problems, such as decentralized control, system robustness and maximization of coverage in a distributed system. This also includes the discovery and analysis of points of interest within an environment. The purpose of this study was to investigate a method to control sensor placement in a world with several sources and multiple types of information autonomously. This includes both controlling the movement of sensor units and filtering of the gathered information depending on individual properties to increase system performance, defined as a good coverage. Additionally, online system health management was examined in this study regarding the case of agent failures and autonomous policy reconfiguration if sensors are added to or removed from the system. Two different solution strategies were devised, one where the environment was fully observable, and one with only partial observability. Both strategies use evolutionary algorithms based on artificial neural networks for developing control policies. For performance measurement and policy evaluation, different multiagent objective functions were investigated. The results of the study show that in the case of a world with multiple types of information, individual control strategies performed best because of their abilities to control the movement of a sensor entity and to filter the sensed information. This also includes system robustness in case of sensor failures where other sensing units must recover system performance. Additionally, autonomous policy reconfiguration after adding or removing of sensor agents was successful. This highlights that intelligent sensor agents are able to adapt their individual control policies considering new circumstances.
In the past ten years, applications of artificial neural networks have changed dramatically. outperforming earlier predictions in domains like robotics, computer vision, natural language processing, healthcare, and finance. Future research and advancements in CNN architectures, Algorithms and applications are expected to revolutionize various industries and daily life further. Our task is to find current products that resemble the given product image and description. Deep learning-based automatic product identification is a multi-step process that starts with data collection and continues with model training, deployment, and continuous improvement. The caliber and variety of the dataset, the design selected, and ongoing testing and improvement all affect the model's effectiveness. We achieved 81.47% training accuracy and 72.43% validation accuracy for our combined text and image classification model. Additionally, we have discussed the outcomes from the other dataset and numerous methods for creating an appropriate model.
This study investigates the impact of global payroll outsourcing on organizational efficiency and cost reduction based on the analysis of diverse implications stemming from thirty one (31) survey results. The findings reveal multifaceted challenges and benefitsassociated with outsourcing global payroll processing.
The research also unveils the most benefits of global payroll outsourcing. Notably, there's a consensus on the reduction in time-to-process payroll, cost per payroll processed, and improved payroll accuracy rate. Outsourcing streamlines processes, enhances operational efficiency, and contributes to faster, more accurate financial reporting.
Despite these benefits and challenges, statistical analysis reveals weak correlations between outsourcing global payroll and cost reduction or improved efficiency in various parameters, indicating a lack of a significant relationship. Consequently, the results, suggest no substantial correlation between global payroll outsourcing and enhanced efficiency or cost reduction based on this study's data.
Self-sufficient enzymes belong to the cytochrome P450 (CYP) group and are known for their superior hydroxylation catalytic activity. In the pursuit of identifying new pesticides to combat antimicrobial-resistant pathogens, we employed BM3 wild type (BM3-WT), the fastest monohydroxylating CYP, along with its seven homologs, to investigate the production of potential hydroxylated derivatives from the established pesticide, 4-oxocrotonic acid using high-pressure liquid chromatography (HPLC) method. Following the recombinant production of BM3-WT and three other homologs in E. coli, and their subsequent purification using Immobilized Metal Affinity Chromatography (IMAC), a novel enzyme assay approach was developed as a substitute for the carbon monoxide (CO) assay. This new method relied on the measurement of NADPH consumption at 340 nm by BM3-WT for palmitic acid. Leveraging this established technique, we explored the substrate specificity of BM3-WT and its homologs not only on palmitic acid but also on other structurally similar compounds, including 4-oxocrotonic acid. The results obtained from the established NADPH assay indicate that all tested enzymes displayed greater catalytic activity on 4-oxocrotonic acid in comparison to other substrates with similar structures. However, the impact of BM3-WT and its homologs on 4-oxocrotonic acid varied in terms of product specificity. Enzymes such as Poh, Trr and Bas-CYP D exhibited specificity in producing solely monohydroxylated products, while others tended to yield dehydroxylated and ketol metabolites.
The goal of this thesis is to thoroughly investigate the concepts of stand-alone and decarbonization of optical fiber networks. Because of their dependability, fast speed, and capacity, optical fiber networks are vital inmodern telecommunications. Their considerable energy consumption and carbon emissions, on the other hand, constitute a danger to global sustainability objectives and must be addressed.
The first section of the thesis presents a summary of the current state of optical fiber networks, their
components, and the energy consumption connected with them. This part also goes over the difficulties of lowering energy usage and carbon emissions while preserving network performance and dependability.
The second section of the thesis focuses on the stand-alone idea, which entails powering the optical fiber network with renewable energy sources and energy-efficient technology. This section investigates and explores the possibilities of renewable energy sources like solar and wind power to power the network. It also investigates energy-efficient technologies like virtualization and cloud computing, as well as their potential to minimize network energy usage.
The third section of the thesis focuses on the notion of decarbonization, which entails lowering carbon emissions linked with the optical fiber network. This section looks at various carbon-reduction measures, such as employing low-carbon energy sources and improving energy efficiency. It also covers the relevance of carbon offsets and the difficulties associated with adopting decarbonization measures in the context of optical fiber networks.
The fourth section of the thesis compares the ideas of stand-alone and decarbonization. It investigates the advantages and disadvantages of each strategy, as well as their potential to minimize energy consumption and carbon emissions in optical fiber networks. It also explores the difficulties in applying these notions as well as potential hurdles to their wider adoption.
Finally, the need of addressing the energy consumption and carbon emissions connected with optical fiber networks is emphasized in this thesis.
It outlines important obstacles and potential impediments to adopting these initiatives and gives insights into potential ways for decreasing them.
It also makes suggestions for further study in this area.
As information technology continues to advance at a rapid speed around the world, new difficulties emerge. The growing number of organizational vulnerabilities is among the most important issues. Finding and mitigating vulnerabilities is critical in order to protect an organization’s environment from multiple attack vectors.
The study investigates and comprehends the complete vulnerability management process from the standpoint of the security officer job role, as well as potential improvements. Few strategies are used to achieve efficient mitigation and the de- velopment of a process for tracking and mitigating vulnerabilities. As a result, a qualitative study is conducted in which the objective is to create a proposed vulner- ability and risk management process, as well as to develop a system for analyzing and tracking vulnerabilities and presenting the vulnerabilities in a graphical dash- board format. This thesis’s data was gathered through an organized literature study as well as through the use of various web resources. We explored numerous ap- proaches to analyze the data, such as categorizing the vulnerabilities every 30, 60, and 90 days to see whether the vulnerabilities were reoccurring or new. According to our findings, tracking vulnerabilities can be advantageous for a security officer.
We come to the conclusion that if an organization has a proper vulnerability tracking system and vulnerability management process, it can aid security officers in having a better understanding of and making plans for reducing vulnerabilities. In terms of system patching and vulnerability remediation, it will also assist the security officer in identifying areas of weakness in the process. As a result, the suggested ways provide an alternate approach to managing and tracking vulnerabilities in an effective manner, although there is still a small area that needs additional analysis and research to make it even better.
Threat Modeling is a vital approach to implementing ”Security by Design” because it enables the discovery of vulnerabilities and mitigation of threats during the early stage of the Software Development Life Cycle as opposed to later on when they will be more expensive to fix. This thesis makes a review of the current threat Modeling approaches, methods, and tools. It then creates a meta-model adaptation of a fictitious cloud-based shop application which is tested using STRIDE and PASTA to check for vulnerabilities, weaknesses, and impact risk. The Analysis is done using Microsoft Threat Modeling Tool and IriusRisk. Finally, an evaluation of the results is made to ascertain the effectiveness of the processes involved with highlights of the challenges in threat modeling and recommendations on how security developers can make improvements.
Organizations striving to achieve success in the long term must have a positive brand image which will have direct implications on the business. In the face of the rising cyber threats and intense competition, maintaining a threat-free domain is an important aspect of preserving that image in today's internet world. Domain names are often near-synonyms for brand names for numerous companies. There are likely thousands of domains that try to impersonate the big companies in a bid to trap unsuspecting users, usually falling prey to attacks such as phishing or watering hole. Because domain names are important for organizations for running their business online, they are also particularly vulnerable to misuse by malicious actors. So, how can you ensure that your domain name is protected while still protecting your brand identity? Brand Monitoring, for example, may assist. The term "Brand Monitoring" applies only to keep tabs on an organization's brand performance, reception, and overall online presence through various online channels and platforms [1]. There has been a rise in the need of maintaining one's domain clear of any linkages to malicious activities as the threat environment has expanded. Since attackers are targeting domain names of organizations and luring unsuspecting users to visit malicious websites, domain monitoring becomes an important aspect. Another important aspect of brand abuse is how attackers leverage brand logos in creating fake and phishing web pages. In this Master Thesis, we try to solve the problem of classification of impersonated domains using rule-based and machine learning algorithms and automation of domain monitoring. We first use a rule-based classifier and Machine Learning algorithms to classify the domains gathered into two buckets – "Parked" and "Non-Parked". In the project's second phase, we will deploy object detection models (Scale Invariant Feature Transform - SIFT and Multi-Template Matching – MTM) to detect brand logos from the domains of interest.
Estimation and projecting total steel industry production costs from 2019 to 2030 for Germany
(2023)
This thesis analyses the total production cost of the German steel industry from 2019 to 2022, as well as a projection of the German steel industry's total production cost until 2030. The research separates the costs of steel production into their primary components, such as raw materials, energy, CO2 cost, capital expenses and operating expenses. The cost of steel production is determined separately for primary steelmaking with the blast furnace and basic oxygen furnace (BF-BOF) and secondary steelmaking with the electric arc furnace (EAF).
The analysis indicates that, following the COVID-19 disaster and the fuel crisis, the overall cost of producing steel in Germany has progressively risen over the previous few years, reaching its peak in the first half of 2022. In addition, there are considerable disparities between the production costs of primary and secondary steelmaking processes, with primary steelmaking generally being more expensive.
In this analysis, the total cost of production for the German steel industry in the year 2030 has been estimated by taking into account historical trends as well as other predictions that are currently available.
This thesis provides overall insights on the economics of the German steel sector. By giving thorough information on production costs and changes over time, this research can assist guide crucial future investment decisions in this essential industry. To ensure long-term success, our findings emphasize the significance of investing in more sustainable and ecologically friendly steel production processes.
Technology advancement has played a vital role in business development; however, it has opened a broad attack surface. Passwords are one of the essential concepts used in applications for authentication. Companies manage many corporate applications, so the employees must meet the password criteria, which leads to password fatigue. This thesis addressed this issue and how we can overcome this problem by theoretically implementing an IAM solution. In this, we disused MFA, SSO, biometrics, strong password policies and access control. We introduced the IAM framework that should be considered while implementing the IAM solution. Implementing an IAM solution adds an extra layer of security.
The Internet of Things is spreading significantly in every sector, including the household, a variety of industries, healthcare, and emergency services, with the goal of assisting all of those infrastructures by providing intelligent means of service delivery. An Internet of Vulnerabilities (IoV) has emerged as a result of the pervasiveness of the Internet of Things (IoT), which has led to a rise in the use of applications and devices connected to the IoT in our day-to-day lives. The manufacture of IoT devices are growing at a rapid pace, but security and privacy concerns are not being taken into consideration. These intelligent Internet of Things devices are especially vulnerable to a variety of attacks, both on the hardware and software levels, which leaves them exposed to the possibility of use cases. This master’s thesis provides a comprehensive overview of the Internet of Things (IoT) with regard to security and privacy in the area of applications, security architecture frameworks, a taxonomy of various cyberattacks based on various architecture models, such as three-layer, four-layer, and five-layer. The fundamental purpose of this thesis is to provide recommendations for alternate mitigation strategies and corrective actions by using a holistic rather than a layer-by-layer approach. We discussed the most effective solutions to the problems of privacy and safety that are associated with the Internet of Things (IoT) and presented them in the form of research questions. In addition to that, we investigated a number of further possible directions for the development of this research.
The purpose of this master's thesis was to set up a test bed for the absorption of chemical compounds by carbon-based sorbents and polymers and to develop a method for the detection of these substances applied by liquid chromatography.
The study made it possible to demonstrate the effectiveness of both polymers and biochars sorbents for the adosorption of specific substances. The results obtained open new paths on the study of biochar for the treatment of contaminated water. Some biochars made from plant-based materials have been shown to be almost as effective as commercial products used in plants. The developed chromatography method allows efficient separation of substances and their detection.
As cyber threats continue to evolve, it is becoming increasingly important for organizations to have a Security Operations Center (SOC) in place to effectively defend against them. However, building and maintaining a SOC can be a daunting task without clear guidelines, policies, and procedures in place. Additionally, most current SOC solutions used by organizations are outdated, lack key features and integrations, and are expensive to maintain and upgrade. Moreover, proprietary solutions can lead to vendor lock-in, making it difficult to switch to a different solution in the future.
To address these challenges, this thesis proposes a comprehensive SOC framework and an open-source SOC solution that provides organizations with a flexible and cost-effective way to defend against modern cyber threats. The research methodology involved conducting a thorough literature review of existing literature and research on building and maintaining a SOC, including using SOC as a service. The data collected from the literature review was analyzed to identify common themes, challenges, and best practices for building and maintaining a SOC.
Based on the data collected, a comprehensive framework for building and maintaining a SOC was developed. The framework addresses essential areas such as the scope and purpose of the SOC, governance and leadership, staffing and skills, technologies and tools, processes and procedures, service level agreements (SLAs), and evaluation and measurement. This framework provides organizations with the necessary guidance and resources to establish and effectively operate a SOC, as well as a reference for evaluating the service provided by SOC service providers.
In addition to the SOC framework, a modern open-source SOC solution was developed, which emphasizes several key measures to help organizations defend against modern cyber threats. These measures include real-time, actionable threat intelligence, rapid and effective incident response, continuous security monitoring and alerting, automation, integration, and customization. The use of open-source technologies and a modular architecture makes the solution cost-effective, allowing organizations to scale it up or down as needed.
Overall, the proposed SOC framework and open-source SOC solution provide organizations with a comprehensive and systematic approach for building and maintaining a SOC that is aligned with the needs and objectives of the organization. The open-source SOC solution provides a flexible and cost-effective way to defend against modern cyber threats, helping organizations to effectively operate their SOC and reduce their risk of security incidents and breaches.
An organized strategy to ensure the security of an organization is an information security management system. During various security crises, hazards, and breaches, this strategy aids an organization in maintaining the confidentiality, integrity, and accessibility of information. Organizations are getting ready to comply with information security management system criteria. Despite this, security concerns continue to plague ineffective controls, have poor connectivity, or cause a silo effect, which is a common cause. One of the causes is a low maturity model that is not synchronized with the organization’s business processes. For a higher level of maturity, it is best to evaluate the practices.
Different maturity models on information security and cyber security capacity, management processes, security controls, implementation level, and many more have already been developed by numerous international organizations, experts, and scholars. The present models, however, do not assess a particular organization's particular practices. The evaluation of the business process is frequently neglected because measurement requirements for models are typically more concentrated on examining specific elements. For this reason, it caused the maturity assessment to not be executed explicitly and broadly.
We developed an organizational information security maturity model, a combination of work of different maturity models currently existing. While making this model, we considered that any size or type of organization could use this model. The model considers the success elements of the information security management system when assessing the implementation's effectiveness. We employed a mixed-method strategy that included both qualitative and quantitative research. With the help of a questionnaire survey, we evaluated the previous research using a qualitative methodology. In the quantitative method, we'll figure out how mature the information security management system is now. The proposed model could be used to reduce security incidents by improving implementation gaps.
The current thesis conducts the study on the integration of digitalization techniques aimed at improving energy supply efficiency in off-grid energy systems. The primary objective is to fortify the security of energy supply in remote areas, particularly in instances of adverse weather conditions, unanticipated changes in load and fluctuations in the performance of renewable energy systems. This objective is to be achieved through the implementation of a smart load management strategy in stand-alone photovoltaic systems (SAPVS). This strategy involves deployment of forecasting algorithms on an edge device that operates with limited processing resources in an environment characterized for the lack of internet connection. The edge device is designed to interact with a smart home gateway that prioritizes, and schedules smart appliances based on the forecasted state of charge (SOC) in the 36-hours ahead of the SAPVS operation (the implementation of the loads schedule deployed on the Home Assistant device is out of the scope of the tasks implemented for this project).
The edge device, developed using a Raspberry Pi 3B+, was specifically intended for being implemented along with a SAPVS, in remote areas such as health stations in Africa and tropical islands, providing communities with a reliable source of electrical energy. The deployment of the strategy was carried out in four phases. The first phase involved the implementation of an Extraction-Transformation-Load (ETL) pipeline, where data was gathered from various heterogeneous hardware sources of an implemented test system that served as the enabler and testbench of this research, this test stand is composed of power electronics components such as an inverter, a MPPT solar charge controller, a smart meter, and a BOS LiFePo4 battery prototype. In the transformation stage, a data model was developed to identify the most critical parameters of the energy system, and to eliminate outliers and null values. In the load stage, a local SQL database was established for saving and structuring the data gathered and to ensure high-quality data with defined units and casting.
The second phase involved data analysis to identify the relevant features and potential exogenous variables for the forecasting model to implement. In the third phase, an Auto Regressive Moving Average (ARMA) model with two selected exogenous variables was implemented to forecast the AC load consumption profile for the 36- hours ahead of the off-grid system operation. The final phase involved the information exchange with the Home Assistant device, by transferring to it from the edge device the battery SOC present value and the predicted 36-hour ahead AC load profile information for prioritization and scheduling of loads; this through an MQTT interface.
The outcome of the experiment was a successful deployment of a data engineering and data forecasting approach that enabled data quality strategy implementation, local database storage, and forecasting algorithms on a processing and internet-constrained edge device. The interface with a home assistant implementation resulted in the successful execution of smart load management endeavors in an off-grid system, thereby enhancing the energy security of supply and contributing to the advancement of data-driven strategies in the rural electrification sector.
This thesis emphasizes the significance of digitalization strategies in smart SAPVS and highlights the potential of edge computing solutions in achieving seamless energy management in smart homes.
This research presents a comprehensive exploration of hydroponic systems and their practical applications, with a focus on innovative solutions for managing environmental and analytical sensors in hydroponic setups. Hydroponic systems, which enable soilless cultivation, have gained increasing importance in modern agriculture due to their resource-efficient and high-yield nature.
The study delves into the development and deployment of the SensVert system, an adaptable solution tailored for hydroponic environments. SensVert offers adaptability and accessibility to farmers across various agricultural domains, addressing contemporary challenges in supervising and managing environmental and analytical sensors within hydroponic setups. Leveraging LoRa technology for seamless wireless data transmission, SensVert empowers users with a feature-rich dashboard for real-time monitoring and control. The study showcases the practical implementation of SensVert through a single sensor node, seamlessly integrating temperature, humidity, pressure, light, and pH sensors. The system automates pH regulation, employing the Henderson-Hasselbalch equation, and precisely controls liquid dosing using a PID controller. At the core of SensVert lies an architecture comprising The Things Stack as the network server, Node-Red as the application server, and Grafana as the user interface. These components synergize within a local network hosted on a Raspberry Pi; effectively mitigating challenges associated with data packet transmission in areas with limited internet connectivity.
As part of ongoing research, this work also paves the way for future advancements. These include the establishment of a wireless sensor network (WSN) utilizing LoRa technology, enabling seamless over-the-air sensor node updates for maintenance or replacement scenarios. These enhancements promise to further elevate the system's reliability and functionality within hydroponic cultivation, fostering sustainable agricultural practices.
Automation research has become one of the most important tools for future thinking organizations. It includes studying the economic and social aspects to determine how accountants were affected by automating the accounting profession. Moreover, this research studied the social aspect of automation, including the accountants' satisfaction and agreement towards the shift from manual-based accounting to automated accounting. Additionally, the purpose of the research was to comprehend the aspects that affect the variance of the satisfaction and agreement levels before and after automation and whether there is a relationship between those satisfaction and agreement levels and the demographic profile of accountants.
A quantitative method was used to answer the research questions. The findings and results were gathered through an online survey. The respondents in the study represent forty-three accountants who are located and working in Germany. The implications and conclusions of the research were observed from the accountants' perspective.
The research results presented that the automation of accounting significantly impacted the accountants' profession. It indicated that accountants are satisfied with automated accounting and agree with its effects and impacts on their profession. Accountants agreed that automated accounting tasks made the accounting process more effective and valuable. The findings also showed that educational level and length of experience in automated accounting are correlated with the satisfaction of accountants towards automated accounting. It presented that the more experience in automation and higher education accountants have, the more satisfied they are with automated accounting. Due to this phenomenon, higher qualifications and more basic IT knowledge are required in comparison with previous times.
Server Side Rendering (SSR), Single Page Application (SPA), and Static Site Generation (SSG) are the three most popular ways of making modern Web applications today. If we go deep into these processes, this can be helpful for the developers and clients. Developers benefit since they do not need to learn other programming languages and can instead utilize their own experience to build different kinds of Web applications; for example, a developer can use only JavaScript in the three approaches. On the other hand, clients can give their users a better experience.
This Master Thesis’s purpose was to compare these processes with a demo application for each and give users a solid understanding of which process they should follow. We discussed the step-by-step process of making three applications in the above mentioned categories. Then we compared those based on criteria such as performance, security, Search Engine Optimization, developer preference, learning curve, content and purpose of the Web, user interface, and user experience. It also talked about the technologies such as JavaScript, React, Node.js, and Next.js, and why and where to use them. The goals we specified before the program creation were fulfilled and can be validated by comparing the solutions we gave for user problems, which was the application’s primary purpose.
Quarz crystal microbalances allow the monitoring of the adsorption process of mass from a liquid to their surface. The adsorbed mass can be analysed regarding to its protein content using mass spectromety. To ensure the protein identification the results of several measurements can be combined. A high content QCM-D array was developed to allow up to ten measurements parallel. The samples can be routed inside the array distributing one sample to several chips. The fluidic parts were prototyped using 3D printing. The assembled array was tight and the sample routing function could be demonstrated. A temperature controller was developed and implemented. The parameters for the PID controller were determined and the controller was shown to be able to keep the temperature constant over long time with high accuracy.
Conceptualization and implementation of automated optimization methods for private 5G networks
(2023)
Today’s companies are adjusting to the new connectivity realities. New applications require more bandwidth, lower latency, and higher reliability as industries become more distributed and autonomous. Private 5th Generation (5G) networks known as 5G Non-Public Networks (5G-NPN), is a novel 3rd Generation Partnership Project (3GPP)- based 5G network that can deliver seamless and dedicated wireless access for a particular industrial use case by providing the mentioned application’s requirements. To meet these requirements, several radio-related aspects and network parameters should be considered. In many cases, the behavior of the link connection may vary based on wireless conditions, available network resources, and User Equipment (UE) requirements. Furthermore, Optimizing these networks can be a complex task due to the large number of network parameters and KPIs that need to be considered. For these reasons, traditional solutions and static network configuration are not affordable or simply impossible. Despite the existence of papers in the literature that address several optimization methods for cellular networks in industrial scenarios, more insight into these existing but complex or unknown methods is needed.
In this thesis, a series of optimization methods were implemented to deliver an optimal configuration solution for a 5G private network. To facilitate this implementation, a testing system was implemented. This system enables remote control over the UE and 5G network, establishment of a test environment, extraction of relevant KPI reports from both UE and network sides, assessment of test results and KPIs, and effective utilization of the optimization and sampling techniques.
The research highlights the advantageous aspects of automated testing by using OFAT, Simulated Annealing, and Random Forest Regressor methods. With OFAT, as a common sampling method, a sensitivity analysis and an impact of each single parameter variation on the performance of the network were revealed. With Simulated Annealing, an optimal solution with MSE of roughly 10 was revealed. And, in the Random Forest Regressor, it was seen that this method presented a significant advantage over the simulated annealing method by providing substantial benefits in time efficiency due to its machine- learning capability. Additionally, it was seen that by providing a larger dataset or using some other machine-learning techniques, the solution might be more accurate.
Cloud computing has revolutionized the way businesses operate by providing them with access to scalable, cost-effective, and flexible IT resources. This technology has enabled businesses to store, manage, and process data more efficiently, leading to improved competitiveness and increased revenue. The purpose of this thesis is to explore the impacts of using cloud computing from a business perspective. The research employs both primary and secondary sources of data, including a literature review, interviews with employees who have more than 5 years of experience, a questionnaire, and observations from Billwerk+ company.
The findings of this research indicate that cloud computing has had a significant impact on businesses, providing them with cost savings, improved agility and flexibility, and enhanced access to data and applications. However, it has been revealed that the benefits of cloud computing for companies may vary according to the departments of the employees. The results of this research contribute to the existing body of knowledge on the topic of cloud computing and its impact on businesses. The findings of this thesis can be used by business owners, managers, technology professionals, and students to make informed decisions about the adoption and use of cloud computing technology.
In conclusion, this thesis provides a comprehensive understanding of the impacts of using cloud computing from a business perspective, highlighting the factors that companies consider when deciding to use cloud environments and the views from different departments. The results of this research will be valuable to a wide range of individuals interested in exploring the implications of cloud computing for businesses.
Decarbonisation Strategies in Energy Systems Modelling: APV and e-tractors as Flexibility Assets
(2023)
This work presents an analysis of the impact of introducing Agrophotovoltaic technologies and electric tractors into Germany’s energy system. Agrophotovoltaics involves installing photovoltaic systems in agricultural areas, allowing for dual usage of the land for both energy generation and food production. Electric tractors, which are agricultural machinery powered by electric motors, can also function as energy storage units, providing flexibility to the grid. The analysis includes a sensitivity study to understand how the availability of agricultural land influences Agrophotovoltaic investments, followed by the examination of various scenarios that involve converting diesel tractors to electric tractors. These scenarios are based on the current CO2 emission reduction targets set by the German Government, aiming for a 65% reduction below 1990 levels by 2030 and achieving zero emissions by 2045. The results indicate that approximately 3% of available agricultural land is necessary to establish a viable energy mix in Germany. Furthermore, the expansion of electric tractors tends to reduce the overall system costs and enhances the energy-cost-efficiency of Agrophotovoltaic investments.
Privacy is the capacity to keep some things private despite their social repercussions. It relates to a person’s capacity to control the amount, time, and circumstances under which they disclose sensitive personal information, such as a person’s physiology, psychology, or intelligence. In the age of data exploitation, privacy has become even more crucial. Our privacy is now more threatened than it was 20 years ago, outside of science and technology, due to the way data and technology highly used. Both the kinds and amounts of information about us and the methods for tracking and identifying us have grown a lot in recent years. It is a known security concern that human and machine systems face privacy threats. There are various disagreements over privacy and security; every person and group has a unique perspective on how the two are related. Even though 79% of the study’s results showed that legal or compliance issues were more important, 53% of the survey team thought that privacy and security were two separate things. Data security and privacy are interconnected, despite their distinctions. Data security and data privacy are linked with each other; both are necessary for the other to exist. Data may be physically kept anywhere, on our computers or in the cloud, but only humans have authority over it. Machine learning has been used to solve the problem for our easy solution. We are linked to our data. Protect against attackers by protecting data, which also protects privacy. Attackers commonly utilize both mechanical systems and social engineering techniques to enter a target network. The vulnerability of this form of attack rests not only in the technology but also in the human users, making it extremely difficult to fight against. The best option to secure privacy is to combine humans and machines in the form of a Human Firewall and a Machine Firewall. A cryptographic route like Tor is a superior choice for discouraging attackers from trying to access our system and protecting the privacy of our data There is a case study of privacy and security issues in this thesis. The problems and different kinds of attacks on people and machines will then be briefly talked about. We will explain how Human Firewalls and machine learning on the Tor network protect our privacy from attacks such as social engineering and attacks on mechanical systems. As a real-world test, we will use genomic data to try out a privacy attack called the Membership Inference Attack (MIA). We’ll show Machine Firewall as a way to protect ourselves, and then we’ll use Differential Privacy (DP), which has already been done. We applied the method of Lasso and convolutional neural networks (CNN), which are both popular machine learning models, as the target models. Our findings demonstrate a logarithmic link between the desired model accuracy and the privacy budget.
Even though the internet has only been there for a short period, it has grown tremendously. To- day, a significant portion of commerce is conducted entirely online because of increased inter- net users and technological advancements in web construction. Additionally, cyberattacks and threats have expanded significantly, leading to financial losses, privacy breaches, identity theft, a decrease in customers’ confidence in online banking and e-commerce, and a decrease in brand reputation and trust. When an attacker pretends to be a genuine and trustworthy institution, they can steal private and confidential information from a victim. Aside from that, phishing has been an ongoing issue for a long time. Billions of dollars have been shed on the global economy. In recent years, there has been significant progress in the development of phishing detection and identification systems to protect against phishing attacks. Phishing detection technologies frequently produce binary results, i.e., whether a phishing attempt was made or not, with no explanation. On the other hand, phishing identification methodologies identify phishing web- pages by visually comparing webpages with predetermined authentic references and reporting phishing together with its target brand, resulting in findings that are understandable. However, technical difficulties in the field of visual analysis limit the applicability of currently available solutions, preventing them from being both effective (with high accuracy) and efficient (with little runtime overhead). Here, we evaluate existed framework called Phishpedia. This hybrid deep learning system can recognize identity logos from webpage screenshots and match logo variants of the same brand with high precision. Phishpedia provides high accuracy with low run- time. Lastly, unlike other methods, Phishpedia does not require training on any phishing sam- ples whatsoever. Phishpedia exceeds baseline identification techniques (EMD, PhishZoo, and LogoSENSE), inaccurately detecting phishing pages in lengthy testing using accurate phishing data. The effectiveness of Phishpedia was tested and compared against other standard machine learning algorithms and some state-of-the-art algorithms. The given solutions performed better than different algorithms in the given dataset, which is impressive.
Among the billions of smartphone users in the world, Android still holds more than 80% of the market share. The applications which the users install have a specific set of features that need access to some device functionalities and sensors that may hold sensitive information about the user. Therefore, Android releases have set permission standards to let the user know what information is being disclosed to the application. Along with other security and privacy improvements, significant changes to the permission scheme are introduced with the Android 6.0 version (API level 23). In this master thesis, the Android permission scheme is tested on two devices from different eras. The evolution of Android over the years is examined in terms of confidentiality. For each device, two applications are built; one focused on extracting every piece of information within the confidentiality scope with every permission declared and/or requested, and the other app focused on getting this type of information without user notification. The resulting analysis illustrates whether how and in what way the Android permission scheme declined or improved over time.
Decarbonisation Strategies in Energy Systems Modelling: Biochar as a Carbon Capture Technology
(2022)
The energy system is changing since some years in order to achieve the climate goals from the Paris Agreement which wants to prevent an increase of the global temperature above 2 °C. Decarbonisation of the energy system has become for governments a big challenge and different strategies are being stablished. Germany has set greenhouse gas reduction limits for different years and keeps track of the improvement made yearly. The expansion of renewable energy systems (RES) together with decarbonisation technologies are a key factor to accomplish this objective.
This research is done to analyse the effect of introducing biochar, a decarbonisation technology, and study how it will affect the energy system. Pyrolysis is the process from which biochar is obtained and it is modelled in an open-source energy system model. A sensibility analysis is made in order to assess the effect of changing the biomass potential and the costs for pyrolysis.
The role of pyrolysis is analysed in the form of different future scenarios to evaluate the impact. The CO2 emission limits from the years 2030 and 2045 are considered to create the scenarios, as well as the integration of flexibility technologies. Four scenarios in total are assessed and the result from the sensibility analysis considering pyrolysis are always compared to the reference scenario, where pyrolysis is not considered.
Results show that pyrolysis has a bigger impact in the energy system when the CO2 limit is low. Biochar can be used to compensate the emissions from other conventional power plant and achieve an energy transition with lower costs. Furthermore, it was also found that pyrolysis can also reduce the need of flexibility. This study also shows that the biomass potential and the pyrolysis costs can affect a lot the behaviour of pyrolysis in the energy system.
One of the most critical areas of research and expansion has been exploiting new technologies in supply chain risk management. One example of this is the use of Digital Twins. The performance of physical systems can be analyzed and simulated using digital twins, virtual versions of these systems that use real-time data, and sophisticated algorithms. Inside the supply chain risk management field, digital twins present a one-of-a-kind opportunity to improve an organization's ability to anticipate, address, and react to the possibility of problems within the supply chain.
The objective of this study is to identify and assess the advantages that accrue to supply chain risk management as a result of Digital Twins' adoption into the system, as well as to identify the challenges associated with achieving those benefits. In the context of supply chain risk management, a thorough literature study is conducted to analyze the essential traits and capabilities of digital twins and how these qualities lead to enhanced risk management methods. This study investigates the essential properties and capacities of digital twins. In addition, the state of digital twin technology and its applications in supply chain risk management are evaluated, and prospective areas for further study and development are highlighted.
The primary purpose of this investigation is to provide a comprehensive and in-depth analysis of the digital twins' role in supply chain risk management through the utilization of digital twins, as well as to highlight the potential benefits and challenges associated with the implementation of digital twins. The research was carried out based on the existing body of written material and the replies of 27 individuals who had previous experience making use of digital twins and took part in an online questionnaire.
The results of this study will be relevant to a diverse group of stakeholders, including specialists in risk management and researchers, amongst others.
As the population grows, so does the amount of biowaste. As demand for energy grows, biogas is a promising solution to the problem. Lignocellulosic materials are challenged of slow degradability due to the presence of polymers such as cellulose, lignin and hemicellulose. There are several pretreatment methods available to enhance the degradability of such materials, including enzymatic pretreatment. In this pretreatment, there are few parameters that can influence the results, the most important being the enzyme to solid ratio and the solid to liquid ratio. During this project, experiments were conducted to determine the optimal conditions for those two factors. It was discovered that a solid to liquid ratio of 31 g of buffer per 1 gram of organic dry matter produced the highest reducing sugar release in flasks when combined with 34 mg of protein per 1 gram of organic dry mass. Additionally, another experiment was carried out to investigate the impact of enzymatic pretreatment on biogas production using artificial biowaste as a substrate. Artificial biowaste produced 577,9 NL/kg oDM, while enzymatically pretreated biowaste produced 639,3 NL/kg oDM. This resulted in a 10,6% rise in cumulative biogas production compared to its use without enzymatic pretreatment. By the conclusion of the investigation, specific cumulative dry methane yields of 364,7 NL/kg oDM and 426,3 NL/kg oDM were obtained from artificial biowaste without and with enzymatic pretreatment, respectively. This resulted in a methane production boost of 16,9%. Additionally in case of the reactors with enzymatically pretreated substrate kinetic constant was lower more than double, where maximum volume of biogas increased, comparing to the reactors without enzymatic pretreatment.
It is generally agreed that the development and deployment of an important amount of IoT devices throughout the world has revolutionized our lives in a way that we can rely on these devices to complete certain tasks that may have not been possible just years ago which also brought a new level of convenience and value to our lives.
This technology is allowing us in a smart home environment to remotely control doors, windows, and fridges, purchase online, stream music easily with the use of voice assistants such as Amazon Echo Alexa, also close a garage door from anywhere in the world to cite some examples as this technology has added value to several domains ranging from household environments, cites, industries by exchanging and transferring data between these devices and customers. Many of these devices’ sensors, collect and share information in real-time which enables us to make important business decisions.
However, these devices pose some risks and also some security and privacy challenges that need to be addressed to reach their full potential or be considered to be secure. That is why, comprehensive risk analysis techniques are essential to enhance the security posture of IoT devices as they can help evaluate the robustness and reliability towards potential susceptibility to risks, and vulnerabilities that IoT devices in a smart home setting might possess.
This approach relies on the basis of ISO/IEC 27005 methodology and risk matrix method to highlight the level of risks, impact, and likelihood that an IoT device in smart home settings can have, map the related vulnerability, threats and risks and propose the necessary mitigation strategies or countermeasures that can be taken to secure a device and therefore satisfying some security principles. Around 30 risks were identified on Amazon Echo and the related IoT system using the methodology. A detailed list of countermeasures is proposed as a result of the risk analysis. These results, in turn, can be used to elevate the security posture of the device.
The status quo of PROFINET, a commonly used industrial Ethernet standard, provides no inherent security in its communication protocols. In this thesis an approach for protecting real-time PROFINET RTC messages against spoofing, tampering and optionally information disclosure is specified and implemented into a real-world prototype setup. Therefor authenticated encryption is used, which relies on symmetric cipher schemes. In addition a procedure to update the used symmetric encryption key in a bumpless manner, e.g. without interrupting the real-time communication, is introduced and realized.
The concept for protecting the PROFINET RTC messages was developed in collaboration with a task group within the security working group of PROFINET International. The author of this thesis has also been part of that task group. This thesis contributes by proofing the practicability of the concept in a real-world prototype setup, which consists of three FPGA-based development boards that communicate with each other to showcase bumpless key updates.
To enable a bumpless key update without disturbing the deterministic real-time traffic by dedicated messages, the key update annunciation and status is embedded into the header. By provisioning two key slots, of which only one is in used, while the other is being prepared, a well-synchronized coordinated switch between the receiver and the sender performs the key update.
The developed prototype setup allows to test the concept and builds the foundation for further research and implementation activities, e.g. the impact of cryptographic operations onto the processing time.
Annotated training data is essential for supervised learning methods. Human annotation is costly and laborsome especially if a dataset consists of hundreds of thousands of samples and annotators need to be hired. Crowdsourcing emerged as a solution that makes it easier to get access to large amounts of human annotators. Introducing paid external annotators however introduces malevolent annotations, both intentional and unintentional. Both forms of malevolent annotations have negative effects on further usage of the data and can be summarized as spam. This work explores different approaches to post-hoc detection of spamming users and which kinds of spam can be detected by them. A manual annotation checking process resulted in the creation of a small user spam dataset which is used in this thesis. Finally an outlook for future improvements of these approaches will be made.
In this work, an implementation of the somewhat homomorphic BV encryption scheme is presented. During the implementation, care was taken to ensure that the resulting program will be as efficient as possible i.e. fast and resource-saving. The basis for this is the work of Arndt Bieberstein, who implemented the BV scheme with respect to functionality. The presented implementation supports the basics of the BV scheme, namely (symmetric and asymmetric) encryption, decryption and evaluation of addition as well as multiplication. Additionally, it supports the encoding of positive and negative numbers, various gaussian sampling methods, basically infinitely large polynomial coefficients, the generation of suitable parameters for a use case, threading and relinearization to reduce the size of a ciphertext after multiplications. After presenting the techniques used in the implementation, it’s actual efficiency is determined by measuring the timings of the operations for various parameters.