004 Informatik
Refine
Year of publication
Document Type
- Bachelor Thesis (51)
- Master's Thesis (34)
- Conference Proceeding (6)
- Contribution to a Periodical (5)
- Article (reviewed) (2)
- Doctoral Thesis (2)
- Book (1)
Conference Type
- Konferenzartikel (6)
Keywords
- IT-Sicherheit (12)
- JavaScript (9)
- Deep learning (5)
- Blockchain (4)
- Computersicherheit (4)
- E-Learning (4)
- HTML 5.0 (4)
- Internet der Dinge (4)
- Internet of Things (4)
- Künstliche Intelligenz (3)
Institute
- Fakultät Medien (M) (ab 22.04.2021) (37)
- Fakultät Medien und Informationswesen (M+I) (bis 21.04.2021) (36)
- Fakultät Elektrotechnik, Medizintechnik und Informatik (EMI) (ab 04/2019) (19)
- Fakultät Elektrotechnik und Informationstechnik (E+I) (bis 03/2019) (6)
- ivESK - Institut für verlässliche Embedded Systems und Kommunikationselektronik (4)
- Zentrale Einrichtungen (3)
- Fakultät Wirtschaft (W) (2)
- IMLA - Institute for Machine Learning and Analytics (1)
Open Access
- Closed Access (47)
- Open Access (28)
- Closed (26)
- Diamond (3)
The interaction between agents in multiagent-based control systems requires peer to peer communication between agents avoiding central control. The sensor nodes represent agents and produce measurement data every time step. The nodes exchange time series data by using the peer to peer network in order to calculate an aggregation function for solving a problem cooperatively. We investigate the aggregation process of averaging data for time series data of nodes in a peer to peer network by using the grouping algorithm of Cichon et al. 2018. Nodes communicate whether data is new and map data values according to their sizes into a histogram. This map message consists of the subintervals and vectors for estimating the node joining and leaving the subinterval. At each time step, the nodes communicate with each other in synchronous rounds to exchange map messages until the network converges to a common map message. The node calculates the average value of time series data produced by all nodes in the network by using the histogram algorithm. The relative error for comparing the output of averaging time series data, and the ground truth of the average value in the network will decrease as the size of the network increases. We perform simulations which show that the approximate histograms method provides a reasonable approximation of time series data.
Bereichsübergreifender Einsatz von JavaScript – Aktueller Stand und exemplarische Implementierung
(2021)
Nahezu alle Websites nutzen die Programmiersprache JavaScript zur Darstellung von interaktiven Inhalten und zur Bereitstellung von komplexen Funktionalitäten. Seit ihren Anfängen im Jahr 1995 hat sich die Sprache nicht nur zum Standard in der Webentwicklung etabliert, sondern auch zu einer leistungsfähigen Mehrzweckprogrammiersprache weiterentwickelt.
Diese Arbeit befasst sich mit einer ausführlichen Darstellung der aktuellen Möglichkeiten, welche Ansätze sich durch die Weiterentwicklung JavaScripts zu einer Mehrzweckprogrammiersprache ergeben und wie sich diese heute umsetzen lassen. Anhand des intelligenten Schlüsselkastens „Smart Vault“ wird verdeutlicht, wie dieses Vorgehen praktisch realisiert und die Vorteile einer einzigen Programmiersprache angewendet werden können.
Es hat sich herausgestellt, dass sich JavaScript für Anwendungen unterschiedlicher Bereiche einsetzen lässt und darüber hinaus ein hohes Potenzial für weitere Entwicklungen, Verbesserungen und zusätzliche Einsatzgebiete besitzt. Es lassen sich nicht nur Websites, Web Server und Desktop Apps, sondern auch Mikrocontroller im Internet of Things konfigurieren und miteinander nutzen, ohne eine weitere Programmiersprache zu benötigen. Zahlreiche Bibliotheken und Frameworks machen es möglich, dass die Sprache verschiedene Anwendungen über ihre Einsatzgebiete hinweg miteinander verbindet.
As e-commerce platforms have grown in popularity, new difficulties have emerged, such as the growing use of bots—automated programs—to engage with e-commerce websites. Even though some algorithms are helpful, others are malicious and can seriously hurt e-commerce platforms by making fictitious purchases, posting fictitious evaluations, and gaining control of user accounts. Therefore, the development of more effective and precise bot identification systems is urgently needed to stop such actions. This thesis proposes a methodology for detecting bots in E-commerce using machine learning algorithms such as K-nearest neighbors, Decision Tree, Random Forest, Support Vector Machine, and Neural Network. The purpose of the research is to assess and contrast the output of these machine learning methods. The suggested approach will be based on data that is readily accessible to the public, and the study’s focus will be on the research of bots in e-commerce.
The purpose of the study is to provide an overview of bots in e-commerce, as well as information on the different kinds and traits of bots, as well as current research on bots in e-commerce and associated work on bot detection in e-commerce. The research also seeks to create a more precise and effective bot detection system as well as find critical factors in detecting bots in e-commerce.
This research is significant because it sheds light on the increasing issue of bots in e-commerce and the requirement for more effective bot detection systems. The suggested approach for using machine learning algorithms to identify bots in ecommerce can give e-commerce platforms a more precise and effective bot detection system to stop malicious bot activities. The study’s results can also be used to create a more effective bot detection system and pinpoint key elements in detecting bots in e-commerce.
Organizations striving to achieve success in the long term must have a positive brand image which will have direct implications on the business. In the face of the rising cyber threats and intense competition, maintaining a threat-free domain is an important aspect of preserving that image in today's internet world. Domain names are often near-synonyms for brand names for numerous companies. There are likely thousands of domains that try to impersonate the big companies in a bid to trap unsuspecting users, usually falling prey to attacks such as phishing or watering hole. Because domain names are important for organizations for running their business online, they are also particularly vulnerable to misuse by malicious actors. So, how can you ensure that your domain name is protected while still protecting your brand identity? Brand Monitoring, for example, may assist. The term "Brand Monitoring" applies only to keep tabs on an organization's brand performance, reception, and overall online presence through various online channels and platforms [1]. There has been a rise in the need of maintaining one's domain clear of any linkages to malicious activities as the threat environment has expanded. Since attackers are targeting domain names of organizations and luring unsuspecting users to visit malicious websites, domain monitoring becomes an important aspect. Another important aspect of brand abuse is how attackers leverage brand logos in creating fake and phishing web pages. In this Master Thesis, we try to solve the problem of classification of impersonated domains using rule-based and machine learning algorithms and automation of domain monitoring. We first use a rule-based classifier and Machine Learning algorithms to classify the domains gathered into two buckets – "Parked" and "Non-Parked". In the project's second phase, we will deploy object detection models (Scale Invariant Feature Transform - SIFT and Multi-Template Matching – MTM) to detect brand logos from the domains of interest.
In dieser Forschungsarbeit wird die Datensicherheit von Microsoft Azure analysiert und bewertet. Die Bewertung findet dabei aus der Sicht von Unternehmen statt. Im ersten Abschnitt wird zunächst der grundlegende Aufbau und die unterschiedlichen Formen des Cloud Computing beschrieben. Im zweiten Teil wird ein Vergleich der drei größten Cloud Anbieter vollzogen. Der letzte Teil besteht aus der Evaluation der Datensicherheit von Azure, wobei auf Aspekte wie Datenschutz, Bedrohungen und Schutzmaßnahmen eingegangen wird. Abschließend wird eine Empfehlung für das Unternehmen Bechtle GmbH Offenburg IT-Systemhaus abgegeben.
Im Verlauf der Arbeit stellt sich heraus, dass Azure eine ausreichende Datensicherheit bieten kann. Allerdings wird deutlich, dass durch die Kombination von mehreren Nebenfaktoren wie das Patch-Verhalten oder die Antwortzeit auf Sicherheitsschwachstellen seitens Microsofts, eine große Gefahr für die Daten von Unternehmen entstehen kann. Demnach ist Microsoft als Anbieter ein größeres Problem für die Sicherheit von Daten in Azure als der Cloud-Dienst selbst.
Conceptualization and implementation of automated optimization methods for private 5G networks
(2023)
Today’s companies are adjusting to the new connectivity realities. New applications require more bandwidth, lower latency, and higher reliability as industries become more distributed and autonomous. Private 5th Generation (5G) networks known as 5G Non-Public Networks (5G-NPN), is a novel 3rd Generation Partnership Project (3GPP)- based 5G network that can deliver seamless and dedicated wireless access for a particular industrial use case by providing the mentioned application’s requirements. To meet these requirements, several radio-related aspects and network parameters should be considered. In many cases, the behavior of the link connection may vary based on wireless conditions, available network resources, and User Equipment (UE) requirements. Furthermore, Optimizing these networks can be a complex task due to the large number of network parameters and KPIs that need to be considered. For these reasons, traditional solutions and static network configuration are not affordable or simply impossible. Despite the existence of papers in the literature that address several optimization methods for cellular networks in industrial scenarios, more insight into these existing but complex or unknown methods is needed.
In this thesis, a series of optimization methods were implemented to deliver an optimal configuration solution for a 5G private network. To facilitate this implementation, a testing system was implemented. This system enables remote control over the UE and 5G network, establishment of a test environment, extraction of relevant KPI reports from both UE and network sides, assessment of test results and KPIs, and effective utilization of the optimization and sampling techniques.
The research highlights the advantageous aspects of automated testing by using OFAT, Simulated Annealing, and Random Forest Regressor methods. With OFAT, as a common sampling method, a sensitivity analysis and an impact of each single parameter variation on the performance of the network were revealed. With Simulated Annealing, an optimal solution with MSE of roughly 10 was revealed. And, in the Random Forest Regressor, it was seen that this method presented a significant advantage over the simulated annealing method by providing substantial benefits in time efficiency due to its machine- learning capability. Additionally, it was seen that by providing a larger dataset or using some other machine-learning techniques, the solution might be more accurate.
Immer mehr Unternehmen setzen auf eine Cross-Cloud-Strategie, die es Unternehmen ermöglicht, ihre Anwendungen und Daten über mehrere Cloud-Plattformen hinweg effizient zu verwalten und zu betreiben. Konsistenz und Atomarität zwischen den Cloud-Plattformen zu wahren, stellt eine große Herausforderung dar. Hierzu wird in dieser Arbeit eine Lösung vorgestellt, um Cross-Cloud-Atomarität zu erreichen, welche auf Basis des 2-Phasen-Commit-Protokolls (2PC) beruht. In diesem Zusammenhang wird die Funktionsweise des 2PC-Protokolls erörtert und Erweiterungen sowie Alternativen zum Protokoll kurz angesprochen. Zusätzlich werden alternative Lösungsansätze diskutiert, die für die Erzielung von Cross-Cloud-Atomarität in Betracht gezogen werden können. Dadurch wird ein umfassender Einblick in das Thema sowie mögliche Lösungsansätze für diese Herausforderung gewährt.
The identification of vulnerabilities is an important element of the software development process to ensure the security of software. Vulnerability identification based on the source code is a well studied field. To find vulnerabilities on the basis of a binary executable without the corresponding source code is more challenging. Recent research has shown how such detection can be performed statically and thus runtime efficiently by using deep learning methods for certain types of vulnerabilities.
This thesis aims to examine to what extent this identification can be applied sufficiently for a variety of vulnerabilities. Therefore, a supervised deep learning approach using recurrent neural networks for the application of vulnerability detection based on binary executables is used. For this purpose, a dataset with 50,651 samples of 23 different vulnerabilities in the form of a standardised LLVM Intermediate Representation was prepared. The vectorised features of a Word2Vec model were then used to train different variations of three basic architectures of recurrent neural networks (GRU, LSTM, SRNN). For this purpose, a binary classification was trained for the presence of an arbitrary vulnerability, and a multi-class model was trained for the identification of the exact vulnerability, which achieved an out-of-sample accuracy of 88% and 77%, respectively. Differences in the detection of different vulnerabilities were also observed, with non-vulnerable samples being detected with a particularly high precision of over 98%. Thus, the methodology presented allows an accurate detection of vulnerabilities, as well as a strong limitation of the analysis scope for further analysis steps.
Das tiefe Lernen und die daraus entstehenden Technologien bieten eine neue Herausforderung für Unternehmen und privat Personen beiderlei. Deepfakes sind schon seit über vier Jahren im Internet verbreitet und in dieser Zeit wurden hauptsächlich politische Figuren Opfer der Technologie. Diese Arbeit nimmt sich das Ziel, die möglichen Attacken zu beschreiben und Gegenmaßnahmen dafür vorzustellen. Es wird zunächst Social Engineering erläutert und die technischen Grundlagen von Deepfakes gelegt. Daraufhin folgt ein Fallbeispiel, welches genauer aufzeigt, wie auch Unternehmen Opfer von Deepfake Attacken werden können. Diese Attacken fügen einen erheblichen finanziellen sowie Reputationsschaden an. Daher müssen verschiedene technische und organisatorische Maßnahmen gegenüber Deepfakes im Social Engineering Umfeld eingeführt werden. Durch die ständige Entwicklung der Technik werden diese Attacken in der Zukunft an Komplexität und Häufigkeit zunehmen. Unternehmen, Forscher und IT-Sicherheitsspezialisten müssen daher die Entwicklung dieser Attacken beobachten.