Refine
Document Type
- Bachelor Thesis (1)
- Conference Proceeding (1)
- Master's Thesis (1)
- Working Paper (1)
Conference Type
- Konferenzartikel (1)
Keywords
- Computersicherheit (1)
- Deepfake (1)
- IT-Sicherheit (1)
- PQC (1)
- Post-Quanten Kryptographie (1)
- Public-Key-Verschlüsselung (1)
- Quantencomputer (1)
- Tor (1)
- Verschlüsselung (1)
Institute
Open Access
- Closed (2)
- Closed Access (1)
- Open Access (1)
Durch die Fortschritte im Bereich der Quantencomputer rückt der Zeitpunkt näher, dass Quantencomputer die bestehenden mathematischen Probleme lösen können, welche in den aktuellen Public-Key-Verschlüsselungsverfahren verwendet werden. Als Reaktion darauf wurde ein Standardisierungsprozess für quantensichere Public-Key-Verschlüsselungsverfahren gestartet. Diese Arbeit analysiert diese und vergleicht sie untereinander, um Stärken und Schwächen der einzelnen Verfahren aufzuzeigen.
Synthesizing voice with the help of machine learning techniques has made rapid progress over the last years. Given the current increase in using conferencing tools for online teaching, we question just how easy (i.e. needed data, hardware, skill set) it would be to create a convincing voice fake. We analyse how much training data a participant (e.g. a student) would actually need to fake another participants voice (e.g. a professor). We provide an analysis of the existing state of the art in creating voice deep fakes and align the identified as well as our own optimization techniques in the context of two different voice data sets. A user study with more than 100 participants shows how difficult it is to identify real and fake voice (on avg. only 37% can recognize a professor’s fake voice). From a longer-term societal perspective such voice deep fakes may lead to a disbelief by default.
Synthesizing voice with the help of machine learning techniques has made rapid progress over the last years [1]. Given the current increase in using conferencing tools for online teaching, we question just how easy (i.e. needed data, hardware, skill set) it would be to create a convincing voice fake. We analyse how much training data a participant (e.g. a student) would actually need to fake another participants voice (e.g. a professor). We provide an analysis of the existing state of the art in creating voice deep fakes and align the identified as well as our own optimization techniques in the context of two different voice data sets. A user study with more than 100 participants shows how difficult it is to identify real and fake voice (on avg. only 37 percent can recognize a professor’s fake voice). From a longer-term societal perspective such voice deep fakes may lead to a disbelief by default.