Certified Adversarial Robustness With Domain Constraints

DSpace Repositorium (Manakin basiert)

Zur Kurzanzeige

dc.contributor.advisor Hein, Matthias (Prof. Dr.)
dc.contributor.author Voráček, Václav
dc.date.accessioned 2025-02-06T13:46:08Z
dc.date.available 2025-02-06T13:46:08Z
dc.date.issued 2025-02-06
dc.identifier.uri http://hdl.handle.net/10900/161636
dc.identifier.uri http://nbn-resolving.org/urn:nbn:de:bsz:21-dspace-1616366 de_DE
dc.identifier.uri http://dx.doi.org/10.15496/publikation-102968
dc.description.abstract Deep learning has become a dominant technique in machine learning and with only a little exaggeration it became a synonym to machine learning itself. Despite its great performance on tasks ranging from image classification to text generation, the underlying mechanism remains largely not understood and the deep-learning systems are a black-box, yielding some undesirable properties, such as the presence of adversarial examples. An adversarial example is a tiny modification of an input (usually demonstrated for images) that is imperceivable to humans and does not change the semantics of the input, however, the classifier is fooled and changes output to some absurd value. This is of a major concern in safety-critical applications, such as for autonomous driving where the consequence of such adversarial manipulations are potentially catastrophic. In this thesis, we continue in the effort to mitigate the problem. Namely, we first observe that the problem is not only present for deep learning, but also for simpler classifiers, such as nearest prototype classifiers. In that case, we derive rigorous mathematical guarantees about the robustness and provide tractable lower-bounds for the robustness. Despite using simpler models, this allowed us to establish state-of-the-art results on a popular benchmark. Later, we focus on randomized smoothing, which is a method certifying the robustness of a classifier to adversarial perturbations. In simple terms, in randomized smoothing we add noise to the input many times and output the majority vote over the outputs of the classifier for the noisy inputs. We present three separate contributions to this topic. First, we show that the standard implementation of this procedure does not actually yield the guarantees due to floating point errors and develop a fix. Second, we improve the performance of this technique in a certain setting. Third, we speed up the certification procedure. en
dc.language.iso en de_DE
dc.publisher Universität Tübingen de_DE
dc.rights ubt-podno de_DE
dc.rights.uri http://tobias-lib.uni-tuebingen.de/doku/lic_ohne_pod.php?la=de de_DE
dc.rights.uri http://tobias-lib.uni-tuebingen.de/doku/lic_ohne_pod.php?la=en en
dc.subject.classification Maschinelles Lernen de_DE
dc.subject.ddc 004 de_DE
dc.title Certified Adversarial Robustness With Domain Constraints en
dc.type PhDThesis de_DE
dcterms.dateAccepted 2025-01-24
utue.publikation.fachbereich Informatik de_DE
utue.publikation.fakultaet 7 Mathematisch-Naturwissenschaftliche Fakultät de_DE
utue.publikation.noppn yes de_DE

Dateien:

Das Dokument erscheint in:

Zur Kurzanzeige