Institutsseminar/2023-07-14 IPD Böhm

Aus SDQ-Institutsseminar
Termin (Alle Termine)
Datum Freitag, 14. Juli 2023
Uhrzeit 13:00 – 13:40 Uhr (Dauer: 40 min)
Ort Raum 348 (Gebäude 50.34)
Webkonferenz
Vorheriger Termin Fr 14. Juli 2023
Nächster Termin Fr 21. Juli 2023

Termin in Kalender importieren: iCal (Download)

Vorträge

Vortragende(r) David Schulmeister
Titel Hidden Outliers in Manifolds
Vortragstyp Proposal
Betreuer(in) Jose Cribeiro
Vortragsmodus in Präsenz
Kurzfassung Hidden outliers represent instances of disagreement between a full-space and an ensemble. This adversarial nature naturally replicates the subspace behavior that high-dimensional outliers exhibit in reality. Due to this, they have been proven useful for representing complex occurrences like fraud, critical infrastructure failure, and healthcare data, as well as for their use in general outlier detection as the positive class of a self-supervised learner. However, while interesting, hidden outliers' quality highly depends on the number of subspaces selected in the ensemble out of the total possible. Since the number of subspaces increases exponentially with the number of features, this makes high-dimensional applications of Data Analysis, such as Computer Vision, computationally unfeasible. In this thesis, we are going to study the generation of hidden outliers on the embedded data manifold using deep learning techniques to overcome this issue. More precisely, we are going to study the behavior, characteristics, and performance in multiple use-cases of hidden outliers in the data manifold.
Vortragende(r) Denis Wambold
Titel Subspace Generative Adversarial Learning for Unsupervised Outlier Detection
Vortragstyp Proposal
Betreuer(in) Jose Cribeiro
Vortragsmodus in Präsenz
Kurzfassung Outlier detection is an important yet challenging task, especially for unlabeled, high-dimensional, datasets. Due to their self-supervised generative nature, Generative Adversarial Networks (GAN) have proven themselves to be one of the most powerful deep learning methods for outlier detection. However, most state-of-the-art GANs for outlier detection share common limitations. Oftentimes we only achieve great results if the model’s hyperparameters are properly tuned or the underlying network structure is adjusted. This optimization is not possible in practice when the data is unlabeled. If not tuned properly, it is not unusual that a state-of-the-art GAN method is outperformed by simpler shallow methods.

We propose using a GAN architecture with feature ensemble learning to address hyperparameter sensibility and architectural dependency. This follows the success of feature ensembling in mitigating these problems inside other areas of Deep Learning. This thesis will study the optimization problem, training, and tuning of feature ensemble GANs in an unsupervised scenario, comparing it to other deep generative methods in a similar setting.

Neuen Vortrag erstellen

Hinweise