Hauptseite

Aus IPD-Institutsseminar
Version vom 25. Juli 2018, 13:59 Uhr von Erik.burger@kit.edu (Diskussion | Beiträge)
(Unterschied) ← Nächstältere Version | Aktuelle Version (Unterschied) | Nächstjüngere Version → (Unterschied)
Wechseln zu: Navigation, Suche

Das Institutsseminar des Instituts für Programmstrukturen und Datenorganisation (IPD) ist eine ständige Lehrveranstaltung, die den Zweck hat, über aktuelle Forschungsarbeiten am Institut zu informieren. Insbesondere soll Studierenden am Institut die Gelegenheit gegeben werden, über ihre Bachelor- und Masterarbeiten vor einem größeren Auditorium zu berichten. Schwerpunkte liegen dabei auf der Problemstellung, den Lösungsansätzen und den erzielten Ergebnissen. Das Seminar steht aber allen Studierenden und Mitarbeiter/-innen des KIT sowie sonstigen Interessierten offen.

Ort Gebäude 50.34, Seminarraum 348
Zeit jeweils freitags, 11:30–13:00 Uhr

Die Vorträge müssen den folgenden zeitlichen Rahmen einhalten:

  • Masterarbeit: 30 Minuten Redezeit + 15 Minuten Diskussion
  • Bachelorarbeit: 20 Minuten Redezeit + 10 Minuten Diskussion
  • Proposal: 12 Minuten Redezeit + 8 Minuten Diskussion

Weitere Informationen: https://sdqweb.ipd.kit.edu/wiki/Institutsseminar. Bei Fragen und Anmerkungen können Sie eine E-Mail an das Institutsseminar-Team schreiben.

Nächste Vorträge

Freitag, 14. Dezember 2018, 11:30 Uhr, Raum 301 (Gebäude 50.34)
Vortragende(r) Florian Hennerich
Titel Erkennung Semantischer Wortveränderungen auf Textströmen
Vortragstyp Masterarbeit
Betreuer(in) Adrian Englhardt
Kurzfassung Die natürliche Sprache befindet sich in ständigem Wandel. Mittels Semantic Change Detection kann eine Änderung der Semantik von Wörtern zwischen Zeitpunkten festgestellt werden. Herkömmliche Semantic Change Detection Systeme arbeiten nur auf statischen Korpora. Durch Social Media ist es möglich, Sprache in Echtzeit zu analysieren. Da bisherige Ansätze jedoch nicht auf Textströmen funktionieren, soll in dieser Masterarbeit ein Echtzeitsystem zur Verarbeitung von Textströmen entworfen werden, welches frühzeitig die Änderung einer Wortbedeutung aufzeigt. Grundlage hierfür sind geeignete Worteinbettungen, die zum einen gute Vektoren liefern und zum anderen trotz Begrenzung des Speichers den Textstrom gut repräsentieren. Zur Evaluation soll ein synthetischer Korpus generiert werden, um die verschiedenen Methoden vergleichen zu können. Anschließend wird eine explorative Untersuchung auf Twitterdaten durchgeführt.
Vortragende(r) Alexander Poth
Titel Statistical Generation of High Dimensional Data Streams with Complex Dependencies
Vortragstyp Bachelorarbeit
Betreuer(in) Edouard Fouché
Kurzfassung The evaluation of data stream mining algorithms is an important task in current research. The lack of a ground truth data corpus that covers a large number of desireable features (especially concept drift and outlier placement) is the reason why researchers resort to producing their own synthetic data. This thesis proposes a novel framework ("streamgenerator") that allows to create data streams with finely controlled characteristics. The focus of this work is the conceptualization of the framework, however a prototypical implementation is provided as well. We evaluate the framework by testing out data streams against state-of-the-art dependency measures and outlier detection algorithms.
Freitag, 14. Dezember 2018, 11:30 Uhr, Raum 348 (Gebäude 50.34)
Vortragende(r) Jan-Philipp Jägers
Titel Iterative Performance Model Parameter Estimation Considering Parametric Dependencies
Vortragstyp Masterarbeit
Betreuer(in) Manar Mazkatli
Kurzfassung A main issue of using performance models, is the effort it takes to keep them updated. They must stay in synchronization with the actual system architecture and source code, in order to produce meaningful results. To address this shortcoming, there exist techniques, which extract the structure of a system and derive a performance model from existing artifacts, for example, from source code. Existing approaches estimate performance model parameters like loop iterations, branch transitions, resource demands and external call arguments for a whole system at once. We present an approach which estimates parameters of such performance models iteratively. We use monitoring data to estimate performance model parameters which are affected by an iterative source code change. We use a decision tree to build a predictive model for branch transitions and regression analysis to build a predictive model for loop iterations and resource demands. These predictive models can include dependency relations to service call arguments. To estimate resource demands iteratively, we must know all resource demands that are executed in a system. We monitor only a part of a system during one iteration and must estimate which resource demands are executed, but not monitored. We use the previously estimated performance model parameters and the control flow information of services to traverse the control flow in order to detect non-monitored resource demands.
Vortragende(r) Torsten Syma
Titel Multi-model Consistency through Transitive Combination of Binary Transformations
Vortragstyp Masterarbeit
Betreuer(in) Heiko Klare
Kurzfassung Software systems are usually described through multiple models that address different development concerns. These models can contain shared information, which leads to redundant representations of the same information and dependencies between the models. These representations of shared information have to be kept consistent, for the system description to be correct. The evolution of one model can cause inconsistencies with regards to other models for the same system. Therefore, some mechanism of consistency restoration has to be applied after changes occurred. Manual consistency restoration is error-prone and time-consuming, which is why automated consistency restoration is necessary. Many existing approaches use binary transformations to restore consistency for a pair of models, but systems are generally described through more than two models. To achieve multi-model consistency preservation with binary transformations, they have to be combined through transitive execution.

In this thesis, we explore transitive combination of binary transformations and we study what the resulting problems are. We develop a catalog of six failure potentials that can manifest in failures with regards to consistency between the models. The knowledge about these failure potentials can inform a transformation developer about possible problems arising from the combination of transformations. One failure potential is a consequence of the transformation network topology and the used domain models. It can only be avoided through topology adaptations. Another failure potential emerges, when two transformations try to enforce conflicting consistency constraints. This can only be repaired through adaptation of the original consistency constraints. Both failure potentials are case-specific and cannot be solved without knowing which transformations will be combined. Furthermore, we develop two transformation implementation patterns to mitigate two other failure potentials. These patterns can be applied by the transformation developer to an individual transformation definition, independent of the combination scenario. For the remaining two failure potentials, no general solution was found yet and further research is necessary.

We evaluate the findings with a case study that involves two independently developed transformations between a component-based software architecture model, a UML class diagram and its Java implementation. All failures revealed by the evaluation could be classified with the identified failure potentials, which gives an initial indicator for the completeness of our failure potential catalog. The proposed patterns prevented all failures of their targeted failure potential, which made up 70% of all observed failures, and shows that the developed implementation patterns are applicable and help to mitigate issues occurring from transitively combining binary transformations.