Refine
Document Type
- Master's Thesis (29)
- Bachelor Thesis (23)
- Conference Proceeding (4)
- Diploma Thesis (1)
- Final Report (1)
Year of publication
Keywords
- Maschinelles Lernen (58) (remove)
Institute
Diese Arbeit beschäftigt sich mit dem Erstellen semantischer Encodings von Bilddaten. Um diese Kodierungen aus den Daten zu extrahieren, wird ein künstliches neuronales Netzwerk auf
Videobild Interpolation trainiert. Die daraus erlernten Encodings sollen anschließend auf ihre Anwendbarkeit in einer anderen Aufgabe der KI gestützten Bildverarbeitung, der Extraktion von Landmarken auf Menschen, getestet werden.
Machine learning models for timeseries have always been a special topic of interest due to their unique data structure. Recently, the introduction of attention improved the capabilities of recurrent neural networks and transformers with respect to their learning tasks such as machine translation. However, these models are usually subsymbolic architectures, making their inner working hard to interpret without comprehensive tools. In contrast, interpretable models such learning vector quantization are more transparent in the ability to interpret their decision process. This thesis tries to merge attention as a machine learning function with learning vector quantization to better handle timeseries data. A design on such a model is proposed and tested with a dataset used in connection with the attention based transformers. Although the proposed model did not yield the expected results, this work outlines improvements for further research on this approach.
In this paper, we conduct experiments to optimize the learning rates for the Generalized Learning Vector Quantization (GLVQ) model. Our approach leverages insights from cog- nitive science rooted in the profound intricacies of human thinking. Recognizing that human-like thinking has propelled humankind to its current state, we explore the applica- bility of cognitive science principles in enhancing machine learning. Prior research has demonstrated promising results when applying learning rate methods inspired by cognitive science to Learning Vector Quantization (LVQ) models. In this study, we extend this approach to GLVQ models. Specifically, we examine five distinct cognitive science-inspired GLVQ variants: Conditional Probability (CP), Dual Factor Heuristic (DFH), Middle Symmetry (MS), Loose Symmetry (LS), and Loose Symme- try with Rarity (LSR). Our experiments involve a comprehensive analysis of the performance of these cogni- tive science-derived learning rate techniques across various datasets, aiming to identify optimal settings and variants of cognitive science GLVQ model training. Through this research, we seek to unlock new avenues for enhancing the learning process in machine learning models by drawing inspiration from the rich complexities of human cognition. Keywords: machine learning, GLVQ, cognitive science, cognitive bias, learning rate op- timization, optimizers, human-like learning, Conditional Probability (CP), Dual Factor Heuristic (DFH), Middle Symmetry (MS), Loose Symmetry (LS), Loose Symmetry with Rarity (LSR).
Analysis of Continuous Learning Strategies at the Example of Replay-Based Text Classification
(2023)
Continuous learning is a research field that has significantly boosted in recent years due to highly complex machine and deep learning models. Whereas static models need to be retrained entirely from scratch when new data get available, continuous models progressively adapt to new data saving computational resources. In this context, this work analyzes parameters impacting replay-based continuous learning approaches at the example of a data-incremental text classification task using an MLP and LSTM. Generally, it was found that replay improves the results compared to naive approaches but achieves not the performance of a static model. Mainly, the performances increased with more replayed examples, and the number of training iterations has a significant influence as it can partly control the stability-plasticity-trade-off. In contrast, the impact of balancing the buffer and the strategy to select examples to store in the replay buffer were found to have a minor impact on the results in the present case.
Die immer größer werdenden virtuellen Welten von Computerspielen mit spannendem und glaubhaftem Inhalt zu füllen, ohne die Entwicklungszeit enorm in die Höhe zu treiben, ist eine der großen Herausforderungen für Spielentwickler heutzutage. Eine Möglichkeit dieses Problem anzugehen ist der Einsatz computergestützter Generierungsalgorithmen um manuellen Aufwand zu verringern. Die vorliegende Arbeit befasst sich mit der Umsetzung dreier Data-To-Text-Ansätze zum Zweck der automatischen Generierung von Questtexten aus einer Datenstruktur in der Spieleentwicklungsumgebung Unity. Die entstehenden Implementierungen werden im Anschluss evaluiert auf Eignung für den Anwendungsfall. Folgende Methoden zur Realisierung der Texte werden angewandt: Templating, Templating mit Template-Generierung aus einer kontextfreien Grammatik, sowie die Oberflächenrealisierungsbibliothek SimpleNLG
Recently a deep neural network architecture designed to work on graph- structured data have been capturing notice as well as getting implemented in various domains and application. However, learning representation (feature embedding) from graphical data picking pace in research and constructing graph(s) from dataset remains a challenge. The ability to map the data to lower dimensions further makes the task easier while providing comfort in applying many operations. Graph neural network (GNN) is one of the novel neural network models that is catching attention as it is outperforming in various applications like recommender systems, social networks, chemical synthesis, and many more. This thesis discusses a unique approach for a fundamental task on graphs; node classification. The feature embedding for a node is aggregated by applying a Recurrent neural network (RNN), then a GNN model is trained to classify a node with the help of aggregated features and Q learning supports in optimizing the shape of neural networks. This thesis starts with the working principles of the Feedforward neural network, recurrent units like simple RNN, Long short-term memory (LSTM), and Gated recurrent unit (GRU), followed by concepts of Reinforcement learning (RL) and the Q learning algorithm. An overview of the fundamentals of graphs, followed by the GNN architecture and workflow, is discussed subsequently. Some basic GNN models are discussed in brief later before it approaches the technical implementation details, the output of the model, and a comparison with a few other models such as GraphSage and Graph attention network (GAN).
In vielen Einsatzbereichen sind digitale Nachbildungen realer Gebäude von großer Wichtigkeit. Die Erstellung dieser Nachbildungen erfordert bei älteren bzw. historischen Gebäuden allerdings meist erheblichen Vermessungs- und Nachbearbeitungsaufwand mit großem Personal- und Zeitbedarf. Häufig wurde ein Gebäude stilistisch an die jeweilige Zeit angepasst, sodass einzelne Zustände nur mit historischem Bildmaterial reproduzierbar sind.
Am Beispiel mehrerer ausgewählter, aktuell existierender Gebäude der Stadt Mittweida sind realitätsnahe, digitale und veränderbare Modelle mittels eines möglichst automatisierten Workflows erstellt.
Die äußere Erscheinung dieser Modelle kann mit dem entwickelten System automatisiert an andere Stile anpasst werden, welche durch z.B. historisches Bildmaterials von Gebäuden vorgegeben sind. Aufgrund der vielfältigen Einsatzbereiche und weiten Verbreitung finden hierfür Verfahren der Photogrammetrie für die Erstellung und neuronale Netze für die Stilanpassung Anwendung, welche auf handelsüblicher Hardware eingesetzt werden können. Eine Evaluierung erfolgte durch bildlichen Vergleich der stilangepassten Modelle mit dem zugehörigen Bildmaterial.
Diese Masterarbeit zeigt einen Ansatz zur Vorhersage von Zugverspätungen mit Hilfe von Supervised Learning. Dazu werden Modelle mit verschiedenen Algorithmen getestet und miteinander verglichen. Außerdem wird gezeigt, wie das entwickelte Vorhersagemodell in eine Blockchain-Anwendung integriert werden kann.
In dieser Arbeit werden die algorithmischen Grundlagen der Machine Learning Verfahren LVQ1 und LVQ3 erläutert. Für LVQ3 werden mehrere Ansätze zur Anpassung der Lernrate betrachtet, die anschließend verglichen werden sollen. Dazu werden vier verschiedene Experimente durchgeführt, wobei zwei Datensätze Verwendung finden, deren Ursprung in medizinischen Bilddaten liegt.
This thesis investigates the efficacy of four machine learning algorithms, namely linear regression, decision tree, random forest and neural network in the task of lead scoring. Specifically, the study evaluates the performance of these algorithms using datasets without sampling and with random under-sampling and over-sampling using SMOTE. The performance of each algorithm is measure using various performance metrics, including accuracy, AUC-ROC, specificity, sensitivity, precision, recall, F1 score, and G-mean. The results indicate that models trained on the dataset without sampling achieved higher accuracy than those trained on the dataset with either random under-sampling or random over-sampling using SMOTE. However, the neural network demonstrated remarkable results on each dataset compared to the other algorithms. These findings provide valuable insights into the effectiveness of machine learning algorithms for lead scoring tasks, particularly when using different sampling techniques. The findings of this study can aid lead management practices in selecting the most suitable algorithm and sampling technique for their needs. Furthermore, the study contributes to the literature by providing a comprehensive evaluation of the performance of machine learning algorithms for lead scoring tasks. This thesis has practical implications for businesses looking to improve their lead management practices, and future research could extend the analysis to other machine learning algorithms or more extensive datasets.