Detailseite
Projekt Druckansicht

Lernkonzepte in tiefen neuronalen Netzen

Fachliche Zuordnung Bild- und Sprachverarbeitung, Computergraphik und Visualisierung, Human Computer Interaction, Ubiquitous und Wearable Computing
Softwaretechnik und Programmiersprachen
Förderung Förderung von 2012 bis 2017
Projektkennung Deutsche Forschungsgemeinschaft (DFG) - Projektnummer 227351812
 
Erstellungsjahr 2019

Zusammenfassung der Projektergebnisse

Our work has led to a number of insights into the representations of deep networks, such as the emergence of abstract representations in deep Boltzmann machines (DBMs), the specific role of convolution and pooling layers in neural networks, a consolidated understanding of layer-wise oscillations1 in stacked architectures, as well as the effect of training parameters and loss functions on the qualitative properties of learned representations. In addition to characterizing representations quantitatively, we also aimed at characterizing interaction between representations by identifying modes of interaction such as “preserve” or “complement”, which we further quantified in the context of Boltzmann machine architectures with a newly proposed “layer interaction number”. While our analysis was instrumental at characterizing the representation from small up to mid-scale machine learning models, we observed that due to the curse of dimensionality our kernel RDE analysis of deep networks may not fully identify the properties of a deep representation when the latter is high-dimensional. When analyzing high-dimensional representations, we observed that the aforementioned analysis should preferably be performed locally in the input domain. Our newly developed local RDE analysis (LRDE) achieves this localization. Furthermore, we used the LRDE technique to improve estimates of predictive uncertainty in kernel regression models. While at the beginning of the project, research on deep learning was mainly focused on learning unsupervised representations through techniques such as Boltzmann machines, or auto-encoders, there has been stunning progress on learning large-scale neural networks with GPU-based implementations, which have become state-of-the-art methodology in image recognition or natural language processing. Hence, in the second part of the project, we focused our research effort on understanding deep representations in these new state-of-the-art models. Specifically, we considered the extraction of human-interpretable insights from deep representations. To achieve this, we developed the layer-wise relevance propagation (LRP) method that can explain the predictions of complex state-of-the-art image and text classifiers in terms of input variables. The method was later given theoretical support by viewing the classifier as a composition of multiple functions (one per neuron), and performing a “deep Taylor decomposition” (DTD) of that composition of functions. Recent advances in the field of interpreting deep networks were summarized in a tutorial paper. Our work on interpreting deep representations was presented at numerous conferences and workshops in the form of research talks or tutorials. Furthermore, our work on analyzing representations has served as a source of inspiration, when applying machine learning to practical problems in chemistry and biology.

Projektbezogene Publikationen (Auswahl)

 
 

Zusatzinformationen

Textvergrößerung und Kontrastanpassung