DISSERTATION

Interpretable recurrent neural networks in continuous-time control environments

Ramin Hasani

Year: 2020 University:   reposiTUm (TU Wien)   Publisher: TU Wien

Abstract

Intelligent agents must learn coherent representations of their world, from high-dimensional sensory information, and utilize them to generalize well in unseen situations. Although contemporary deep learning algorithms have achieved noteworthy successes in variform of high-dimensional tasks, their learned causal structure, interpretability, and robustness were largely overlooked. This dissertation presents methods to address interpretation, stability and the overlooked properties of a class of intelligent algorithms, namely recurrent neural networks (RNNs), in continuous-time environments. Accordingly, the contributions of the work lie into two major frameworks: I) Designing interpretable RNN architectures — We first introduce a novel RNN instance that is formulated by computational models originally developed to explain the nervous system of small species. We call these RNNs liquid time-constant (LTCs) because they possess nonlinear compartments that regulate the state of a neuron through a variable time-constant. LTCs form a dynamic causal model capable of learning causal relationships between the input, their neural state, and the output dynamics directly from supervised training data. Moreover, we demonstrate that LTCs are universal approximators and can be advantageously used in continuous-time control domains. We then combine LTCs with contemporary scalable deep neural network architectures and structural inspirations from the C. elegans connectome, to develop novel neural processing units, that can learn to map multidimensional inputs to control commands by sparse, causal, interpretable and robust neural representations. We extensively evaluate the performance of LTC-based neural network instances in a large category of simulated and real-world applications ranging from time-series classification and prediction to autonomous robot and vehicle control. II) Designing interpretation methods for trained RNN instances — In this framework, we develop a quantitative method to interpret the dynamics of modern RNN architectures. As opposed to the existing methods that are proactively constructed by empirical feature visualization algorithms, we propose a systematic pipeline for interpreting individual hidden state dynamics within the network using response characterization methods. Our method is able to uniquely identify neurons with insightful dynamics, quantify relationships between dynamical properties and test accuracy through ablation analysis, and interpret the impact of network capacity on a network’s dynamical distribution. Finally, we demonstrate the scalability of our method by evaluating a series of different benchmark sequential datasets. The findings of this dissertation notably improves our understanding of neural information processing systems in continuous-time environments.

Keywords:
Recurrent neural network Artificial neural network Control (management) Computer science Artificial intelligence

Metrics

7
Cited By
0.44
FWCI (Field Weighted Citation Impact)
0
Refs
0.69
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Neural Networks and Applications
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Group of Continuous Time Recurrent Neural Networks

Neeraj SahuAdwitiya Sinha

Journal:   International Journal of Advanced Research in Computer Science Year: 2012 Vol: 3 (3)Pages: 622-625
JOURNAL ARTICLE

Continuous attractors of discrete-time recurrent neural networks

Jiali YuHuajin TangHaizhou Li

Journal:   Neural Computing and Applications Year: 2012 Vol: 23 (1)Pages: 89-96
JOURNAL ARTICLE

Self-Optimization in Continuous-Time Recurrent Neural Networks

Mario ZarcoTom Froese

Journal:   Greater South Information System Year: 2018
© 2026 ScienceGate Book Chapters — All rights reserved.