@christinawarren

Inner-Process Visualization of Hidden States in Recurrent Neural Networks

, and . Proceedings of the 13th International Symposium on Visual Information Communication and Interaction, New York, NY, USA, Association for Computing Machinery, (2020)
DOI: 10.1145/3430036.3430047

Abstract

In this paper, we introduce a visualization technique aimed to help machine learning experts to analyze the hidden states of layers in recurrent neural networks (RNNs). Our technique allows the user to visually inspect how hidden states store and process information throughout the feeding of an input sequence into the network. It can answer questions such as which parts of the input data had a higher impact on the prediction and how the model correlates each hidden state configuration with a certain output. Our visualization comprises several components: our input visualization shows the input sequence and how it relates to the output (using color coding); hidden states are visualized by nonlinear projection to 2-D visualization space via t-SNE in order to understand the shape of the space of hidden states; time curves are employed to show the details of the evolution of hidden state configurations; and a time-multi-class heatmap matrix visualizes the evolution of expected predictions for multi-class classifiers. To demonstrate the capability of our approach, we discuss two typical use cases for long short-term memory (LSTM) models applied to two widely used natural language processing (NLP) datasets.

Links and resources

Tags

community

  • @sfbtrr161
  • @christinawarren
@christinawarren's tags highlighted