Sam Greydanus | 2017 | MIT License
Karpathy, Johnson, and Fei-Fei
Curated abstract. Performance and limitations of RNNs are poorly understood. Using character-level language models as a testbed, authors analyze the internal representations, predictions and error types. Results include: 1) existence of interpretable cells that track long-range dependencies such as line lengths, quotes and brackets. 2) comparative analysis with finite horizon n-gram models shows LSTM improvements are due to longer time horizon. 3) analysis of the remaining errors with a pie chart error breakdown.
Math.
Pros.
Cons.
Montavon, Bach, Binder, Samek, Müller
Math.
Words.
Pros.
Cons.
Springenberg, Dosovitskiy, Brox, Riedmiller
Math.
Words.
Pros.
Cons.