-
randomly initialized LSTMs: Results
-
LSTM learnability: Results
-
Transformer learnability (done after the paper was finished): Results
-
generating alternatives: Code for XLNet, Code for u-PMLM
-
creating RoBERTa predictions: see https://github.com/m-hahn/fairseq
-
results across tasks: GLUE, Parsing, Syntax, Text Classification
-
Sensitivity and length: Text Classification and CoLA, and By Task Group
-
per-input analysis: Sensitivity and Label Dispersion and Sensitivity and Accuracy
Online Experiments: RTE, SST-2, SST-2