Language Interpretability Tool (LIT)
The Language Interpretability Tool (LIT) is a visual, interactive model-understanding tool for NLP models.
natural-language-processing interpretability library code

The Language Interpretability Tool (LIT) is a visual, interactive model-understanding tool for NLP models.

LIT is built to answer questions such as:

  • What kind of examples does my model perform poorly on?
  • Why did my model make this prediction? Can this prediction be attributed to adversarial behavior, or to undesirable priors in the training set?
  • Does my model behave consistently if I change things like textual style, verb tense, or pronoun gender?

Don't forget to tag @PAIR-code in your comment, otherwise they may not be notified.

Authors community post
Code repositories for projects from the People+AI Research (PAIR) Initiative
Share this project
Similar projects
Lda2vec: Tools for interpreting natural language
The lda2vec model tries to mix the best parts of word2vec and LDA into a single framework.
AllenNLP Interpret
A Framework for Explaining Predictions of NLP Models
BertViz
Tool for visualizing attention in the Transformer model (BERT, GPT-2, Albert, XLNet, RoBERTa, CTRL, etc.)
Visualizing Memorization in RNNs
Inspecting gradient magnitudes in context can be a powerful tool to see when recurrent units use short-term or long-term contextual understanding.