Skip to content

astrajoan/sequential-learning-interpretation

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

46 Commits
 
 

Repository files navigation

sequential learning interpretation

Deep learning, and in particular, recurrent neural networks, has in recent years gained nonstop interest with its successful application in a broad range of areas. These include handwriting recognition, natural language processing, speed recognition and so on. However, with the ever expanding use of such models, their interpretability or the mechanism of their decision making process have been understudied. Such interpretation can not only help users trust the models and predictions more, but also provide valuable insights into various areas, such as genetic modeling and linguistics, and help with model designs.

Here, we organized papers and articles from difference sources to provide a somewhat full-around overview of developments in this area.

directions in sequential learning interpretation

definition

to start with

papers

articles

controversies

  • Attention is not Explanation (arXiv, 2019)
    keywords: RNN, BiLSTM, binary text classification, question answering, feature importance, Kendall τ correlation, counterfactual attention weights, adversarial attention
  • Attention is not not Explanation (arXiv, 2019)
    keywords: LSTM, binary text classification, uniform attention weights, model variance, MLP diagnostic tool, model-consistent adversarial training, TVD/JSD plots

based on interpretation methods

SHAP

composition

gradient

backpropagation

attention

saliency

erasure

prototypes

interpretable model

adversarial training

counterfactual

interpretation evaluation metrics

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published