TY - GEN
T1 - Understanding Convolutional Neural Networks for Text Classification
AU - Jacovi, Alon
AU - Shalom, Oren Sar
AU - Goldberg, Yoav
N1 - Publisher Copyright:
© 2018 Association for Computational Linguistics
PY - 2018/1/1
Y1 - 2018/1/1
N2 - We present an analysis into the inner workings of Convolutional Neural Networks (CNNs) for processing text. CNNs used for computer vision can be interpreted by projecting filters into image space, but for discrete sequence inputs CNNs remain a mystery. We aim to understand the method by which the networks process and classify text. We examine common hypotheses to this problem: that filters, accompanied by global max-pooling, serve as ngram detectors. We show that filters may capture several different semantic classes of ngrams by using different activation patterns, and that global max-pooling induces behavior which separates important ngrams from the rest. Finally, we show practical use cases derived from our findings in the form of model interpretability (explaining a trained model by deriving a concrete identity for each filter, bridging the gap between visualization tools in vision tasks and NLP) and prediction interpretability (explaining predictions).
AB - We present an analysis into the inner workings of Convolutional Neural Networks (CNNs) for processing text. CNNs used for computer vision can be interpreted by projecting filters into image space, but for discrete sequence inputs CNNs remain a mystery. We aim to understand the method by which the networks process and classify text. We examine common hypotheses to this problem: that filters, accompanied by global max-pooling, serve as ngram detectors. We show that filters may capture several different semantic classes of ngrams by using different activation patterns, and that global max-pooling induces behavior which separates important ngrams from the rest. Finally, we show practical use cases derived from our findings in the form of model interpretability (explaining a trained model by deriving a concrete identity for each filter, bridging the gap between visualization tools in vision tasks and NLP) and prediction interpretability (explaining predictions).
UR - https://www.scopus.com/pages/publications/85122039541
U2 - 10.18653/v1/w18-5408
DO - 10.18653/v1/w18-5408
M3 - Conference contribution
AN - SCOPUS:85122039541
T3 - EMNLP 2018 - 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, Proceedings of the 1st Workshop
SP - 56
EP - 65
BT - EMNLP 2018 - 2018 EMNLP Workshop BlackboxNLP
PB - Association for Computational Linguistics (ACL)
T2 - 1st Workshop on BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, co-located with the 2018 Conference on Empirical Methods in Natural Language Processing, EMNLP 2018
Y2 - 1 November 2018
ER -