Learning Explainable Linguistic Expressions with Neural Inductive Logic Programming for Sentence Classification

Prithviraj Sen, Marina Danilevsky, Yunyao Li, Siddhartha Brahma, Matthias Boehm, Laura Chiticariu, Rajasekar Krishnamurthy

Research output: Chapter in Book/Report/Conference proceedingConference paperpeer-review

Abstract

Interpretability of predictive models is becoming increasingly important with growing adoption in the real-world. We present RuleNN, a neural network architecture for learning transparent models for sentence classification. The models are in the form of rules expressed in first-order logic, a dialect with well-defined, human-understandable semantics. More precisely, RuleNN learns linguistic expressions (LE) built on top of predicates extracted using shallow natural language understanding. Our experimental results show that RuleNN outperforms statistical relational learning and other neuro-symbolic methods, and performs comparably with black-box recurrent neural networks. Our user studies confirm that the learned LEs are explainable and capture domain semantics. Moreover, allowing domain experts to modify LEs and instill more domain knowledge leads to human-machine co-creation of models with better performance.
Original languageEnglish
Title of host publicationProceedings of the 2020 Conference on Empirical Methods in Natural Language Processing
Pages4211–4221
DOIs
Publication statusPublished - 2020
EventConference on Empirical Methods in Natural Language Processing -
Duration: 16 Nov 2020 → …

Conference

ConferenceConference on Empirical Methods in Natural Language Processing
Abbreviated titleEMNLP
Period16/11/20 → …

Fingerprint

Dive into the research topics of 'Learning Explainable Linguistic Expressions with Neural Inductive Logic Programming for Sentence Classification'. Together they form a unique fingerprint.

Cite this