1 of 10

Neural models for Factual Inconsistency Classification with Explanations

Tathagata Raha, Mukund Choudhary, Abhinav Menon, Harshit Gupta, K V Aditya Srivatsa, Manish Gupta, Vasudeva Verma

2 of 10

What’s the work about?

FICLE! A new task, dataset, and baselines for kickstarting a much needed look into Factual Inconsistencies in text:

  • Manually annotated inconsistency tags and other 5 kinds of based-in-linguistics ontology of explanations.
  • Baselines on standard Transformer-based NLU & NLG models.
  • Weighted F1 of ∼86% and ∼76% for coarse (20-class) and fine-grained (60-class) inconsistent entity-type prediction respectively; and an IoU of ∼94% and ∼65% for claim and context span detection respectively.

All of this, without an external knowledge graph!

3 of 10

An overview of FICLE

A specialized dataset derived from the FEVER task's refuting examples, enriched with an ontology for identifying textual inconsistencies and annotated for both syntactic and semantic explanations.

4 of 10

Structural Explanations Annotation

Inconsistent Claim Fact Triple (Source-Relation-Target), Context Span, and which Structural Component the inconsistency corresponds to.

Some structural annotations’ distribution overview statistics

5 of 10

Semantic Explanations Annotation

Inconsistency Types are adapted from lexical relation types in linguistics with “Set” and “Negation” added in to handle other popular cases.

6 of 10

Baseline Neural Model Pipelines for FICLE

Stage B

Structural explanations

Stage A

Span Prediction

Stage C

Semantic explanations

7 of 10

We also experimented with predicting just “Source, Relation and Target Prediction from Claim Sentence” and just “ Inconsistent Context Span Prediction” which mostly resulted to inferior results.

Experiments & Results - I

8 of 10

Note that the two problems are 5-class and 6-class classification respectively. We observe that joint multi-task model outperforms the other two methods.

Experiments & Results - II

9 of 10

DeBERTa outperforms all other models. Embeddings & Two-step methods help!

Experiments & Results - III

10 of 10

Thank You!