Representation Learning for Conversational Data using Discourse Mutual Information Maximization
Bishal Santra, Sumegh Roychowdhury, Aishik Mandal, Vasu Gurram, Atharva Naik, Manish Gupta, Pawan Goyal
CNeRG Lab, IIT Kharagpur & Microsoft, India
NAACL 2022
NLU and Pretrained Representations
Discourse Mutual Information
3
DMI Objective
4
InfoNCE-S
Experimental Setup
5
Baselines
Results – Probing
Results – Finetuning
Case Studies
9
Context:
User 1: Are you busy tomorrow morning ?
Response:
By User 2 (Human): I'm free . What's up ?
By DMI Model: No , not this weekend . I have too much work to do .
Context:
User 1: Hi , Dan . What's new ?
User 2: Where have you been ? I've been trying to get in touch with you for two days .
User 1: I went to Salt Lake City on business . What's up ?
Response
By User 2 (Human): I got fired .
By DMI Model: Not much . I had to pay an unexpected bill , so I needed the money back .
Conclusions
Links
Thank you for your attention!