Top Banner
Language Model as an Annotator: Exploring DialoGPT for Dialogue Summarization Xiachong Feng 1 , Xiaocheng Feng 1,2 , Libo Qin 1 , Bing Qin 1,2 , Ting Liu 1,2 1 Harbin Institute of Technology, 2 Peng Cheng Laboratory
22

Language Model as an Annotator: Exploring DialoGPTfor ...

Mar 14, 2022

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Language Model as an Annotator: Exploring DialoGPTfor ...

Language Model as an Annotator: Exploring DialoGPT for Dialogue Summarization

Xiachong Feng1, Xiaocheng Feng1,2, Libo Qin1, Bing Qin1,2, Ting Liu1,2

1 Harbin Institute of Technology, 2 Peng Cheng Laboratory

Page 2: Language Model as an Annotator: Exploring DialoGPTfor ...

Dialogue Summarization• Dialoguesummarizationaimstogenerateasuccinctsummarywhileretainingessentialinformationofthedialogue.

2

Remember we are seeing the wedding planner after workSure, where are we meeting her? At Nonna Rita’s I want to order seafood tagliatelle Haha why notWe remmber spaghetti pomodoro disaster from our last meetingOmg it was over her white blouse:D:P

Blair and Chuck are going to meet the wedding planner after work at Nonna Rita’s. The tagliatelle served at Nonna Rita’s are very good.

Blair:

Chuck:Blair:Chuck: Blair:Chuck:

Blair:Chuck:Blair:

Dialogue

Summary

Page 3: Language Model as an Annotator: Exploring DialoGPTfor ...

A Good Summary?

3

Informativeness Redundancy Relevance

Peyrard (2019):agoodsummaryisintuitivelyrelatedtothreeaspects

Page 4: Language Model as an Annotator: Exploring DialoGPTfor ...

Related Works• Forinformativeness• Linguisticallyspecificwords• Domainterminologies• Topicwords

• Forredundancy• Similarity-basedmethodstoannotateredundantutterances

• For relevance• Topicsegmentation

4

Page 5: Language Model as an Annotator: Exploring DialoGPTfor ...

Problems• Reliedonhumanannotations.

• labor-consuming

• Obtainedviaopen-domaintoolkits

• Dialogueagnostic

• notsuitablefordialogues

5

Page 6: Language Model as an Annotator: Exploring DialoGPTfor ...

Pre-trained Language Models

6

Page 7: Language Model as an Annotator: Exploring DialoGPTfor ...

DialoGPT Annotator

7

Page 8: Language Model as an Annotator: Exploring DialoGPTfor ...

Overview• KeywordsExtraction:Extractsunpredictablewordsaskeywords.• TopicSegmentation:Insertsatopicsegmentationpointbeforeoneutteranceifitisunpredictable.• RedundancyDetection:Detectsutterancesthatareuselessforcontextrepresentationasredundant.

8

Page 9: Language Model as an Annotator: Exploring DialoGPTfor ...

Keywords Extraction: DialoGPTKE• Motivation:ifonewordinthegoldenresponseisdifficulttobeinferredfromDialoGPT,weassumethatitcontainshighinformationandcanbeviewedasakeyword.• Extractsunpredictablewordsaskeywords.

9

Context: yoo guys. EOS1Response: hey wassup. EOS2Context: hey wassup. EOS2Response: Remmber the meeting EOS3Context: Remmber the meeting EOS3Response: I almost forget it. EOS4Context: I almost forget it. EOS4Response: fine EOS5Context: fine EOS5Response: Where? EOS6Context: Where? EOS6Response: at Barbara's place. EOS7

Tom: yoo guys. EOS1John: hey wassup. EOS2Tom: Remmber the meeting EOS3John: I almost forget it. EOS4Tom: fine EOS5John: Where? EOS6Tom: at Barbara's place. EOS7

Context-response Pairs

Original Dialogue

DialoGPT

hey

Remmber

wassup.

the meeting

EOS2

EOS3

Remmber the meeting

Golden:

loss31

Word-level and Utterance-level Loss

Prediction:Keywords Extraction

loss3loss32 loss33 loss34... Extracted

KeywordsAvg

loss31 loss32 loss33 loss34

...

loss71 loss72 loss73 loss34

Page 10: Language Model as an Annotator: Exploring DialoGPTfor ...

Topic Segmentation: DialoGPTTS• Motivation:iftheresponseisdifficulttobepredictedgiventhecontextbasedonDialoGPT,weassumetheresponsemaybelongtoanothertopicandthereisatopicsegmentationbetweenthecontextandresponse.• Insertsatopicsegmentationpointbeforeoneutteranceifitisunpredictable.

10

Context: yoo guys. EOS1Response: hey wassup. EOS2Context: hey wassup. EOS2Response: Remmber the meeting EOS3Context: Remmber the meeting EOS3Response: I almost forget it. EOS4Context: I almost forget it. EOS4Response: fine EOS5Context: fine EOS5Response: Where? EOS6Context: Where? EOS6Response: at Barbara's place. EOS7

Tom: yoo guys. EOS1John: hey wassup. EOS2Tom: Remmber the meeting EOS3John: I almost forget it. EOS4Tom: fine EOS5John: Where? EOS6Tom: at Barbara's place. EOS7

Context-response Pairs

Original Dialogue

DialoGPT

hey

Remmber

wassup.

the meeting

EOS2

EOS3

Remmber the meeting

Golden:

loss31

Word-level and Utterance-level Loss

Prediction:

Topic Segmentation

loss3

SegmentationPoint

loss32 loss33 loss34Avg

loss2 loss3 loss4 loss5 loss6 loss7

SegmentationPoint

Page 11: Language Model as an Annotator: Exploring DialoGPTfor ...

Redundancy Detection: DialoGPTRD• Motivation:If oneutterancebringsbringslittleinformationandhassmalleffectsonpredictingtheresponse,thisutterancebecomesaredundantutterance.• Detectsutterancesthatareuselessforcontextrepresentationasredundant.

11

Tom: yoo guys. EOS1John: hey wassup. EOS2Tom: Remmber the meeting EOS3John: I almost forget it. EOS4Tom: fine EOS5John: Where? EOS6Tom: at Barbara's place. EOS7

Dialogue Sequence

DialoGPT

yoo guys. EOS1 hey wassup. EOS2Remmber the meeting EOS3 I almostforget it. EOS4 fine EOS5 Where? EOS6 at Barbara's place. EOS7

Original Dialogue

yoo guys. EOS1... at Barbara's place. EOS7

...Dialogue Context

Representation

𝒉𝑬𝑶𝑺𝟏 𝒉𝑬𝑶𝑺𝟐 𝒉𝑬𝑶𝑺𝟕

0.711

0.998

0.991

SimilarityCalculation

Step: 1

Step: 2

Step: 3

0.642

Step: 4

Step: 50.573

Step: 60.993

5 4

Step: 0

5 4

5 4

5

Redundantutterances

Steps

2

ℎ!"#!ℎ!"#" ℎ!"## ℎ!"#$ ℎ!"#% ℎ!"#& ℎ!"#'

ℎ!"#!ℎ!"#" ℎ!"## ℎ!"#$ ℎ!"#% ℎ!"#& ℎ!"#'

ℎ!"#!ℎ!"#" ℎ!"## ℎ!"#$ ℎ!"#% ℎ!"#& ℎ!"#'

ℎ!"#!ℎ!"#" ℎ!"## ℎ!"#$ ℎ!"#% ℎ!"#& ℎ!"#'

ℎ!"#!ℎ!"#" ℎ!"## ℎ!"#$ ℎ!"#% ℎ!"#& ℎ!"#'

ℎ!"#!ℎ!"#" ℎ!"## ℎ!"#$ ℎ!"#% ℎ!"#& ℎ!"#'

ℎ!"#!ℎ!"#" ℎ!"## ℎ!"#$ ℎ!"#% ℎ!"#& ℎ!"#'

Page 12: Language Model as an Annotator: Exploring DialoGPTfor ...

Annotation Tags

12

Page 13: Language Model as an Annotator: Exploring DialoGPTfor ...

Summarizer

13

BARTPre-trained PGN

Nonpre-trained

Page 14: Language Model as an Annotator: Exploring DialoGPTfor ...

Dataset and Metrics• Datasets• SAMSum• AMI

• Evaluation Metrics• ROUGE• BERTScore

14

Statistics for SAMSum and AMI datasets

Page 15: Language Model as an Annotator: Exploring DialoGPTfor ...

Automatic Evaluation

15TestsetresultsontheSAMSum dataset BERTScore

TestsetresultsontheAMIdataset

Page 16: Language Model as an Annotator: Exploring DialoGPTfor ...

Human Evaluation

16

modelcangetthebestscoreinconcisenessmodelcanperform

betterincoverage

Page 17: Language Model as an Annotator: Exploring DialoGPTfor ...

Effect of DialoGPTKE

17

• Entitiesplayanimportantroleinthesummarygeneration.

• CombinedwithDialoGPT embeddings,KeyBERT cangetbetterresults.

Intrinsic Evaluation For Keywords

• Viewreferencesummarywordsasgoldenkeywords

• BothTextRank andEntitiesperformpoorlyinrecall

• Ourmethodcanextractmorediversekeywords.

Page 18: Language Model as an Annotator: Exploring DialoGPTfor ...

Effect of DialoGPTRD

18

• Rule-basedmethod:annotatesutteranceswithoutnoun,verbandadjectiveasredundant.

• OurmethodshowsmoreadvantagesforlongandverbosemeetingtranscriptsintheAMI.

Page 19: Language Model as an Annotator: Exploring DialoGPTfor ...

Effect of DialoGPTTS

19

• OurmethodcangetcomparableresultswiththestrongbaselineC99(w/DialoGPT emb).

Page 20: Language Model as an Annotator: Exploring DialoGPTfor ...

Ablation Studies for Annotations

20

• Forbothdatasets,trainingsummarizersbasedon datasetswithtwoofthreeannotationscansurpasscorrespondingsummarizersthataretrainedbasedondatasetswithonetypeofannotation.

• SummarizersthataretrainedonDKE+TS stillgetimprovementsonbothdatasets.

Page 21: Language Model as an Annotator: Exploring DialoGPTfor ...

Conclusion• WeinvestigatetouseDialoGPT asunsupervisedannotatorsfordialoguesummarization,includingkeywordsextraction,redundancydetectionandtopicsegmentation.

• Experimentalresultsshowthatourmethodconsistentlyobtainsimprovementsuponpre-traind summarizer(BART)andnonpre-trainedsummarizer(PGN)onbothdatasets.

• Combiningallthreeannotations,oursummarizercanachievenewstate-of-the-artperformanceontheSAMSum dataset.

21

Page 22: Language Model as an Annotator: Exploring DialoGPTfor ...

Thanks!