1 icassp paper survey presenter: chen yi-ting. 2 improved spoken document retrieval with dynamic key...
DESCRIPTION
3 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis (PLSA)TRANSCRIPT
![Page 1: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/1.jpg)
1
ICASSP Paper Survey
Presenter: Chen Yi-Ting
![Page 2: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/2.jpg)
2
• Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis (PLSA)
• Improved Spoken Document Summarization Using Probabilistic Latent Semantic Analysis (PLSA)
• Topic and Stylistic Adaptation for Speech Summarisation• Automatic Sentence Segmentation of Speech for Automa
tic Summarization
![Page 3: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/3.jpg)
3
Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent
Semantic Analysis (PLSA)
![Page 4: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/4.jpg)
4
• In this paper, using a “dynamic key term lexicon” automatically extracted from the ever-changing document archives as an extra feature set in the retrieval task.
![Page 5: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/5.jpg)
5
• An important part of the proposed approach is the automatic key term extraction from the archives
• The second important part of the proposed approach is the key term recognition from the user spoken query– Here special approaches to recognize correctly the key term in
the user query were developed, including emphasizing the possible key term candidates during search through the phone lattice, and key term matching using a phone similarity matrix including two different distance measures.
– Two different versions of models can be used: the general lexicon including all terms except those stop terms deleted and the other based on the much smaller but semantically rich key term lexicon
![Page 6: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/6.jpg)
6
• Named Entity Recogniion– The first is to recognize the NEs from a text document (or the tra
nscription of a spoken document) using global information– The second special approach used here is for spoken document
s to recover the OOV NEs using external knowledge
• Key Term extraction by term entropy based on PLSA
)()|(
)()()|(
)|(j
kj
j
kkjjk tP
TtPtP
TPTtPtTP
)|(log)|()(1
jk
K
kjkj tTPtTPtH
![Page 7: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/7.jpg)
7
• Key term recognition from user query– The user spoken query is transcribed not only into a word graph as usua
l recognition process, but into a phone lattice as well– Then matching the phone lattice with the phone sequences of the key te
rms in the dynamic lexicon using dynamic programming (threshold)– The price paid here is of course the overall word error rate may be incre
ased• The experimental conditions
– Word error rates 27%、 character error rates 14.29%、 syllable error rate 8.91%
– 32 topics were used in PLSA modeling– 1000new stories (test set) 50 queries – The length of the queries is roughly 8-11 words– A lexicon of 61521 word was used here– A total of 1708 NE were obtained (from 9836 news stories)– Picked up the top 2000 terms ranked by term entropy
![Page 8: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/8.jpg)
8
• Experimental Results
![Page 9: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/9.jpg)
9
Improved Spoken Document Summarization Using Probabilistic Latent Semantic Analysis (PLSA)
![Page 10: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/10.jpg)
10
• where some statistical measure s(tj) (such as TF/IDF or the like)linguistic measure l(tj) (e.g., named entities)c(tj) is calculated from the confidence scoreg(tj) is N-gram score for the term tj b(S) is calculated from the grammatical structure of the sentence Sλ1, λ2, λ3, λ4 and λ5 are weighting parameters
• Two useful measures, referred to as topic significance and term entropy in this paper are proposed based on the PLSA modeling to determine the terms and thus sentences important for the document which can then be used to construct the summary
• The statistical measure s(tj) which has been proved extremely useful is called “significance score”:
)()()()()(1)( 51
4321 Sbtgtctltsn
SIn
jjjjj
t
Aijj F
Fdtnts log),()(
![Page 11: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/11.jpg)
11
• Topic significance– The topic significance score of a term tj with respect to a topic Tk
– The statistical measure:
• Term Entropy
– is a scaling factor
klTil
ik
Ddijkt
l
i
j dTPdTPdtnTS
,
)|()|(),()(
)|()()(1
ikk
K
ktj dTPTStsj
)()|(
)()()|(
)|(j
kj
j
kkjjk tP
TtPtP
TPTtPtTP
K
kjkjkijjEN tTPtTPdtnts
1
)|(log)|(),()(
)|( jk tTP )|( jk tTP
![Page 12: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/12.jpg)
12
• Experiments configuration– The test corpus included 200 news stories broadcast– Word accuracy 66.46%、 character accuracy 74.95%、 syllable
accuracy 81.70%– Sentence recall/precision is the evaluation metric for automatic s
ummarization of documents
![Page 13: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/13.jpg)
13
• Experiments configuration
![Page 14: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/14.jpg)
14
Topic and Stylistic Adaptation for Speech Summarisation
![Page 15: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/15.jpg)
15
• In this paper they investigate LiM topic and stylistic adaptation using combinations of LiMs each trained on different adaptation data
• Focusing on adapting the linguistic component, which is not related at all to the language model used during the recognition process, to make it more suited for the summarisation task
• Experiments were performed both on spontaneous speech, using 9 talks from the Translanguage English Database (TED) corpus, and speech read from text, using 5 talks from CNN broadcast news from 1995
• The measure of summary quality used in this paper is summarisation accuracy (SumACCY)
[%]100*/)( LenDelInsSubLenAccuracy
![Page 16: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/16.jpg)
16
• Automatic speech summarisation system
![Page 17: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/17.jpg)
17
• Summarisation Method• Important sentences are first extracted according to the followi
ng score for each sentence, obtained from the automatic speech recognition out (ASR)
• Starting with a baseline LiM (LiMB) we perform LiM adaptation by linearly interpolating the baseline model with other component models trained on different data
Where • Different types of component LiM are built, coming from differe
nt sources of data, and using either unigram, bigram or trigram information
n
iiLiIiC wLwIwC
NWS
1
)()()(1)(
)....|(.....)....|()....|(
11
111111
iniinn
iniiinii
wwwPwwwPwwwP
k k 1
)....|(log)( 11 iniii wwwPwL
![Page 18: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/18.jpg)
18
• Experimental Setup– Due to lack of data we had to use the talks both for development
and evaluation with a rotating form of cross-validation: all talks but one are used for development, the remaining talk being used for testing
– Summaries from the development talks are generated automatically by the system using different sets of parameters
– For the TED data, two type of component linguistic models:• The first type are built on the small corpus of hand-made sum
maries, made for the same summarisation ratio.• The second type are built from the papers in the conference
proceeding for talk we want to summarise– For the CNN data, one type of component linguistic models
• the small corpus of hand-made summaries
![Page 19: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/19.jpg)
19
• Experimental Setup– Reference results: random summarisation, the humman summari
es and the baseline
CNNTED
![Page 20: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/20.jpg)
20
![Page 21: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/21.jpg)
21
Automatic Sentence Segmentation of Speech for Automatic Summarization
![Page 22: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/22.jpg)
22
• This paper presents an automatic sentence segmentation method for an automatic speech summarization system
• The segmentation method is based on combining word- and class-based statistical language models to predict sentence and non-sentence boundaries
• Studying both the effect of the segmentation on the sentence segmentation system itself and the effect of the segmentation on the summarization accuracy
• To judge the quality of the sentence segmentation we used the F-measure metrics
![Page 23: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/23.jpg)
23
• Automatic sentence segmentation
• This probability was combined with the matching recursive path probability from
• Three LMs were used in sentence segmentation, two word-based LMs and a class-based LM
• The LMs were combined by linear interpolation as follows:
)|()|()....(
)|()|()....()....(
1211
111 1
SwpwwSpwwP
SwpSSpwwPwwP
iiiiSNO
iwiSiS i
)|()....(
)|()....()....(
1211
111 1
iiiiSNO
wiiSiSNO
wwwpwwP
SwpwwPwwPi
1iw
))(),...,(|)(())(|()|(),....,|()|(
)|()|(
11
11
iniiiiim
iniiim
mimmi
wCwCwCPwCwPhwPorwwwPhwP
hwPhwP
![Page 24: 1 ICASSP Paper Survey Presenter: Chen Yi-Ting. 2 Improved Spoken Document Retrieval With Dynamic Key Term Lexicon and Probabilistic Latent Semantic Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062504/5a4d1b947f8b9ab0599c2d7e/html5/thumbnails/24.jpg)
24
• Experimental results