Top Banner
Author's personal copy On the effectiveness of subwords for lexical cohesion based story segmentation of Chinese broadcast news L. Xie a,, Y.-L. Yang a , Z.-Q. Liu b a Shaanxi Provincial Key Laboratory of Speech and Image Information Processing, School of Computer Science, Northwestern Polytechnical University, Xi’an, China b Media Computing Group, School of Creative Media, City University of Hong Kong, Hong Kong article info Article history: Received 27 January 2010 Received in revised form 22 January 2011 Accepted 22 February 2011 Available online 2 March 2011 Keywords: Story segmentation Topic segmentation Topic detection and tracking Spoken document retrieval Subwords Lexical cohesion abstract Story segmentation divides a multimedia stream into homogenous regions each addressing a central topic. Lexical cohesion is a reasonable indicator for story boundaries. However, for story segmentation of Chinese broadcast news, directly measuring word level lexical cohe- sion is not applicable, because the texts transcribed from audio is highly unreliable and the inevitable speech recognition errors may significantly break word cohesion, thus heavily degrading the segmentation performance. To address the problem, we propose to use sub- word level cohesion in story segmentation of Chinese broadcast news, because Chinese subwords play great semantic roles and show robustness to speech recognition errors. We provide a comprehensive study on the effectiveness of subword units in story segmen- tation of Chinese speech recognition transcripts, and analyze the influence of recognition errors to the segmentation performance. Specifically, we study subword-based TextTiling and lexical chaining approaches to story segmentation, in which lexical cohesion is mea- sured using either character or syllable n-grams (n = 1, 2, 3, 4). Our extensive experiments demonstrate performance improvement of subword unigrams and bigrams over word- based methods. For instance, tested on the CCTV corpus, character unigram lexical chaining obtains a relative F1-measure gain of 12% over words on erroneous brief news transcripts (with word error rate of 40.9%). Generally, we find that subword-based methods can often obtain better segmentation than word-based ones for both error-free and erroneous transcripts. Ó 2011 Elsevier Inc. All rights reserved. 1. Introduction With the exponential growth of multimedia data containing speech, such as TV and radio broadcast news, meetings, lec- tures, voice mails and web-sharing videos, the development of automatic methods to semantically access and efficiently manage the spoken content has become increasingly important. Speech signal is semantically rich and usually covering sub- jects, concepts, topics, identities and emotions. For long streams such as a one-hour broadcast news episode, it is desirable to segment them into shorter clips that represent specific topics or stories. This would ideally allow users to swiftly jump to the start of relevant segments rather than have to search through the whole episode. Story segmentation aims to fulfill this task, which partitions a text, audio or video stream into a sequence of topically coherent segments known as stories [1]. It is an important precursor because various tasks, e.g., topic categorization and tracking, summarization, information extraction, indexing and retrieval usually assume the presence of individual topical documents [17,25]. Manual segmentation requires 0020-0255/$ - see front matter Ó 2011 Elsevier Inc. All rights reserved. doi:10.1016/j.ins.2011.02.013 Corresponding author. Tel.: +86 29 88431532; fax: +86 29 88431533. E-mail addresses: [email protected] (L. Xie), [email protected] (Y.-L. Yang), [email protected] (Z.-Q. Liu). Information Sciences 181 (2011) 2873–2891 Contents lists available at ScienceDirect Information Sciences journal homepage: www.elsevier.com/locate/ins
19

Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Jun 30, 2020

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

On the effectiveness of subwords for lexical cohesion based storysegmentation of Chinese broadcast news

L. Xie a,⇑, Y.-L. Yang a, Z.-Q. Liu b

a Shaanxi Provincial Key Laboratory of Speech and Image Information Processing, School of Computer Science, Northwestern Polytechnical University, Xi’an, Chinab Media Computing Group, School of Creative Media, City University of Hong Kong, Hong Kong

a r t i c l e i n f o

Article history:Received 27 January 2010Received in revised form 22 January 2011Accepted 22 February 2011Available online 2 March 2011

Keywords:Story segmentationTopic segmentationTopic detection and trackingSpoken document retrievalSubwordsLexical cohesion

a b s t r a c t

Story segmentation divides a multimedia stream into homogenous regions each addressinga central topic. Lexical cohesion is a reasonable indicator for story boundaries. However, forstory segmentation of Chinese broadcast news, directly measuring word level lexical cohe-sion is not applicable, because the texts transcribed from audio is highly unreliable and theinevitable speech recognition errors may significantly break word cohesion, thus heavilydegrading the segmentation performance. To address the problem, we propose to use sub-word level cohesion in story segmentation of Chinese broadcast news, because Chinesesubwords play great semantic roles and show robustness to speech recognition errors.We provide a comprehensive study on the effectiveness of subword units in story segmen-tation of Chinese speech recognition transcripts, and analyze the influence of recognitionerrors to the segmentation performance. Specifically, we study subword-based TextTilingand lexical chaining approaches to story segmentation, in which lexical cohesion is mea-sured using either character or syllable n-grams (n = 1,2,3,4). Our extensive experimentsdemonstrate performance improvement of subword unigrams and bigrams over word-based methods. For instance, tested on the CCTV corpus, character unigram lexical chainingobtains a relative F1-measure gain of 12% over words on erroneous brief news transcripts(with word error rate of 40.9%). Generally, we find that subword-based methods can oftenobtain better segmentation than word-based ones for both error-free and erroneoustranscripts.

� 2011 Elsevier Inc. All rights reserved.

1. Introduction

With the exponential growth of multimedia data containing speech, such as TV and radio broadcast news, meetings, lec-tures, voice mails and web-sharing videos, the development of automatic methods to semantically access and efficientlymanage the spoken content has become increasingly important. Speech signal is semantically rich and usually covering sub-jects, concepts, topics, identities and emotions. For long streams such as a one-hour broadcast news episode, it is desirable tosegment them into shorter clips that represent specific topics or stories. This would ideally allow users to swiftly jump to thestart of relevant segments rather than have to search through the whole episode. Story segmentation aims to fulfill this task,which partitions a text, audio or video stream into a sequence of topically coherent segments known as stories [1]. It is animportant precursor because various tasks, e.g., topic categorization and tracking, summarization, information extraction,indexing and retrieval usually assume the presence of individual topical documents [17,25]. Manual segmentation requires

0020-0255/$ - see front matter � 2011 Elsevier Inc. All rights reserved.doi:10.1016/j.ins.2011.02.013

⇑ Corresponding author. Tel.: +86 29 88431532; fax: +86 29 88431533.E-mail addresses: [email protected] (L. Xie), [email protected] (Y.-L. Yang), [email protected] (Z.-Q. Liu).

Information Sciences 181 (2011) 2873–2891

Contents lists available at ScienceDirect

Information Sciences

journal homepage: www.elsevier .com/locate / ins

Page 2: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

annotators to work through the entire audio/video stream, which is tedious and costly. The need for automated segmenta-tion approaches has become very pressing recently as a result of huge multimedia data produced.

Recently, lexical cohesion-based methods have drawn much interest for story segmentation [12,7,29,27,2,4]. Lexicalcohesion [11] indicates that words in a story (or topic) hang together by semantic relations and different stories tend toemploy different sets of words. The TextTiling method [12] claims lexical similarity minimums as story boundaries througha word similarity measure across the text. The lexical chaining method [29] chains up related words in a text and a highconcentration of chain starts and ends is declared as a story boundary.

Traditionally, lexical cohesion-based story segmentation has been studied at word level. In this paper, we perform a com-prehensive study on subword-based approaches to story segmentation of Chinese broadcast news. Our motivations aretwofold.

1. First, different from western languages, Chinese is a character-based language and monosyllabic [45]. Chinese subwords,e.g., characters and syllables, play important semantic roles. The latent effectiveness of subwords warrants an investiga-tion of lexical cohesion-based story segmentation of Chinese broadcast news.

2. Second, story segmentation of broadcast news is performed largely on erroneous textual transcripts. Previous approachesmeasure word relations on inaccurate texts transcribed from audio (via a speech recognizer) and they did not take intoaccount any error compensation methods. However, it is known that speech recognition errors break lexical cohesionamong words, leading to performance degradation. Our previous preliminary study shows that measuring lexical cohe-sion at subword units hold much promise to solving the problem [42]. Subword units may be robust to speech recogni-tion errors because of their partial matching merit. At subword levels, the mis-recognized words may include somesubword units correctly recognized and matching on the subword level can thus recover word relations in noisytranscripts. However, the effectiveness of subwords for Chinese story segmentation desires a comprehensive study usingdifferent lexical methods, data sets from different sources and different speech recognition error rates.

Therefore, in this paper, we complete an extensive study on the effectiveness of various subword representations in Chi-nese story segmentation. Our experimental study on two popular methods, two Mandarin corpora (TDT2 and CCTV) andtranscripts with different speech recognition error rates demonstrates that Chinese subwords can achieve considerable per-formance gains over words in lexical cohesion-based story segmentation of Chinese broadcast news, both in error-free anderroneous transcripts.

The rest of this paper is organized as follows: Section 2 makes a brief survey on related work. Section 3 describes theTextTiling and lexical chaining methods for story segmentation. In Section 4, we study the robustness of Chinese subwordsand subword-based story segmentation approaches. Section 5 provides our experiments and analysis on the results. Finally,conclusions are drawn in Section 6.

2. Related work

Automatic story segmentation on multimedia documents is a challenging task. Text documents are often clearly orga-nized with titles, sentences and paragraphs via typographic cues, e.g. punctuation and capitalization. However, spoken orvideo documents do not have such structural or typographic merits. Previous efforts on multimedia segmentation focuson three categories of cues: visual cues such as presence of an anchor face [14] and motion changes [13], audio cues suchas significant pauses and pitch resets [41,38], and lexical cues such as word similarity measures from speech recognitiontranscripts or closed captions of video [12,29,8,44]. Cues from different modalities (audio, video and text) can be fused toachieve a better segmentation performance [14,27,32,18].

Hui et al. [16] proposed to detect studio-to-field transitions by spatial and color histogram differences among consecutivevideo frames. They discovered that story boundaries often coincide with studio-to-field shot boundaries in many broadcastnews programs. Anchor face detection has been extensively studied for the story segmentation task since the presence of ananchorperson is another salient visual feature [21,14]. Anchors often play introductive or conclusive roles of news reports.Visual cues are widely studied in the TREC video retrieval evaluations (TRECVID) [30,31,15].

Audio signal implies rich structural information for story segmentation [39]. For example, program directors usually usesalient pauses between consecutive news stories [28,20]. News programs often use short music clips to switch topics andnewscasts involve multiple speakers like anchors, reporters and interviewees. Hence speech/music shifts and speakerchanges may indicate story boundaries [28,14]. Speech prosodic cues have lately raised interest in decision tree [35,36]based topic segmentation [28,20]. Speakers naturally separate their speech into different topics or subtopics through into-national, durational and energy cues.

Audio and visual cues depend on the editorial and production rules and these rules can vary from different media sources.Lexical cues are more generic since they probe story shifts by monitoring the intrinsic semantic variations. Since closed cap-tions are not always available for multimedia documents, lexical-based segmentation approaches are usually carried out onspeech recognition transcripts. As compared with video and audio cues, lexical cues are more popular as they work on bothtext and multimedia sources. Major lexical approaches include word cohesiveness, the use of key phrases and topicmodeling.

2874 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 3: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

Many news programs use cue terms such as ‘stay tuned’ and ‘reports’ at the beginning or end of a news story. Thusdetecting such key phrases can help to locate story boundaries [3,13]. Hsu et al. [13] selected frequent cue phrases in thevicinities of story boundaries from a training set and combined them with other cues in an exponential model-basedapproach. Topic modeling based approaches include hidden Markov models (HMM) [44], maximum entropy (ME) models[3,14,33,34], local content analysis (LCA) [26] and genetic algorithm (GA) [37,24]. In Yamron’s HMM approach [44], topicsare modeled by nodes and words are the emitting observations of topics. Under the HMM framework [40], topic shiftsare signaled by transitions between nodes.

Lexical cohesion is a textual quality that makes the sentences in a topic seems to hang together via inter-word semanticrelations [11]. Text segments with similar vocabulary are more likely to be part of a coherent topic. Repetition (i.e. co-occurrence) of words is the most common appearance of the lexical cohesion phenomenon. Based on this principle, mucheffort has been devoted to lexical cohesion approaches for text segmentation. Major approaches involve TextTiling [12],C99 [7] and lexical chaining [29].

Hearst et al. [12] proposed the TextTiling approach based on the straightforward observation that different topics usuallyemploy different sets of words and shifts in vocabulary usage are indicative of topic changes. As a result, pairwise sentencesimilarities are measured across the text and a local similarity minimum implies a story boundary. Stokes et al. [29] embod-ied word cohesion by a lexical chaining approach, in which related words in a text were linked into chains and a high con-centration of chain starting and ending points is an indication as a story boundary. These two approaches have been recentlyintroduced to segment multimedia documents such as broadcast news [27] and meetings [2] on speech recognitiontranscripts.

Despite of considerable attention from TREC SDR [10], TRECVID [31] and TDT1 evaluations, performances of lexical-basedstory segmentation on spoken documents remain inferior. The high level of speech recognition error rate is one of the majorobstacles. Lexical methods detect story boundaries on noisy texts transcribed from audio via a large vocabulary continuousspeech recognizer (LVCSR). The inevitable errors can cause word matching failures [23], break lexical cohesion and thus degradestory segmentation performance. According to TRECVID 2006, the word error rate (WER) is about 30% for English broadcastnews and about 40% for Mandarin broadcast news. Adverse acoustic conditions, diverse speaking styles and OOV words arethe primary contributors to the speech recognition errors. The existence of out-of-vocabulary (OOV) words (i.e., words outsidethe vocabulary of the speech recognizer) is more common for Chinese than other languages such as English due to the flexibleword-building nature of Chinese. Chinese OOV words are largely named entities (e.g. Chinese person names and transliteratedforeign names) that are keys to topic discrimination.

Because of the complexity of Chinese, some researchers searched for effective approaches for Chinese story segmentationfrom its language point of view. Levow [20] performed an initial investigation on the pitch features in prosody-based Chinesestory segmentation. Our previous work discovered that the pitch reset cues (formally effective in English) are affected by theChinese lexical tones and that tone-normalized pitch resets are more effective in Chinese story segmentation [41,38]. Somerecent researches tried to integrate multi-modal features (lexical, acoustic, video) in Chinese story segmentation [18,19,27].

Recently, subword indexing units (e.g. phonemes, syllables and sub-phonetic segments) have shown robustness to speechrecognition errors and OOV words in spoken document retrieval (SDR) [23] tasks. Especially for Chinese, retrieval based oncharacter or syllable indexing is superior to words due to the special features of Chinese [5,22]. We believe that subwordsshould also be effective in story segmentation of erroneous broadcast news transcripts through partial matching. It oftenhappens that, although words have been incorrectly recognized, they may still contain subword units that have been cor-rectly recognized. From which, it’s possible to recover useful word relations that are essential to story segmentation. Ourpreliminary study has demonstrated the potential of subwords in story segmentation of Chinese broadcast news transcriptsin the presence of speech recognition errors [43,42].

In this paper, we perform an intensive study of the effectiveness of subword units in lexical cohesion-based story segmen-tation of Chinese broadcast news. We examine the performance and behavior of various Chinese subword units in storysegmentation on error-free manual transcripts and speech recognition transcripts with various error rates. We investigatethe feasibility of using subword representations for story segmentation as an alternative to words in TextTiling and lexicalchaining. Chinese is highly different from English and other western languages. Chinese subword units (i.e. characters andsyllables) play important semantic roles due to the character-based-wording and monosyllabic natures of Chinese. Withtheir robustness to speech recognition errors, modeling lexical cohesion at subword levels may lead to superior storysegmentation performance on Chinese speech recognition transcripts.

3. Lexical cohesion based story segmentation

3.1. Lexical cohesion

Lexical cohesion describes that a text with a central topic is created by using words with related meanings and the wordshang together as a whole through cohesive relations [11]. Major lexical cohesion relations include word repetition, synonym/

1 http://projects.ldc.upenn.edu/TDT/.

L. Xie et al. / Information Sciences 181 (2011) 2873–2891 2875

Page 4: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

antonym, specialization/generalization, and part/whole relation. Some examples are shown in Table 1. Among these rela-tions, repetition is a strong, frequently used cohesion indicator.

A plenty of research [12,7,29] has shown that lexical cohesion is a useful device to detect story changes since words in anindividual topical story are semantically cohesive while different stories tend to have different vocabulary. We can thus iden-tify story boundaries by detecting shifts of vocabulary in a text document. Previous research shows that it is ‘‘counterintu-itive and disappointing’’ that using more semantic relations has a negative effect on the performance of story segmentation[29], because additional semantic relations other than repetition can incur noise. According to Stokes [29], the use of repe-tition-only exhibits the best performance for story segmentation. Therefore, we consider only term repetitions in our study.

3.2. TextTiling

TextTiling measures consecutive sentence similarities across the text and a local similarity minimum implies a possibletopic shift [12]. The TextTiling algorithm includes three steps: tokenizaiton, lexical score determination and boundary iden-tification. As a precursor, the tokenization step divides the input text into individual lexical units (usually words). For broad-cast news, tokenization is carried out at the speech recognition stage and the resultant LVCSR transcript has already beensegmented to a sequence of words.

In lexical score determination, the text stream is first segmented into sentences or pseudo-sentences such as a fixed win-dow of words. The lexical similarity is calculated for each sentence pair (i, i + 1) via lexical score:

Sði; iþ 1Þ ¼ cosðvi;viþ1Þ ¼PN

n¼1vn;ivn;iþ1ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiPNn¼1v2

n;i

PNn¼1v2

n;iþ1

q ; ð1Þ

where vi and vi+1 are term frequency vectors for two adjacent sentences i and i + 1, respectively, and vn,i is the frequency ofthe term wn occurred in i with a vocabulary size of N. For LVCSR transcripts, lexical scores are usually determined betweentwo adjacent pseudo-sentences defined as a fixed number of terms (T). This is because: (1) real sentence boundaries are notreadily available in speech recognition transcripts and sentence segmentation is another challengeable task; (2) the numberof shared terms between two long sentences and between a long and a short sentence would probably yield incomparablescores [43]. Since story boundaries are searched at each inter-sentence point, we increase the boundary hypothesizes by asliding step. Lexical scores are calculated at {T,T + D,T + 2D, . . .} term positions, where D is the sliding length and D 6 T. Fig. 1shows a lexical score curve calculated on a broadcast news transcript. We can clearly observe that story boundaries consis-tently correspond to similarity valleys.

TextTiling uses depth score rather than lexical score to identify the story boundaries in the final boundary identificationstep. Depth scores are calculated at valley points of the lexical score time trajectory:

DðvÞ ¼ ðSðplÞ � SðuÞÞ þ ðSðprÞ � SðuÞÞ; ð2Þ

where u is a valley point, and pl and pr are the nearest left and right peaks around u, respectively. For those non-valley points,their depth scores are assigned to 0. It is indicative from the depth score curve that a sharp drop in lexical similarity is moreprobable to be a story boundary, as shown in Fig. 1. Finally, boundaries are identified on the time trajectory of the depthscore, in which a time point whose depth score exceeds a pre-defined threshold h is determined as a story boundary.

Table 1Different lexical cohesion types in broadcast news.

Type Example

Repetition

Synonym/antonym

Generalization/specialization

Part/whole

2876 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 5: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

3.3. Lexical chaining

Lexical chaining is another embodiment of lexical cohesion, where a chain links related terms (e.g. words) across a textstream [29]. We expect that a high concentration of starting and/or ending chains is an indicator of a story boundary. After asimilar tokenization step as in TextTiling, the chaining procedure performs a single-pass clustering, where the first token inthe input text stream forms the first lexical chain and each subsequent token is linked to an existing chain if it is related to atleast one token in that chain by any pre-defined lexical cohesion relations such as repetition and synonymy. As describedbefore, lexical chains are usually formed by repetitions only since additional semantic relations induce noises to the segmen-tation process [29]. Fig. 2 shows an example of lexical chaining for a broadcast news transcript excerpt. We usually set up amaximal chain length and beyond which no chains are allowed. This is because some terms in a news story may re-appear inanother story. For example, some chains may span across the entire text if two stories reporting the same topic are situatedat the beginning and end of a news episode.

After chaining, we measure the boundary strength for each pair of adjacent sentences by

Cði; iþ 1Þ ¼ EðiÞ þ Bðiþ 1Þ; ð3Þ

where EðiÞ and Bðiþ 1Þ denote the number of chains ending at sentence i and the number of chains beginning at sentencei + 1, respectively. Fig. 2 shows the boundary strength scores calculated at inter-sentence positions for a broadcast newstranscript excerpt. Fig. 3 plots the boundary strength curve for a broadcast news transcript and we can clearly observe thatstory boundaries consistently meet boundary strength peaks. An inter-sentence position whose boundary strength exceeds apre-tuned threshold h is considered as a story boundary. Pseudo-sentences with a fixed number of terms (T) are usually usedfor story segmentation of LVCSR transcripts.

4. Subword lexical cohesion approaches

Lexical-based story segmentation approaches usually involve word matching, e.g., word frequency counts in sentencesimilarity measure of TextTiling [12] and connecting word repetitions in lexical chaining [29]. However, speech recognition

0 500 1000 1500 2000 2500 3000 35000

0.10.20.30.40.50.60.70.80.91

Time(#ofwords)

Lexical similarityDepth score

Fig. 1. Lexical score and depth score curves for a broadcast news transcript. Vertical red lines denote reference story boundaries. (For interpretation of thereferences in colour in this figure legend, the reader is referred to the web version of this article.)

Fig. 2. Lexical chaining and boundary strength calculation for a broadcast news excerpt.

L. Xie et al. / Information Sciences 181 (2011) 2873–2891 2877

Page 6: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

errors induce severe word matching failures. In subword levels, we can conduct partial matching or ‘‘sound-like’’ matchingthat can partially recover the relations among words. This merit is especially important for Chinese due to its special char-acteristics. In this section, we first describe these language-specific characteristics and then show the robustness of Chinesesubwords in lexical matching. After that, we present our subword-based TextTiling and lexical chaining methods.

4.1. Special characteristics of Chinese

Chinese is significantly different from western languages such as English in both written and spoken aspects, as listed inTable 2. Instead of alphabetic, Chinese is a character-based language and a word is composed of one or more characters.There are about 6500 commonly used characters2 and almost every character is a morpheme with its meaning. Different fromEnglish, there is no space between words to mark word boundaries in a Chinese text. In fact, word is not clearly defined in Chi-nese. Therefore, word segmentation is a particularly difficult task for Chinese and the segmentation result of a sentence is usu-ally not unique and often confusing [9].

In spoken aspect, Chinese is monosyllabic, i.e., each character is pronounced as a syllable. Different from English, Chinese isa tonal language and each syllable is associated with a lexical-meaningful tone. Syllables with different tones indicate dif-ferent lexical meanings. Mandarin syllable tones are expressed acoustically in pitch trajectories. There are four tones plus aneutral tone for Mandarin (Putonghua) and other dialects may have different tones. In Mandarin, about 1200 phonologicallyallowed tonal syllables correspond to over 6500 commonly used simplified Chinese characters. When tones are disregarded,the number is reduced to only about 400, known as base syllables. Fig. 4 shows the building blocks of a Chinese sentence.

4.2. Robustness of subwords to speech recognition errors

The high character-syllable ratio (16:1) results in a large numbers of homophones in Chinese. Tones are often misrecog-nized by the speech recognizer, which contributes a lot to the speech recognition errors. In Chinese LVCSR transcripts, it is

0 200 400 600 800 1000 1200 1400 1600 1800 20000

5

10

15

20

25

30

Time (# of words)

Cha

in s

tren

gth

Fig. 3. Boundary strength curve for a broadcast news transcript. Vertical red lines denote reference story boundaries. (For interpretation of the references incolour in this figure legend, the reader is referred to the web version of this article.)

Table 2The differences between English and Chinese.

English Chinese

Written Alphabetic, word-based, space as word delimiters Character-based, no word delimiters

Spoken Non-tonal, accentual Tonal, monosyllabic

Fig. 4. A Chinese sentence with its component words, characters and syllables.

2 Simplified characters that are widely used in mainland China and Singapore.

2878 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 7: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

common that a word is subsituted by another character sequence with the same or similar pronunciation, where homo-phone characters are the probable substitutions. Table 3 shows some word matching failures due to speech recognition er-rors. For example, word matching fails between words ‘‘ ’’ (Dianchi Lake) and its substitution ‘‘ ’’ (television).However, they have similar pronunciations and we still can link them together via syllable ‘‘dian’’. Similarly, a word match-ing failure occurs between the foreign name (Albright) and its LVCSR resultant ‘‘ ’’ (two stepWright). We can still link the two together via subword character string ‘‘ ’’.

4.3. Robustness of subwords to OOV words

Flexible word formation in Chinese makes limited characters to produce unlimited words. This is the open vocabularynature, i.e., there does not exist a commonly accepted lexicon for Chinese. In broadcast news domain that probes timelyevents, new words are born almost everyday. As a result, the OOV problem is acute in Chinese LVCSR transcripts. ManyOOV words in BN are named entities (NE) that are important to topic discrimination. An OOV word distributed in differentplaces of a news story may share part of the characters or be substituted by several different character strings with the same(or partially same) syllable sequence. Some examples are shown in Table 4. For example, the place name ‘‘ ’’ is anOOV word that is substituted as other phrases ‘‘ ’’ (cannot afford) and ‘‘ ’’ (vice-ministerial). We cannotmatch them together at word level. However, they both share the same syllable ‘‘bu’’ that still can recover their relations.Foreign proper names are common OOV words in Chinese spoken documents as they are transliterated to Chinese charactersequences based on the pronunciations (i.e. phonetic transliteration). Consequently, speech recognizer may return differentcharacter sequences with the same or similar pronunciations, probably their homophones. For example, it is possible to linkthe three LVCSR resultants of the foreign person name ‘‘ ’’, which have completely different meanings, via char-acters ‘‘ ’’ or syllables ‘‘si-ji’’, as shown in Table 4.

Table 4Some OOV words in Chinese LVCSR transcripts. Subword units for partial matching are underlined.

Characters Base syllables

OOV word: (a Chinese place) ku-bu-qi

LVCSR output (cannot afford) fu-bu-fi

(vice-ministerial level) fu-bu-ji

OOV word: (a Chinese name) wang you cai

LVCSR output (when have money) dang you cai

(king rape) wang you cai

(national friendship talent) bang you cai

OOV word: (Lewinsky) lai wen si ji

LVCSR output (come article this base) lai wen si ji

(come ask driver) lai wen si ji

(show-up driver) lai de si ji

Table 3Some speech recognition errors in Chinese LVCSR transcripts. Subword units for partial matching are underlined.

Character sequence Syllable sequence English translation

Original word dian-chi Dianchi Lake

LVCSR result dian-shi television

Original word zhong-you heavy oil

LVCSR result zhong-yao important

Original word ao-er-bu-lai-te Albright

LVCSR result er bu lai-te two step Wright

Original word hu-lian-wang internet

LVCSR result hu lian-wang mutual connection

Original word a-er-ji-li-ya Algeria

LVCSR result bao-er-ji li yao Bauer drive want

L. Xie et al. / Information Sciences 181 (2011) 2873–2891 2879

Page 8: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

4.4. Merit of subwords in semantic matching

Besides recovering matching failures from noisy transcripts, the superiority of subwords also lies in their ability to linksemantically related words that share some component characters. As we described in Section 4.1, a considerable numberof Chinese words are compositional, i.e., the meaning of the word is related to its component characters [6]. We show someexamples in Table 5. For example, the words ‘‘ ’’ (nuclear energy), ‘‘ ’’ (nuclear fuel) and ‘‘ ’’ (nucle-ar reactor) are all extracted from a news story about the North Korean nuclear crisis. They share the same component char-acter ‘‘ ’’ (nuclear). Obviously, the cohesive relations between these words cannot be captured by rigid word matching.However, character level matching can discover their relations.

The flexibility of Chinese word segmentation also result in word level speech recognition errors. The same character se-quence may be segmented as several different word sequences that are both syntactically valid and semantically meaningful.For example, the proper noun ‘‘ ’’ (The UN General Assembly) can appear as a unique word and also can besegmented to two words, ‘‘ ’’ (UN) and ‘‘ ’’(General Assembly), in the same LVCSR transcript. Rigid word match-ing cannot link them together, while their relations can be found via character or syllable matching.

4.5. Subword-based TextTiling and lexical chaining

Motivated by the merits of subwords, we propose to use different Chinese subword representations i.e., character andsyllable n-gram units, in TextTiling and lexical chaining. Given a sequence of words {w1w2w3� � �wQ} and the sequence ofits component characters (or syllables) {c1c2c3� � �cL}, the overlapping subword n-grams are defined in Fig. 5. Higher-ordersubword overlapping n-grams can be formed accordingly. We use unit overlap to reduce the possibility of missing any usefulinformation embedded in the subword sequence.

Subword based TextTiling: We measure the lexical similarity on the sequence of subword overlapping n-gram trans-formed from the LVCSR word transcripts. Lexical score on subword level is defined as

Table 5Examples of semantically related words sharing component characters. Subword units forsemantic matching are underlined.

Character sequence Syllable sequence English translation

he-neng nuclear energy

he-ran-liao nuclear fuel

he-fan-ying-dui nuclear reactor

jie-fang-zhan-zheng war of liberation

ye-zhan-jun field army

zhan-yi campaign

shui-wei water level

shui-li-shu-niu water control pivot

shui-yu water area

Fig. 5. Forming rules for subword overlapping n-grams.

2880 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 9: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

Sði; iþ 1Þ ¼ cosðv̂i; v̂iþ1Þ ¼PM

m¼1v̂m;iv̂m;iþ1ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiPMm¼1v̂2

m;i

PMm¼1v̂2

m;iþ1

q ; ð4Þ

where v̂i and v̂iþ1 are the subword n-gram frequency vectors of pseudo-sentences i and i + 1, respectively. v̂m;i and v̂m;iþ1 arethe frequency of the subword n-gram unit cm occurred in sentences i and i + 1, respectively. M is the size of the subwordvocabulary. Depth scores on the subword n-gram sequence are calculated from the lexical score trajectory according toEq. (2) and story boundaries are detected by a preset threshold.

Subword based lexical chaining: For each n-gram scale (n = 1,2,3,4), we form lexical chains by connecting repetitions ofsubword n-gram units produced by the rules described in Fig. 5. Fig. 6 shows an example of lexical chaining that connectsrepetitions on the character unigram sequence of an LVCSR transcript. Boundary strengthes are calculated via Eq. (3) andstory boundaries are identified by a predefined threshold.

5. Experiments and analysis

5.1. Corpus

5.1.1. TDT2Topic detection and tracking Phase 2 (TDT2) Mandarin corpus3 is released by LDC, which contains about 53 h of Mandarin

broadcast news audio from Voice of America. The 177 VOA recordings span from February to June 1998, accompanied withmanually annotated meta-data including story boundaries, manual word transcripts (namely TDT2-Ref) and LVCSR transcripts(namely TDT2-LVCSR). The TDT2 audio was transcribed by the Dragon LVCSR system with word, character and base syllableerror rates of 37%, 20% and 15%, respectively.

We separate the corpus into two non-overlapping subsets: a development set of 90 recordings with 1321 story bound-aries and a test set of 87 recordings with 1262 story boundaries. The development set is used for empirical parameter tuningand the test set is for story segmentation performance evaluation. According to TDT2, a detected story boundary is consid-ered correct if it lies within a 15-s tolerant window (about 30 words on average) on each side of a manually-annotated ref-erence boundary.

5.1.2. CCTV broadcast newsWe collect another broadcast news corpus from China Central Television (CCTV) in order to test story segmentation per-

formance at different levels of speech recognition errors. The corpus contains 71 news episodes with 27 h of CCTV newsaudio and 2101 story boundaries.

Fig. 6. Lexical chaining and boundary strength calculation on the character unigram representation of a broadcast news transcript excerpt.

3 http://projects.ldc.upenn.edu/TDT2/.

L. Xie et al. / Information Sciences 181 (2011) 2873–2891 2881

Page 10: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

We use Julius LVCSR4 to transcribe the CCTV corpus. A set of 60 h of CCTV broadcast news audio is adopted to train acousticmodels, i.e. triphone hidden Markov models (HMMs). The textual data for bigram and trigram language model training comesfrom CCTV news transcripts (37 M characters) and People’s Daily (431 M characters). To perform story segmentation experi-ments, we achieve three sets of transcript with different speech recognition errors by using different language models, as sum-marized in Table 6.

We divide the corpus into two parts: a development set with 40 audio files (1209 story boundaries) and a test set with 31audio files (892 story boundaries). Each CCTV broadcast news episode is made up of a detailed news session (about 25 min)and a brief news session (about 5 min). Fig. 7 shows the story length distribution of detailed and brief stories in the corpus.To accord with the TDT2 standard, we assume that a detected story boundary is correct if it lies within a K-word-length tol-erance window on each side of a manually-annotated story boundary (K = 10 for brief story and K = 30 for detailed story).

5.2. Experimental setup

We have carried out experiments on the TDT2 corpus and the CCTV corpus to evaluate the effectiveness of Chinese sub-words in story segmentation of Chinese broadcast news. TextTiling and lexical chaining approaches are involved in theexperiments. Story segmentation experiments are performed on different lexical scales, i.e., words and subword n-grams(n = 1,2,3,4) in form of characters and syllables. We investigate story segmentation performance on both manual transcriptsand LVCSR transcripts with different speech recognition errors. Our goal is to determine whether subword units have enoughrepresentational power to capture the information needed to story discrimination. Experiments on error-free manual tran-scripts are used to show the performance upper bounds of different lexical scales.

We first conduct empirical parameter tuning on the development sets to obtain optimal parameters that achieve the bestperformance of story segmentation. Experiments are then carried out on the test sets using the tuned parameters. Theparameters for TextTiling are pseudo-sentence length (T), sliding length (D) and boundary identification threshold (h).The parameters for lexical chaining include maximal chain length L, pseudo-sentence length (T) and boundary identificationthreshold (h).

The evaluation criterion for story segmentation is F1-measure, i.e., the harmonic mean of recall and precision, defined as

recall ¼ Ncor

Nref; ð5Þ

precision ¼ Ncor

Nret; ð6Þ

Table 6Performance of LVCSR with different language models on CCTV corpus. Manual transcription is also listed.

Transcript Nature Error rate on CCTV corpus (%)

Word Character Base-syllable

CCTV-Ref Manual transcription (reference) 0 0 0CCTV-LVCSR1 LVCSR transcription, recognizer uses a trigram trained on 468 M characters 25.0 18 15.5CCTV-LVCSR2 LVCSR transcription, recognizer uses a trigram trained on 37 M characters 33.4 24.6 20.3CCTV-LVCSR3 LVCSR transcription, recognizer uses a bigram trained on 37 M characters 40.9 29.7 24.1

0 200 400 600 800 10000

0.05

0.1

0.15

0.2

0.25

0.3

0.35

Length (# of words)

Rel

ativ

efre

q uen

cy

TDT2 news

0 200 400 600 800 10000

0.05

0.1

0.15

0.2

0.25

0.3

0.35

Length (# of words)

Rel

ativ

e fre

quen

cy

CCTV Brief newsCCTV Detailed news

Brief news average length: 61Detailed news average length: 237

Average length: 147

Fig. 7. Story length distribution in CCTV corpus (left) and TDT2 corpus (right).

4 http://julius.sourceforge.jp/.

2882 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 11: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

and

F1-measure ¼ 2 � recall � precisionrecallþ precision

; ð7Þ

where Ncor is the number of correctly detected story boundaries, Nref is the number of actual story boundaries (manual anno-tation), and Nret is the number of boundaries returned by the story segmentation system.

5.3. Effectiveness of subwords on error-free transcripts

We have experimented story segmentation on error-free manual transcripts. In these experiments, we aim to determinewhether subword units have enough representative power to perform effective story segmentation as an alternative towords. The experimental results are summarized in Fig. 8, which provide upper bounds on the performance of different sub-word representations for story segmentation.

In general, most subword unigrams and bigrams achieve considerable performance improvements as compared withwords on manual transcripts for both the TextTiling and the lexical chaining methods. As shown in Fig. 8, syllable bigramperforms the best for the TDT2 corpus and character unigram achieves the best performance for the CCTV corpus. Perfor-mance comparisons and relative performance gains of subwords are listed in Table 7 (TextTiling) and Table 8 (lexical

0.63

15

0.66

99

0.66

59

0.57

33

0.43

64

0.61

36

0.61

39

0.66

74

0.58

49

0.44

6

0.59

53

0.59

79

0.46

38

0.44

68

0.57

4

0.46

37 0.55

93

0.48

9

0.44

47

0.59

95

0.64

19

0.60

04

0.28

57

0.13

19

0.56

41

0.60

18

0.60

35

0.29

0.13

45

0.53

75

0.59

17

0.53

34

0.48

61

0.48

82

0.52

33

0.56

84

0.52

41

0.48

28

0.48

70.56

73

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

Word Unigram Bigram Trigram Quadgram

CCTV-Ref

Character TextTiling

Syllable TextTiling

Character Lexical Chaining

Syllable Lexical Chaining

Character TextTiling

Syllable TextTiling

Character Lexical Chaining

Syllable Lexical Chaining

Brief newsDetailed news

0.55

89 0.58

36

0.59

58

0.50

38

0.43

21

0.53

92

0.55

98 0.59

73

0.50

66

0.43

2

0.51

57

0.57

23

0.48

73

0.46

53

0.51

36

0.47

47

0.54

07

0.48

72

0.46

93

0.55

11

0.4

0.45

0.5

0.55

0.6

0.65

Word Unigram Bigram Trigram Quadgram

TDT2-Ref Character TextTilingSyllable TextTilingCharacter Lexical ChainingSyllable Lexical Chaining

Fig. 8. Story segmentation results (F1-measure) for error-free manual transcripts on TDT2 and CCTV. The performances of the component character andsyllable sequence of word are also drawn (bars in the most left bin).

Table 7Performance comparison between word and subword for TextTiling on manual transcripts.

Transcripts F1-measure Subword scale with best F1-measure

Word Subword Relative improv. (%)

TDT2-Ref 0.5589 0.5973 6.4 Syllable bigram

CCTV-Ref Detailed news 0.6315 0.6699 6.1 Character unigramBrief news 0.5995 0.6419 7.1 Character unigram

L. Xie et al. / Information Sciences 181 (2011) 2873–2891 2883

Page 12: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

chaining). Using subwords we were able to achieve improvements ranging from 6.1% to 11.3%. We believe the superiorperformance of character unigram on manual transcripts attributes to the following reasons. Words linked together by longersubword n-grams also can be matched through character unigram. Character unigrams can enhance the cohesive relation ofword repetitions with the increase of the word length. For example, for a 5-character word, the two appearances of the wordin a story can be captured for five times when using character unigram matching. That is to say, character unigram matchinghas a cohesive weighting effect to words with different length, where longer words are associated with higher weights. As weknow in Chinese, long words are likely to be proper nouns that are highly topic-related. The weighting can enhance theimpact of these proper nouns. Fig. 9 shows an example of the effects of character unigram. In this example, the multiplematchings of proper nouns ‘‘ ’’ (Darfur, a place in Sudan) and ‘‘ ’’ (Federal Reserve Board,FRB) using character unigrams help to remove several boundary false alarms (the area with arrow in Fig. 9). Anothercontribution of character unigram attributes to the compositional building feature in most Chinese words, i.e., the meaningof a word is related to the meaning of its component characters, as described in Section 4.1. Different words with relatedsemantic meaning can be matched together through their component characters.

When comparing different subword scales, we observe that trigram and quadgram perform worse than the word scaleand they do not have enough representative power to show effective story segmentation. This is because majority of Chinesewords are one-character or two-character long, as shown in Table 9. Hence using trigram and quadgram significantly reducesthe matching probability of these words.

If we compare the story segmentation performance between syllables and characters for each n-gram scales, there is nosignificant difference except for unigrams. We can observe that syllable unigrams perform much worse than characterunigrams. When we change from character unigram to syllable unigram in the lexical chaining method, the F1-measuredegrades dramatically from 0.5723 to 0.4747 for the TDT2 corpus, from 0.5979 to 0.4637 for the CCTV detailed news. Theinferior performance is due to the fact that the large number of homophone characters in Chinese renders the syllableunigram units less discriminative. However, with the increase of word length, the number of homophones also drops

Table 8Performance comparison between word and subword for lexical chaining on manual transcripts.

Transcripts F1-measure Subword scale for best F1-measure

Word Subword Relative improv. (%)

TDT2-Ref 0.5157 0.5723 11.0 Character unigram

CCTV-Ref Detailed news 0.5953 0.5979 0.4 Character unigramBrief news 0.5375 0.5917 11.3 Character unigram

0 50 100 150 200 250 300 350 4000

0.10.20.30.40.50.60.70.80.9

1

Time (# of words)0 50 100 150 200 250 300 350 400

00.10.20.30.40.50.60.70.80.9

1

Time (# of words)

Lexical similarityDepth score

Lexical similarityDepth score

ThresholdThreshold

Word CharacterUnigram

Fig. 9. Lexical similarity and depth score curves calculated on word representation and character unigram representation (right) of a CCTV transcriptexcerpt.

Table 9Statistics for word and base syllable homophones in the manual transcripts of TDT2 and CCTV. Diff. pron.: different pronunciations.

Word length Word count Diff. pron. Ratio Word count Diff. pron. Ratio

Total Unique Total Unique

TDT2 CCTV1 152675(39%) 2229(13%) 382 5.835 121962(37%) 2517(14%) 386 6.5212 204321(52%) 10117(60%) 8682 1.165 180141(55%) 12566(68%) 10764 1.1673 23061(6%) 2451(14%) 2313 1.060 16692(5%) 2070(11%) 2060 1.0044 7139(2%) 1391(8%) 1391 1.0 5022(1.5%) 1099(6%) 1099 1.04+ 3407(1%) 808(5%) 808 1.0 1698(0.5%) 212(1%) 212 1.0

2884 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 13: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

significantly, as shown in Table 9. This reveals that when the subword length (n) is increased, syllable sequence has similardiscriminative power with character sequence and they lead to a comparable performance in story segmentation.

Previous work has shown that the TextTiling method outperforms the lexical chaining method in story segmentation. Wecan also obtain this observation from the performances of word and lower n-gram transcripts in Fig. 8. However, lexicalchaining surprisingly surpasses TextTiling for quadgram in TDT2 and trigram (brief news) and quadgram in CCTV. For exam-ple, quadgram TextTiling shows very poor performance to CCTV brief news with the lowest F1-measure of 0.1319. That is tosay, TextTiling is more sensitive to large n as compared to lexical chaining. This may be explained as follows. As mentionedbefore, since most Chinese words are one or two characters, there are few repeating (matching) pairs in a short distance forn-grams with n P 3. As we know, long Chinese words tend to be more discriminative in topics because they are usually prop-er nouns; which, however, are usually scattered within a story. In trigram and quadgram cases, the local pairwise windowcomparison strategy of TextTiling results in sustaining low lexical scores due to the lack of matches and story boundariescannot stand out. However, the chaining method can link up repeating long terms (e.g. proper nouns) in a relatively longerrange and thus is not sensitive to the local matching sparseness problem. Fig. 10 shows a real example from the CCTV corpus.For character unigram, we can see that story boundaries can pop out in both the lexical similarity curve and the chainstrength curve. Character quadgrams show a different story. For character quadgram, lexical similarity cure remains flat (al-most zero) over time due to rare matches; while chain strength can still present decent peaks at story boundaries despite of alarge reduction of matches.

We also observe the performance difference between brief news and detailed news from the CCTV corpus. For the samelexical scale and story segmentation method, the F1-measure for detailed news is generally higher as compared with that ofbrief news. From Fig. 7, we can see that the length of a brief news is much shorter than a detailed news. For brief news ses-sions, we observe that some stories even last for only one or two sentences and few repeating pairs can be found in such ashort range, resulting in low similarities or burst chain ends/starts. This affects the stand-out of brief news boundaries thatshould be salient with low similarities or high concentration of chain ends and starts.

5.4. Effectiveness of subwords on LVCSR transcripts

Experimental results on LVCSR transcripts for TDT2 and CCTV are shown in Fig. 11. We can see that the observations onerroneous LVCSR transcripts are consistent with the manual transcripts. Many unigrams and bigrams bring considerable per-formance gains as compared with words on speech recognition transcripts with various error rates and the performances oftrigram and quadgrams cannot catch up with words. As shown in Tables 10 and 11, character unigram performs the best onmost LVCSR transcripts, followed by character and syllable bigrams. The average relative performance gains achieved bythese superior subwords are 7% for TDT2 and 5.5% for CCTV respectively. Besides their language advantages such as semanticmatching, the robustness of these subword units to speech recognition errors and OOV words contribute to the performancegains over words, as studied in Section 4.

We observe many examples that show the partial matching merit of subwords in the LVCSR transcripts. Fig. 12 plots thelexical similarity and depth score curves for two LVCSR transcript excerpts in the TDT2 and CCTV corpora. We can see that

0 500 1000 1500 2000 2500 3000

0

0.2

0.4

0.6

0.8

1

Time (# of words)

Lexi

cals

imila

r ity

0 500 1000 1500 2000 2500 30000

10

20

30

40

50

60

70

Time (# of words)

Cha

inst

reng

t h

Character quadgram

Character unigram

Character quadgram

Character unigram

TextTiling

LexicalChaining

Fig. 10. Comparison between TextTiling and lexical chaining on character unigram and quadgram transcripts of a CCTV broadcast news clip. Vertical redlines denote story boundaries. (For interpretation of the references in colour in this figure legend, the reader is referred to the web version of this article.)

L. Xie et al. / Information Sciences 181 (2011) 2873–2891 2885

Page 14: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

the lexical similarity curves of syllable bigram in excerpt A and the character unigram in excerpt B show much clearer sim-ilarity valleys as compared with that of words. For example, the boundary between Story A4 and A5 (missed by word-basedTextTiling) is successfully detected by syllable bigram-based TextTiling. Also, due to the partial matching of subwords,

0.60

26

0.6 0

9

0.62

55

0.49

0.34

8 6

0.59

97

0.5 8

38

0.62

72

0 .50

3

0 .36

78

0 .54

6

0.56

17

0.44

7

0.41

96

0.56

37

0.4 7

41 0.53

88

0.4 5

17

0.41

83

0 .5 7

7

0.61

92

0.62

55

0.28

31

0.13

69

0.54

63

0.57

1

0.59

45

0 .29

33

0.14

28

0.51

17

0.55

36

0.51

0 8

0.46

62

0.43

530.4 9

89

0.5 3

11

0.5 0

63

0.47

2

0.43

0 20.53

61

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

Word Unigram Bigram Trigram Quadgram

CCTV-LVCSR1

0 .59

39

0.6 1

78

0.61

23

0.50

43

0 .35

1 2

0.58

75

0 .5 7

43

0 .61

7

0.51

41

0.35

4 3

0.5 2

0.53

35

0.49

1 9

0 .48

3 4

0 .52

27

0.48

56

0.52

18

0 .48

2 1

0 .48

3 40 .5 5

36

0.60

5 5

0.58

32

0 .27

43

0 .14

04

0.54

34

0.5 4

96

0.58

32

0.30

63

0.15

06

0 .52

43

0 .54

84

0 .50

04

0 .46

8 1

0.43

970.5 2

51

0 .52

67

0 .50

69

0 .47

65

0.44

890.52

16

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8CCTV-LVCSR2

0 .5 6

32

0.59

39

0.56

29

0.42

25

0.29

06

0.56

75

0.56

72

0.56

6

0.43

24

0.29

7

0.52

13

0.50

57

0.48

05

0.4 5

180.5 1

6

0.44

72

0.46

78

0.47

25

0.45

770 .55

24

0.6 0

05

0.56

99

0.25

53

0.11

29

0 .54

52

0.55

38

0.56

99

0 .28

52

0 .13

76

0.4 8

81 0.54

67

0.5 2

69

0.46

4 2

0.42

950.50

76

0 .5 2

45

0.5 1

18

0.46

6 7

0.43

70.52

4 9

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

Word Unigram Bigram Trigram Quadgram

CCTV-LVCSR3

Word Unigram Bigram Trigram Quadgram

Character TextTiling

Syllable TextTiling

Character Lexical Chaining

Syllable Lexical Chaining

Character TextTiling

Syllable TextTiling

Character Lexical Chaining

Syllable Lexical Chaining

Brief newsDetailed news

0.53

19 0.56

04

0.57

87

0.4 7

54

0 .3 6

89

0.5 2

17

0.54

38 0 .57

76

0.48

8 7

0.38

45

0.51

36 0.55

07

0.5 0

33

0.47

870.51

7

0 .4 5

7 9

0.53

5 5

0.50

37

0.47

45

0 .54

38

0.35

0.4

0.45

0.5

0.55

0.6

0.65

0.7

Word Unigram Bigram Trigram Quadgram

TDT2-LVCSR Character TextTilingSyllable TextTilingCharacter Lexical ChainingSyllable Lexical Chaining

Fig. 11. Story segmentation results (F1-measure) for speech recognition transcripts on TDT2 and CCTV.

2886 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 15: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

several boundary false alarms are removed. Table 12 lists some subword matching examples which help to recover wordrelations in the same LVCSR transcripts with Fig. 12. The word matching failures because of OOV, recognition error and wordsegmentation are recovered by syllable bigram matching or character unigram matching. For example, the two speech rec-ognition results ‘‘ ’’ and ‘‘ ’’, of the OOV word ‘‘ ’’ (Kursk) share the same syllable bi-gram /si-ke/. Due to the flexibility of Chinese word segmentation, ‘‘ ’’ (nuclear reactor) is segmented as three

Table 10Performance comparison between word and subword for TextTiling on LVCSR transcripts.

Transcripts F1-measure Subword scale with best F1-measure

Word Subword Relative improv. (%)

TDT2-LVCSR 0.5319 0.5787 8.8 Character bigramCCTV-LVCSR1 Detailed news 0.6026 0.6272 4.1 Syllable bigram

Brief news 0.5770 0.6192 7.3 Character bigram

CCTV-LVCSR2 Detailed news 0.5939 0.6178 4.0 Character unigramBrief news 0.5536 0.6055 9.4 Character unigram

CCTV-LVCSR3 Detailed news 0.5632 0.5939 5.6 Character unigramBrief news 0.5524 0.6005 8.7 Character unigram

Table 11Performance comparison between word and subword for lexical chaining on LVCSR transcripts.

Transcripts F1-measure Subword scale for best F1-measure

Word Subword Relative improv. (%)

TDT2-LVCSR 0.5136 0.5507 7.2 Character bigramCCTV-LVCSR1 Detailed news 0.546 0.5617 2.9 Character unigram

Brief news 0.5117 0.5536 8.2 Character unigram

CCTV-LVCSR2 Detailed news 0.5200 0.5335 2.6 Character unigramBrief news 0.5243 0.5484 4.6 Character unigram

CCTV-LVCSR3 Detailed news 0.5213 0.5249 0.7 Character unigramBrief news 0.4881 0.5467 12.0 Character unigram

0 50 100 150 200 250 300 350 400 450 5000

0.1

0.2

0.3

0.4

0.5

0 50 100 150 200 250 300 350 400 450 5000

0.1

0.2

0.3

0.4

0.5

0 50 100 150 200 250 300 3500

0.050.1

0.150.2

0.250.3

0.350.4

0.450.5

0 50 100 150 200 250 300 3500

0.050.1

0.150.2

0.250.3

0.350.4

0.450.5

Time (# of words) Time (# of words)

Time (# of words) Time (# of words)

Word

Syllable bigramWord

Character unigram

Story A1 Story A2 Story A3 Story A4 Story A5

Story B1 Story B2 Story B3 Story B4 Story B5

Story A1 Story A2 Story A3 Story A4 Story A5

Story B1 Story B2 Story B3 Story B4 Story B5

A

B

Fig. 12. Word and subword lexical similarity curves (blue) and depth score curves (green) for two LVCSR transcript excerpts in the TDT2 and CCTV corpus.(For interpretation of the references in colour in this figure legend, the reader is referred to the web version of this article.)

L. Xie et al. / Information Sciences 181 (2011) 2873–2891 2887

Page 16: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

words, i.e., ‘‘ ’’ (nuclear), ‘‘ ’’ (reaction) and ‘‘ ’’ (stack). Using syllable bigram we can recover their relations. Thesesyllable bigram matches help TextTiling successfully detect the boundary between Story A4 and A5.

We observe that there are only minor performance differences between syllable and character at the same n-gram scale,except for unigrams. As discussed before, there are homonyms with the increase of n. The discriminative ability betweenlonger syllable sequences and their character counterparts are comparable. That is, the discriminative ability of subwordn-grams comes from the sequential contextual information. The inferior performance of syllable unigram is due to the largenumber of homophones and the lack of contextual information.

When comparing TextTiling with lexical chaining, we achieve the same conclusion with the error-free transcripts: TextT-iling wins at lower n-grams and lexical chaining excels at higher n-grams. The explanations can be found in Section 5.3. If wecompare between brief news and detailed news in CCTV, we find different observations with error-free transcripts: briefnews and detailed news are comparable in F1-measure and sometimes brief news sessions even achieve higher F1-measure.From Tables 10 and 11, we see that relative improvements achieved by the use of subwords is more salient for brief news ascompared with detailed news and the major performance gain attributes to the character unigram. This is probably becausebrief news session may have fewer speech recognition errors as compared with detailed news session that have complicatedaudio scenes (e.g. field speech).

Table 12Examples show how subword matching helps to recover word relations in the same LVCSR transcript excerpts in Fig. 12.

Story # Type Original word Speech recognition result

A1 OOV

A3 OOV

A4 Recognition error

A5 OOV

Word segmentation

B1 OOV

B2 OOV

OOV

B3 Recognition error

B4 OOV

B5 OOV

Word TextTilingWord Lexical Chaining

Character TextTilingSyllable TextTilingCharacter Lexical ChainingSyllable Lexical Chaining

Character TextTilingSyllable TextTilingCharacter Lexical ChainingSyllable Lexical Chaining

Baseline

Unigram

Bigram

0.45

0.47

0.49

0.51

0.53

0.55

0.57

0.59

0.61

0 37%TDT2-Ref TDT2-LVCSR

F1-m

easu

re

Fig. 13. Relationship between story segmentation performance (F1-measure) and speech recognition performance (WER) on the TDT2 corpus.

2888 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 17: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

5.5. LVCSR error rate versus story segmentation performance

We have demonstrated the effectiveness of Chinese subword representations in both error-free transcripts and LVCSRtranscripts. We further examine the sensitivity of different subwords to speech recognition errors. We made comparisononly on unigrams, bigrams and words because trigrams and quadgrams did not show decent performance, as discussedbefore. Figs. 13 and 14 show the relationships between story segmentation performance (in terms of F1-measure) and LVCSRword error rate (WER) on the TDT2 corpus and the CCTV corpus, respectively. We combine the results on brief news anddetailed news for the CCTV corpus in order to achieve a general impression on the relation between LVCSR error rate andstory segmentation performance.

Not surprisingly, the story segmentation performance is affected by the speech recognition performance. In general, lowerspeech recognition error rate leads to inferior F1-measure for both the word and subword scales. For example, when worderror rate climbs to 40.9% from error-free on the CCTV corpus, the F1-measure of the word TextTiling degrades from 0.6147to 0.5586. The degradation of story segmentation performance is largely due to the broken lexical cohesion among words.However, using subwords can restrain the performance degradation caused by noisy texts. From the TDT2 results in Fig. 13,all unigram and bigram approaches outperform the word baselines, except for the syllable unigram-based lexical chaining.Similarly, many subword approaches outperform the word baselines in the CCTV corpus, as shown in Fig. 14. For instance,we observe from the CCTV results that character unigram TextTiling on the transcripts with WER of 25% and 33.4% even out-perform word TextTiling on error-free transcripts. The syllable bigram TextTiling on the transcript with WER of 37% outper-forms the word TextTiling on the error-free transcripts (0.5787 versus 0.5589) for the CCTV corpus. Such performance gainsare particularly significant for story segmentation on LVCSR transcripts with speech recognition errors.

5.6. Fusion of word and subword representations

Words have specificity in describing meanings, but show low robustness to speech recognition errors; subwords arerobust to speech recognition errors, but do not have enough specificity in describing meanings. Hence we conduct word-subword fusion TextTiling experiments in order to show the complimentarity between different lexical representations.

Word TextTilingWord Lexical Chaining

Character TextTilingSyllable TextTilingCharacter Lexical ChainingSyllable Lexical Chaining

Character TextTilingSyllable TextTilingCharacter Lexical ChainingSyllable Lexical Chaining

Baseline

Unigram

Bigram0

.5

58

90

.5

319

0.

61

47

0.

59

07

0.5

77

10

.5

58

60

.5

97

30.

57

87

0.

65

49

0.

61

40.

612

40.

59

67

0.6

19

0.

59

01

0.

68

34

0.

63

27

0.

62

58

0.

61

06

0.50.550.60.650.70.750.8

Word-BasedT e x t T i l i n g

Subword-basedT e x t T i l i n g

Fusion-basedT e x t T i l i n g T D T 2 - R e f C C T V - R e f T D T 2 - L V C S R C C T V - L V C S R 2 C C T V - L V C S R 3 C C T V - L V C S R 1

Fig. 15.Fusion results of word and subword representations in TextTiling on TDT2 and CCTV corpora.L. Xie et al. / Information Sciences 181 (2011) 2873–28912889

Page 18: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

Specifically, we sum the lexical scores between the best-performed subword scale and the word scale, and boundary iden-tification is carried out on the corresponding depth score. For example, for CCTV-LVCSR1, lexical score at the character-bigram level is summed with the word level lexical score. Experimental results are shown in Fig. 15. We can clearly see thatword–subword fusion can consistently improve story segmentation performance on the two corpora. The best performanceis achieved on CCTV-Ref (F-measure = 0.6834), with relative improvements of 10% and 4.4% over the corresponding word andsubword based TextTiling, respectively.

6. Conclusions

In this paper, we have proposed to use Chinese subword representations, i.e. character and syllable n-gram units in lexicalcohesion-based story segmentation of Chinese broadcast news transcripts. Different from western languages, Chinese char-acters and syllables play important semantic roles. Subwords are robust to speech recognition errors and can recover lexicalcohesion in erroneous text via partial matching. We have studied the merits of Chinese subwords and performed an inves-tigation on the effectiveness of subwords in lexical cohesion methods for story segmentation. We experimented for:

1. different subword representations: character and syllable n-grams;2. different embodiments of lexical cohesion: TextTiling and lexical chaining;3. different textual transcripts: manual transcripts (error-free) and speech recognition transcripts with different error rates;4. two corpus: the TDT2 corpus and the CCTV corpus.

We have found that measuring lexical cohesion at subword levels is more effective than the word level. From extensivestory segmentation experiments, we are able to reach some important findings:

1. Most unigrams and bigrams achieve considerable performance gains while trigrams and quadgrams perform worse ascompared with words on both manual transcripts and LVCSR transcripts.

2. Character unigram and syllable bigram achieve superior story segmentation performances in general. For example, syl-lable bigram-based lexical chaining obtains a relative F1-measure improvement of 11% over word-based lexical chainingin TDT2 error-free manual transcript; character unigram-based lexical chaining achieves a relative F1-measure gain of12% over word-based lexical chaining in CCTV brief news transcripts with a word error rate of 40.9%.

3. TextTiling outperforms lexical chaining at lower n-grams and lexical chaining surpasses TextTiling at higher n-grams.TextTiling is more sensitive to large n as compared with lexical chaining in story segmentation.

4. The performance of lexical cohesion based story segmentation is affected by speech recognition errors. F1-measuredegrades with the increase of speech recognition error rate. Using subwords can contain performance degradation. Forexample, character unigram TextTiling on the CCTV transcripts with WER of 25% and 33.4% even outperform word TextT-iling on error-free transcripts.

Our study shows that subword-based lexical cohesion approaches are promising in improving performance of story seg-mentation of Chinese broadcast news. The promising results indicate that, to further improve story segmentation perfor-mance in the future work, it is possible to apply subword features to other lexical methods or integrate subword featureswith features from other sources (i.e. acoustic and visual). Also, we desire an automatic parameter tuning method to achievea more robust and smarter story segmentation system.

Acknowledgements

This paper was partially supported by a Grant from the Research Grants Council of Hong Kong, China (CityU 118608),CityU Grant 7008026 and a Grant from the National Natural Science Foundation of China (60802085).

References

[1] J. Allan (Ed.), Topic Detection and Tracking: Event-Based Information Organization, Kluwer Academic Publishers, 2002.[2] S. Banerjee, I.A. Rudnicky, A TextTiling based approach to topic boundary detection in meetings, in: Interspeech: Annual Conference of the

International Speech Communication Association, 2006, pp. 57–60.[3] D. Beeferman, A. Berger, J. Lafferty, Statistical models for text segmentation, Machine Learning 34 (1-3) (1999) 177–210.[4] S.K. Chan, L. Xie, H. Meng, Modeling the statistical behavior of lexical chains to capture word cohesiveness for automatic story segmentation, in:

Interspeech: Annual Conference of the International Speech Communication Association, 2007, pp. 2581–2584.[5] B. Chen, H.M. Wang, L.S. Lee, Discriminating capabilites of syllable-based features and approaches of utilizing them for voice retrieval of speech

information in Mandarin Chinese, IEEE Transactions on Speech and Audio Processing 10 (5) (2002) 202–214.[6] B. Chen, H.M. Wang, L.S. Lee, Spoken document retrieval and summarization, in: Advance in Chinese Spoken Language Processing, 2007, pp. 301–320.[7] F.Y.Y. Choi, Advances in domain independent linear text segmentation, in: Human Language Technology Conference – North American Chapter of the

Association for Computational Linguistics Annual Meeting, 2000, pp. 26–33.[8] S. Dharanipragada, M. Franz, J. Mccarley, S. Roukos, T. Ward, Story segmentation and topic detection in the broadcast news domain, in: Proceedings of

the DARPA Broadcast News Workshop, 1999.[9] G. Fu, C. Kit, J.J. Webster, Chinese word segmentation as morpheme-based lexical chunking, Information Sciences 178 (2008) 2282–2296.

2890 L. Xie et al. / Information Sciences 181 (2011) 2873–2891

Page 19: Author's personal copy - nwpu-aslp.orglxie.nwpu-aslp.org/papers/2011-INS-A1-SCI-EI-JNL.pdf · Author's personal copy Many news programs use cue terms such as stay tuned and reports

Author's personal copy

[10] J.S. Garofolo, C.G.P. Auzanne, E.M. Voorhees, The trec spoken document retrieval track: a success story, in: Text Retrieval Conference (TREC) 8, 2000, pp.16–19.

[11] M. Halliday, R. Hasan, Cohesion in English, Longman Group, New York, 1976.[12] M.A. Hearst, TexTiling: segmentation text into multi-paragraph subtopic passages, Computational Linguistics 23 (1) (1997) 33–64.[13] W. Hsu, S.F. Chang, A statistical framework for fusing mid-level perceptual features in news story segmentation, in: International Conference on

Multmeida and Expo, vol. 1, 2003, pp. 413–416.[14] W. Hsu, S.F. Chang, C.W. Huang, L. Kennedy, C.Y. Lin, G. Iyengar, Discovery and fusion of salient multi-modal features towards news story

segmentation, in: Proceedings of SPIE, vol. 5307, 2004, pp. 244–258.[15] W. Hsu, L. Kennedy, S.-F. Chang, M. Franz, J. Smith, Columbia-IBM news video story segmentation, in: TRECVID 2004, Tech. rep., Columbia ADVENT

Technical Report 209-2005-3, New York, 2005.[16] P.-Y. Hui, X.-O. Tang, H. Meng, W. Lam, X. Gao, Automatic story segmentation for spoken document retrieval, in: The 10th IEEE International

Conference on Fuzzy Systems, 2001, pp. 1319–1322.[17] L.-S. Lee, B. Chen, Spoken document understanding and organization, IEEE Signal Processing Magazine 22 (5) (2005) 42–60.[18] G.A. Levow, Assessing prosodic and text features for segmentation of Mandarin broadcast news, in: Proceedings of HLT-NAACL 2004 Workshop, 2004,

pp. 28–32.[19] G.A. Levow, Combining prosodic and text features for segmentation of mandarin broadcast news, in: O. Streiter, Q. Lu, (Eds.), ACL SIGHAN Workshop

2004, Association for Computational Linguistics, Barcelona, Spain, July 2004, pp. 102–108.[20] G.A. Levow, Prosody-based topic segmentation for mandarin broadcast news, in: Human Language Technology Conference – North American Chapter

of the Association for Computational Linguistics Annual Meeting, 2004, pp. 137–140.[21] Z. Liu, Q. Huang, Adaptive anchor detection using on-line trained audio/visual model, in: Proceedings of SPIE, 2000.[22] W.K. Lo, H. Meng, P.C. Ching, Multi-scale spoken document retrieval for Cantonese broadcast news, International Journal of Speech Technology 7 (2–3)

(2004) 1381–2416.[23] K. Ng, V.W. Zue, Subword-based approaches for spoken document retrieval, Speech Communication 32 (3) (2000) 157–186.[24] Y. Ni, L. Xie, Z.-Q. Liu, Minimizing the expected complete influence time of a social network, Information Sciences 180 (13) (2010) 2514–2527.[25] H.-J. Oh, S.-H. Myaeng, M.-G. Jang, Semantic passage segmentation based on sentence topics for question answering, Information Sciences 117 (2007)

3696–3717.[26] J.M. Ponte, W.B. Croft, Text segmentation by topic, in: Proceedings of ECDL, 1997, pp. 113–125.[27] A. Rosenberg, J. Hirschberg, Story segmentation of broadcast news in English, Mandarin and Arabic, in: North American Chapter of the Association for

Computational Linguistics Annual Meeting, 2006, pp. 125–128.[28] E. Shriberg, A. Stolcke, D. Hakkani-Tür, G. Tür, Prosody based automatic segmentation of speech into sentences and topics, Speech Communication

(2000).[29] N. Stokes, J. Carthy, A. Smeaton, SeLeCT: a lexical cohesion based news story segmentation system, Journal of AI Communication 17 (1) (2004) 3–12.[30] TRECVID2003, 2003. <http://www-nlpir.nist.gov/projects/tv2003/>.[31] TRECVID2004, 2004. <http://www-nlpir.nist.gov/projects/tv2004/>.[32] G. Tür, D. Hakkani-Tür, Integrating prosodic and lexical cues for automatic topic segmentation, Computational Linguistics 27 (1) (2001) 31–57.[33] X.-Z. Wang, C.-R. Dong, Improving generalization of fuzzy if-then rules by maximizing fuzzy entropy, IEEE Transactions on Fuzzy Systems 17 (3) (2009)

556–567.[34] X.-Z. Wang, C.-R. Dong, T. Fan, Training t-s norm neural networks to refine weights for fuzzy if-then rules, Neurocomputing 70 (2007) 2581–2587.[35] X.-Z. Wang, E. Tsang, S. Zhao, D. Yeung, Learning fuzzy rules from fuzzy examples based on rough set techniques, Information Sciences 177 (20) (2007)

4493–4514.[36] X.-Z. Wang, J.-H. Zhai, S.-X. Lu, Induction of multiple fuzzy decision trees based on rough set technique, Information Sciences 178 (16) (2008) 3188–

3202.[37] C.-H. Wu, C.-H. Hsieh, Story segmentation and topic classification of broadcast news via a topic-based segmental model and a genetic algorithm, IEEE

Transactions on Audio, Speech and Language Processing 17 (8) (2009) 1612–1623.[38] L. Xie, Discovering salient prosodic cues and their interactions for automatic story segmenation in Mandarin broadcast news, Multimedia Systems 14

(2008) 237–253.[39] L. Xie, Z.-H. Fu, W. Feng, Y. Luo, Pitch-density-based features and an SVM binary tree approach for multi-class audio classification in broadcast news,

Multimedia Systems (2010).[40] L. Xie, Z.-Q. Liu, A coupled HMM approach for video-realistic speech animation, Pattern Recognition 40 (10) (2007) 2325–2340.[41] L. Xie, H. Meng, Combined use of speaker and tone-normalized pitch reset with pause duration for automatic story segmentation in Mandarin

broadcast news, in: Human Language Technology Conference – North American Chapter of the Association for Computational Linguistics AnnualMeeting, 2007, pp. 193–169.

[42] L. Xie, Y. Yang, J. Zeng, Subword lexical chaining for automatic story segmentation in Chinese broadcast news, in: 9th Pacific Rim Conference onMultimedia, LNCS, vol. 5353, Springer, 2008, pp. 248–258.

[43] L. Xie, J. Zeng, W. Feng, Multi-scale texttiling for automatic story segmentation in Chinese broadcast news, in: Aisa Information Retreival Conference,LNCS, vol. 4993, Springer, 2008, pp. 345–355.

[44] J. Yamron, I. carp, L. Gillick, P. Mulbregt, A hidden Markov model approach to text segmentation and event tracking, in: IEEE International Conferenceon Acoustics, Speech and Signal Processing, 1999, pp. 333–336.

[45] J. Zeng, W. Feng, L. Xie, Z.-Q. Liu, Cascade Markov random fields for stroke extraction of Chinese characters, Information Sciences 180 (2010) 301–311.

L. Xie et al. / Information Sciences 181 (2011) 2873–2891 2891