Printer Friendly

Study on hot topic discovery from Chinese texts.

1. Introduction

With the rapid development of the Internet, text information has become widely used in the creation of web pages. Pieces of relevant Internet information on the same topic are separated in different places and at different times. Achieving a reliable and common cognition with isolated information is difficult [1]. Hence, obtaining vast amounts of information and extracting the hot topics have become important problems requiring immediate solutions. Topic detection and tracking (TDT) satisfies the demand of selecting information on the same topic within a reasonable time. The main task of TDT is to facilitate effective information classification.

Hot topic discovery is the premise and foundation of TDT. A topic is usually defined as an event on which people focus publicly. The hot topics are the highly condensed summary of the texts; they appear, develop, and disappear within a certain period and are characterized as having obvious timeliness. The extraction of the hot topic words is helpful in quickly determining the useful information. Currently, the main technique of hot topic discovery research includes hot TDT [2,3], new event detection (NED) [4], hierarchical topic detection (HTD) [5], and so on. The adopted approaches include the clustering algorithm, natural language processing, and lexical semantic chaining (LSC).

This paper discusses hot topic discovery from Chinese text that combines statistics and linguistic technology. The rest of the paper is organized as follows: The related background is introduced in Section 2, the proposed approach is described in Section 3, the experiment is presented and the results are analyzed in Section 4, and the study is concluded in Section 5.

2. Background

The study of hot TDT technology originated in the 1970s, and focused on the research of correlation detection, topic detection, topic tracking, cross-language TDT research, and so on.

NED is an important part of topic detection. Allan [4] and Yang [6] established online identification systems to inspect emerging events. Meanwhile, another work [6] defined a new topic detection task, i.e., HTD. HTD adopted the non-circulation of a root node (directed acyclic graph) to describe the hierarchical topics included. Allan [4] was one of the earliest scholars who used natural language processing (NLP) technology to solve the TDT problem. NLP technology uses the vector space model (VSM) to describe the topics and stories; it provides higher weights to the named entity in the model to imply the NED. Hasan [8] used LSC to describe lexical cohesion, while another study [8] proposed that any word in a sentence could be related to multiple other words in that chain. Morris and Hirst [9] designed the algorithm of automatic construction of LSC based on lexical resources. The lexical chains provided a semantic context for interpreting words, concepts, and sentences [8]. The idea of LSC can be used on topic discovery.

Meanwhile, various dimension-reduction methods have been used to detect topics. For example, latent semantic analysis using singular value decomposition based on VSM [10] compresses a highly dimensional vector space to a lower space. Sriurai [11] applied the topic-model approach to cluster the words into a set of topics, in which words assigned into the same topic must be semantically related. Then these hot topics were used in text categorization. The experimental results showed that the classification model based on hot topics yielded the best performance. Grun [12] used the probabilistic model of R package to obtain the topics in documents. Khodra [13] employed the sentential features to extract the topic sentences of a paragraph. The experiment showed that position and meta-discourse features were important in extracting the topic sentence.

2. The Proposed Approach

The use of statistical method based on the hot word extraction technology has become common. However, the expression of the hot word is usually unclear with a single word. The proposed idea of hot topic discovery is to utilize the hot single words and explore them forward and backward. The approach then forms the long two- and three-word pairs, from which hot topic words are discovered more efficiently. For example, "Beijing" and " Olympics" are two hot words, but when the proposed approach is used, the more useful hot topic, "Beijing Olympics," can be generated.

3.1 Data Preprocessing

Each Chinese text should be cut into words when the statistical method is employed. ICTCLAS [14] is used to conduct the word segmentation, and the space vector model is adopted to express the document and select the single hot topic words. In this process, the stop words are deleted. Further processing is conducted to remove the words that are not suitable for the topic words based on the parts of speech, such as the following characteristics of a word: /r (pronoun), /q (quantifier), /p (preposition), /c (conjunction), /u (auxiliary word), /e (interjection), /y (modal words), /o (onomatopoetic word), /k (suffix), /x (character string), and /w (punctuation). Finally, term frequency (TF) and document frequency (DF) are used to decrease the dimension.

D = {[d.sub.1], [d.sub.2], [d.sub.n]} is the document collection, F = {[w.sub.1], [w.sub.2], ...., [w.sub.[absolute value of F]]} is the feature set of document collection, and [absolute value of F] represents the total number of the features. Each document d is represented as a vector d = ([d.sub.(1)], [d.sub.(2)], ... [d.sub.([absolute value of F])]), and [d.sub.(i)] is the weight of feature w. in document d. Each feature value in the vector is calculated using the formula [TF.sup.*] IDF (inverse document frequency).

3.2 Combination Rules of The Hot Topic Words

The combination rules of the hot topic words are introduced based on the linguistic characteristic of the Chinese text. If a word "W is a single hot topic, moving one step forward or backward around it forms the hot topic of the two-word phase. Moving two steps forward and backward around "W forms the hot topic of the three-word phase. Sample formats would be "Pre + W, "W+ Be" and "Pre + W + Be," where "W is a word, "Pre" is the word before it, and "Be" is the word after it. If only the abovementioned method is used, some incorrect combinations may be generated because the grammar rules are not considered. For this reason, the phrase-structure analysis method is used to define the combination rules before the first step, and then the correctness of the word combination is further judged. For example, the three-word phrase made up of "noun + noun + noun" is considered correct, whereas the "noun + adjective + adjective" combination is incorrect. The combination rules of the two- and three-word phrases are shown in Table 1.

3.3 Detailed Algorithm

The feature selection is conducted based on the statistical idea of obtaining the word and document frequencies. From the obtained primary candidate words, the candidate word table is built. Considering that the information expression of the individual word may be not perfect, the method combines the word pairs to form phrases and obtain more comprehensive subject information. Grammar rules are also utilized to generate the possible topic phrases. The grammar rules are shown in Table 1. One word is selected in the candidate table, and then the phrase is formed because of the adjacent relation of the words.

Finally, the weights of the phrases are computed, and the hot topics are generated. The detailed algorithm is presented below.

Input: a //TF threshold of feature selection b //DF threshold of feature selection c //the number of hot topic words


1). For each document [D.sub.i] in the text set:

2). Obtain the word table [DW.sub.i] after the word segmentation of [D.sub.i].

3). For [DW.sub.i], remove stop and single-character words, and exclude the words that are not used as the topic words dpending on the part of speech, such as the words with /r/q/p/c/u/e/y/ o/k/x/w. Then the candidate word table [CW.sub.i] is generated.

4). Set the document frequency of each word W as 1 in [CW.sub.i].

5). Divide the document [D.sub.i] into sentences and obtain the sentence table [DS.sub.i].

6). For each sentence S in [DS.sub.i]:

7). Look up each candidate word Wcontained in [CW.sub.i]; if the word W is in the sentence S, then update its word frequency (TF).

8). Add the three-word phrase "Pre + W + Be" into [CW.sub.i] depending on the combination rules, such as nnn/nan/

9). If the three-word phrase is not generated, then add the two-word phrase "Pre + W' or "W + Be" into [CW.sub.i] depending on the combination rules, such as an/nn/vn/nv/na. Update its frequency (TF).

10). End for

11). End for

12). Merge the words and phrases of all [CW.sub.i] into the DW_ Base table. Obtain the sum of the TF and DF of the words in DW_Base.

13). Calculate the weight of the word using the formula TF * IDF.

14). For multi-word phrase, its weight is equals the sum of the weights of its sub-words belonging to the DW_Base table.

15). For DW_Base, filter out the words whose TF values are less than a.

16). For DW_Base, filter out the words whose DF values are less than b.

17). Sort the three-word phrases with weight in a descending order.

18). Sort the two-word phrases with weight in a descending order.

19). Sort the words with weight in a descending order.

20). Output c three-word phrases as long-length hot topics.

21). Output c two-word phrases as middle-length hot topics.

22). Output c words as short-length hot topics.


4. Experiment and Results

Fifty XML files about sports are used in the experiment. The format of the XML file includes information, such as title, date, and content. Figure 1 shows an example of the XML file.

The application system of the Chinese version is developed based on the proposed approach. The following modes are used in discovering the hot topics: (1) extracting hot topic words from the XML file set, (2) extracting the hot topic words from the XML file set based on the date scope, and (3) extracting the hot topic words from the XML file set based on the time interval.

Figure 2 shows the interface of hot topic word extraction from the XML file set. Figure 3 shows the interface of hot topic word extraction of the XML file set from 2013/01/10 to 2013/01/14 and every day thereafter.

To show the hot topic words, four options are provided for every discovery mode, including "pre + single + be," "pre + single," "single + be," and "single". The "pre + single + be" option shows the hot topics of three words, the "pre + single" and "single + be" options of two words, and the "single" option of one word. The TF threshold is equal to 8 and the DF threshold is equal to 5. The number of hot topics is equal to 15. Figure 5 shows the result interface of the "pre + single + be" option. Figure 3 also shows the result interface of the "single" option.

The hot topic extraction is conducted on 50 XML files. Table 2 shows three-word topics and their weights among the top 10, Table 3 shows two-word topics and their weights among the top 10, and Table 4 shows one-word topics and their weights among the top 10.

Tables 2 and 3 prove that two- and three-word topics pro vide more comprehensive information, respectively, where the topic of each phrase is a combination of the adjacent words. In addition, the co-occurrence of the phrase is higher than the TF threshold, and its DF value is also higher than the DF threshold. The weight of the phrase is higher and ranks among the top 15, indicating that the phrase appears to reach a certain level. The hot topic of the phrase plays an important role in the texts and should frequently appear; this reflects the central idea, which is the advantage of the proposed approach. Table 4 shows that one-word topics are short and incomplete, hence, they cannot fully express the central idea. The abovementioned tables show that if the weight of the one-word topic is higher, the probability of its occurrence in two- and three-word topics would be larger. For example, one-word topics [TEXT NOT REPRODUCIBLE IN ASCII] and [TEXT NOT REPRODUCIBLE IN ASCII] have higher weights, and these words appear in two- and three-word topics because of the cumulative calculation of the long topic weight.

The combination of adjacent words before and after can be further improved. Only two- or three-word topics are extracted in the proposed approach. Some longer and meaningful topics can be used in future works on this topic

5. Conclusion

Hot topic discovery from Chinese texts is studied in this work to facilitate effective information classification. The proposed approach combines the statistics and linguistic grammar rules to track the hot topics of the texts. The main purpose of the proposed approach is to overcome the shortcomings of the one-word topic, which is unable to clearly express information. The candidate frequent words are obtained based on statistics. Two- and three-word topics are generated by utilizing the grammar rules and the candidate words. The application system of the Chinese version is then developed based on the proposed approach, and more hot topics are achieved by adjusting some relative parameters. The experiment results prove that the proposed method is effective in extracting two- and three-word hot topics that can express stronger and more comprehensive information. In future works, the approach should be expanded and considerably longer and more meaningful topics should be used.

Received: 17 April 2014, Revised 30 June 2014, Accepted 20 July 2014

6. Acknowledgment

This study was funded by Beijing Social Sciences Funding Program (No. 14WYB040), the National Natural Science Foundation of China (No. 61309012) and the First-Class University and First-Class Discipline Transition Construction Funds of Minzu University of China.


[1] Hu, H. (2014). Research on Ontology Construction and Information Extraction Technology Based on WordNet. Journal of Digital Information Management, 12 (2) 114-119.

[2] Holz, F., Teresniak, S. (2010). Towards automatic detection and tracking of topic change. Computational Linguistics and Intelligent Text Processing. Berlin, Germany: Spring-Verlag, p. 327-339.

[3] Song, D., Wang, W., Chen, Y. (2006). Topic Detection and Tracking with a Developed Vector Space Model. Computer Technology and Development, 9 (16) 62-67.

[4] Allan, J., Papka, R., Lavrenko, V. (1998). On-Line New Event Detection and Tracking. In: Proceedings of SIGIR' 98: 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, p. 3745. New York: ACM Press.

[5] Zhao, H., Zhao, T., Zhang, S., Wang, H. (2006). Topic detection research based on content analysis. Journal of Harbin Institute of Technology, 10 (38) 1740-1743.

[6] Yang, Y., Pierce, T., Carbonell, J. (1998). A study on Retrospective and On-Line Event detection. In: Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval, p. 28-36. CMU, USA: ACM.

[7] Feng, A., Allan, J. (2004). Hierarchical Topic Detection in TDT-2004. In: Proceeding of the 7th Conference of Topic Detection and Tracking. publication/228682428_Hierarchical_topic_detection_in_T DT-2004.

[8] Hasan, R. (1984). Coherence and cohesive harmony. Flood L, eds. Understanding Reading Comprehension. Newark, Delaware: International Reading Association, p. 181-219.

[9] Morris, J., Hirst, G.. (1991). Lexical Cohesion by The saural Relations as an Indicator of the Structure of Text. Computational Linguistics, 17(1) 21-48.

[10] Deerwester, S., Dumais, S. T., Furnas, G. W., Landauer, T. K., Harshman, R. (1990). Indexing by latent semantic analysis. Journal of the American Society for Information Science, 41 391-407.

[11] Wongkot Sriurai. (2011) Improving Text Categorization by Using a Topic Model. Advanced Computing: An International Journal, 2 (6) 21-27.

[12] Grun, B., Hornik, K. (2011). Topicmodels: An R Package for Fitting Topic Models. Journal of Statistical Software, 40 (13) 1-30.

[13] Khodra M. L., Widyantoro D.H., Aziz, E. A., Trilakson, B. R. (2011). Free Model of Sentence Classifier for Automatic Extraction of Topic Sentences. Journal of Information and Communication Technology, 5 (1) 1734.

[14] Huaping, Z., Hongkui, Y., Deyi, X., Qun, L. (2003). HHMM-based Chinese Lexical Analyzer ICTCLAS. In: The Second SIGHAN Workshop on Chinese Language Processing affiliated with 41th Association for Computational Linguistics, p. 184-187.

Guixian Xu (1), Lirong Qiu (1), Hongfang Liu (2)

(1) Minzu University of China Beijing, China (2) Mayo Clinic Minnesota, USA,

Table 1. Combination rules of the two- and
three-word phrases

Three-word phrase                Two-word phrase
combination rules                combination rules

Noun + noun + noun (nnn)         Adjective + noun (an)
Noun + adjective + noun (nan)    Verb + adv (va)
Noun + prep + noun (npn)         Verb + noun (vn)
Verb + noun + noun (vnn)         Noun + Verb (nv)
Noun + noun + verb (nnv)         Noun + noun (nn)
Noun + Verb + noun (nvn)
Adjective + noun + verb (anv)
Noun + conj + noun (ncn)
COPYRIGHT 2014 Digital Information Research Foundation
No portion of this article can be reproduced without the express written permission from the copyright holder.
Copyright 2014 Gale, Cengage Learning. All rights reserved.

Article Details
Printer friendly Cite/link Email Feedback
Author:Xu, Guixian; Qiu, Lirong; Liu, Hongfang
Publication:Journal of Digital Information Management
Article Type:Report
Geographic Code:9CHIN
Date:Aug 1, 2014
Previous Article:Integrating multidimensional information for the benefit of collaborative enterprises.
Next Article:Hierarchical community detection algorithm based on local similarity.

Terms of use | Privacy policy | Copyright © 2022 Farlex, Inc. | Feedback | For webmasters |