Next Article in Journal
Effects of Simulated Gastric Acid Exposure on Surface Topography, Mechanical and Optical Features of Commercial CAD/CAM Ceramic Blocks
Previous Article in Journal
Automated Detection of Premature Flow Transitions on Wind Turbine Blades Using Model-Based Algorithms
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Nested Chinese Restaurant Topic Model for Short Texts with Document Embeddings

Department of Computer Science and Technology, University of Science and Technology of China, Hefei 230052, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(18), 8708; https://doi.org/10.3390/app11188708
Submission received: 31 August 2021 / Revised: 15 September 2021 / Accepted: 16 September 2021 / Published: 18 September 2021
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
In recent years, short texts have become a kind of prevalent text on the internet. Due to the short length of each text, conventional topic models for short texts suffer from the sparsity of word co-occurrence information. Researchers have proposed different kinds of customized topic models for short texts by providing additional word co-occurrence information. However, these models cannot incorporate sufficient semantic word co-occurrence information and may bring additional noisy information. To address these issues, we propose a self-aggregated topic model incorporating document embeddings. Aggregating short texts into long documents according to document embeddings can provide sufficient word co-occurrence information and avoid incorporating non-semantic word co-occurrence information. However, document embeddings of short texts contain a lot of noisy information resulting from the sparsity of word co-occurrence information. So we discard noisy information by changing the document embeddings into global and local semantic information. The global semantic information is the similarity probability distribution on the entire dataset and the local semantic information is the distances of similar short texts. Then we adopt a nested Chinese restaurant process to incorporate these two kinds of information. Finally, we compare our model to several state-of-the-art models on four real-world short texts corpus. The experiment results show that our model achieves better performances in terms of topic coherence and classification accuracy.

1. Introduction

With the growth of social media and applications of mobile phones, short texts have been a kind of prevalent and important information on the internet. There is abundant semantic information that can be found in short texts. However, unlike documents with regular size, the average length of each short text is very short. This characteristic makes the knowledge discovery on short texts being a challenging research problem.
To discover knowledge from texts, the topic modeling method is widely used [1]. For documents with regular size, conventional topic models like LDA [2] and HDP [3] perform well. These methods can automatically generate topics according to word co-occurrence information. However, for short texts, word co-occurrence information is very sparse because of the short length of each text [4]. So the performances of these conventional methods are very poor on short texts [5]. To overcome this problem, many researchers propose different kinds of customized topic models dealing with the sparsity of word co-occurrence information [6].
As word co-occurrence information in short texts is sparse, researchers propose two kinds of strategies to overcome this problem. The first strategy is to incorporate auxiliary information. Metadata like hashtags or authors is available in short texts. Some methods aggregate short texts according to metadata [7,8,9]. However, these methods are effective only for a few short texts corpus because metadata is not always available. So other methods incorporate word embeddings generated from an auxiliary corpus with documents of regular length [10,11,12,13,14]. However, these methods need to construct a semantically corresponding auxiliary corpus of short texts. If the auxiliary corpus is not semantically related to target short texts, the performance will be very poor [15].
The second strategy is to design specific models without auxiliary information. Several models restrict the number of topics each short text corresponds to [16,17,18]. The local word co-occurrence information of each topic is very sparse. So when a short text corresponds to one or a few topics, the local word co-occurrence information will be less sparse. However, this strategy cannot add sufficient word co-occurrence information. Several models use the global word co-occurrence information [19,20]. These methods rebuild short texts corpus and construct a global word pairs set. The word pairs set is the summary of local word co-occurrence information. So the global word co-occurrence information will be less sparse. However, in these models, variables of documents are missing and the probability distribution between short texts and topics is unable to be obtained directly. Self-aggregated topic models [21,22] automatically aggregate short texts into latent long documents. Then topics are generated from these long documents. Long documents incorporate sufficient word co-occurrence information and make local word co-occurrence information no longer sparse. These models seem more reasonable than other strategies and auxiliary information is not needed. However, the aggregation process of short texts may incorporate non-semantic information. Short texts that are not semantically relevant may be aggregated into one long document. Then this long document will bring a lot of non-semantic word co-occurrence information into topics and makes topics incoherent.
As there exist a lot of deficiencies of the state-of-the-art topic models, we propose a self-aggregated topic model by incorporating document embeddings [23] (DESTM) to deal with the sparsity of word co-occurrence information. Document embedding information of short texts is the mapping of short texts’ content in vector space. This information contains two kinds of information: word co-occurrence information and context semantic information. Then we can get the similarities between every two short texts according to document embeddings. So, if short texts are aggregated according to the similarities calculated from document embeddings, semantically related short texts are easier to be aggregated together. Long documents generated by our model can effectively avoid incorporating non-semantic word co-occurrence information.
However, document embeddings of short texts also contain a lot of noisy information resulting from the sparsity of word co-occurrence information. So the challenge is that we need to exclude noisy information in document embeddings. Noisy information affects the value of distances between short texts and makes short texts not similar. So we changed document embeddings into global and local semantic information to discard noisy information and retain as much reliable semantic information as possible.
The global semantic information is the probability distribution of the content similarities on the entire dataset. Although the distances of similarities are affected by the noisy information, the distribution of the entire dataset will not change. For example, if the distribution of similarities is a normal distribution, noisy information cannot change this distribution to other kinds of distributions. However, the noisy information may have little impact on the parameter of the similarity distribution. So we only incorporate this distribution as the prior in order to minimize the impact of its parameters.
The local semantic information is the distances between similar short texts. The sparsity of word co-occurrence information will reduce the similarities between short texts. So we use a threshold to retain distances between short texts that are relatively similar. Distances between short texts that are not so similar are discarded.
Then we adopt the nested Chinese restaurant process to incorporate global and local semantic information in two steps. In the first step, we generate the distribution of similarities of short texts by a Chinese restaurant process. In the second step, we incorporate this distribution as the prior distribution and generate long documents according to the distances of similar short texts. Contributions of our model are as follows:
  • Firstly, our model aggregates short texts into latent long documents according to document embeddings. Document embedding information provides similarities of short texts and can avoid incorporating non-semantic word co-occurrence information.
  • Secondly, we change document embeddings into global and local semantic information to discard noisy information. Global information is the similarity probability distribution across all short texts and local information is the distances of similar short texts. Our model adopts the nested Chinese restaurant process to incorporate these two kinds of information in two steps.
  • Thirdly, we compare our model to several state-of-the-art models. By generating topics on four real-world short texts corpus, the experimental results demonstrate that our model outperforms other methods in terms of topic coherence and classification accuracy.
The remainder of this paper is organized as follows. In Section 2, we analyze the state-of-the-art methods of topic models for short texts. In Section 3, we introduce the details of our model. In Section 4, we show experiments and analyze results. In Section 5, we conclude our work and discuss future works.

2. Related Works

There are mainly two kinds of strategies. One is incorporating auxiliary information to provide additional word co-occurrence information. Another is designing specific models without auxiliary information.

2.1. Models with Auxiliary Information

For the first strategy, one kind of auxiliary information is the metadata of short texts. The AT model [7] aggregates short texts according to the author of texts. ET-LDA [24] generates topics from events and each event is constructed by associated tweets. The pooling model [8] aggregates tweets according to hashtags. Other models combine different kinds of metadata. mLDA [9] aggregates short texts according to authors and labels. rrPLSA [25] aggregates short texts according to authors and social roles. The AOTM model [26] incorporates authors and authors’ regular sized documents. However, for different kinds of short texts, metadata may not always be available. Another kind of auxiliary information is the corpus of documents with regular size. Word embeddings [27,28] are always used to incorporate semantic information from the long document set into topic models. The models tf-lda and tf-dmm [10] generate words from topics by a multi-nominal distribution or by feature vectors from word embeddings. The Gaussian LDA model [11] proposes a multivariate Gaussian distribution between topics and word vectors instead of the multinominal distribution. The GPU-DMM model [12] incorporates word embedding through the Pólya Urn strategy. The model GLTM [13] combines both local word embeddings from short texts and global word embeddings from auxiliary long documents. The TRNMF [29] model incorporates word embeddings based on regularized non-negative matrix factorization. The TSSE-DMM [30] model divides topics into four aspects and generates topics along with word embeddings. The WEPTM model [14] generates priors of the distribution between topics and words according to word embeddings. However, if the auxiliary document set is not carefully designed and becomes semantically irrelevant with target short texts, word embeddings will be noisy information and lead to poor performance [15].

2.2. Models without Auxiliary Information

The second strategy is to design specific models without auxiliary information. The GSDMM model [16] suggests that each short text only corresponds to one topic. The PDMM model [17] uses a Poisson distribution to limit the number of topics for each short text. The DSPTM model [18] restricts the number of topics for each short text by the “Spike and Slab” prior. All these models can improve the problem of the sparsity of local word co-occurrence information, but can not provide sufficient additional word co-occurrence information. The BTM model [19] rebuilds short texts into word pairs according to global word co-occurrence information. The WNTM model [20] transforms short texts into a network of words according to global word co-occurrence information. These two methods use global word co-occurrence information but cannot provide distributions between topic and short texts directly. The SATM model [21] proposes a self-aggregated topic model. This model uses two phases sampling long documents and sampling topics. However, it is time-consuming and may become overfitting. The PTM model [22] proposes a model heuristically generating latent long documents and generating topics from these long documents. This model is more simple and effective compared to SATM model. The SPTM model [22] avoids long documents being ambiguous when the number of long documents is defined as too small. Although all the self-aggregated models provide sufficient additional word co-occurrence information, the aggregation process will incorporate non-semantic word co-occurrence information inevitably.
The classification of the state-of-the-art topic models for short texts are shown in Table 1 according to their strategies.

3. Model and Inference

In this section, we introduce the details of our model, DESTM. Firstly, we introduce the generation process and the probabilistic graphical model. Then we show how we incorporate document embeddings through the nested Chinese restaurant process. Then we show the inference equations and the method for sampling latent topics and latent long documents according to Gibbs sampling.

3.1. Overview

Topic modeling methods sample topics through a joint probability distribution constructed by a generation process. We chose the conventional method LDA as an example. The LDA method designs three variables: text d, word w, and topic z. Then model constructs a generation process among these variables. It suggests that for each word w in each text d, a topic z is sampled from text d according to a multi-nominal distribution, and word w is sampled from the topic z according to another multi-nominal distribution. Then the joint probability distribution P ( d , z , w ) is define by two multi-nominal distributions P ( z | d ) and P ( w | z ) , where P ( d , z , w ) = P ( w | z ) P ( z | d ) . The two variables w and d are already known and variable z is the latent variable. So we can sample z according to the Gibbs sampling method [31].
To deal with short texts, self-aggregated topic models design a new latent variable l as the long document. The state-of-the-art model PTM proposes a generation process among four variables: short text d, word w, topic z, and long document l. Firstly, for each short text d, sample a long document l according to a multi-nominal distribution. Then for each word w of short text d, sample a topic z from a long document l according to a multi-nominal distribution. Finally, sample the word w from topic z according to another multi-nominal distribution. So, this generation process constructs a joint probability distribution P ( d , w , z , l ) according to three multi-nominal distributions P ( w | z ) , P ( z | l ) and P ( l | d ) , where P ( d , w , z , l ) = P ( w | z ) P ( z | l ) P ( l | d ) . Every short text shares a same multi-nominal distribution P ( l | d ) . So if the number of long documents is less than the number of short texts, short texts will be aggregated into long documents.
However, the aggregation according to a multi-nominal distribution cannot avoid aggregating semantically irrelevant short texts into one long document. So we propose a nested Chinese restaurant process to incorporate document embeddings instead of using a multi-nominal distribution. As the Chinese restaurant process is a specific sampling method of the Dirichlet process. So firstly, we use the Dirichlet process (DP) to describe the generalized mathematical representation of the generation process of our model. Then we will introduce details of the nested Chinese restaurant process in our model.
The generation process is as follows:
  • Sample G 0 D P ( α , H )
  • Sample G D P ( δ , G 0 )
  • For each topic z
    Sample η z D i r ( γ )
  • For each latent long document l
    Sample θ l D i r ( β )
  • For each short text i
    (a)
    Sample a long document l G
    (b)
    For each word w i , l in short text i l
    i.
    Sample a topic z M u l t i ( θ l )
    ii.
    Sample the word w i , l M u l t i ( η z )
The Dirichlet process is a stochastic process that samples distributions. In our generation process, we first sample a distribution G 0 from a Dirichlet process D P ( α , H ) , where α is the prior parameter and H is the prior distribution. Then using G 0 as another prior distribution, we sample a distribution G from a Dirichlet process D P ( δ , G 0 ) , where δ is the prior parameter. Therefore, the expectation of G is G 0 . Then, for each short text i, we sample a long document l according to distribution G. Here, G can be regarded as a probability measure that divides the space of short texts. So, short texts are aggregated into long documents according to G. Then, for each word w i , l in short text i l aggregated to long document l, we firstly sample a topic z from a multi-nominal distribution M u l t i ( θ l ) . The parameter θ l belongs to the long document l. Secondly, we sample the word w i , l from another multi-nominal distribution M u l t i ( η z ) . The parameter η z belongs to the topic z.
Through this generation process, we build the joint probabilistic distribution P ( w , d , z , l ) = P ( w | z ) P ( z | l ) G G 0 . The relations between these variables are shown in the graphical model in Figure 1.
As shown in the graphical model. N means the number of short texts. N w means the number of words in a short text. T means the number of topics and K means the number of vocabulary. Solid node W means words that are known variables. Nodes Z and L are the latent variables that we need to sample. Nodes γ , β , δ , α , and H are prior parameters that we need to define. Nodes θ and η are parameters of multi-nominal distributions. G 0 is a probability measure sampled from the Dirichlet process and a probability measure G is sampled according to prior G 0 .
In our model, we use the nested Chinese restaurant process to sample G 0 and G. Document embeddings are separated into two kinds of semantic information. One is the distribution of similarities of short texts and another is the distances of similar short texts. Then we incorporate the distribution into G 0 and the distances into G. As the expectation of G is G 0 , G will contain both two kinds of semantic information. By aggregating short texts according to the distribution G, short texts that are semantically related are more likely to be aggregated into one long document and can significantly avoid incorporating non-semantic word co-occurrence information.

3.2. Nested Chinese Restaurant Process

In our model, we use the nested Chinese restaurant process to sample G 0 and G from two Dirichlet processes. The nested Chinese restaurant process can be described as follows:
1.
Produce a Chinese restaurant r with unlimited number of tables
2.
The first customer sits at the first table
3.
The customer c sits at
(a)
The table k with probability n k α 0 + n 1
(b)
A new table with probability α 0 α 0 + n 1
4.
For each table k with customers c in restaurant r
(a)
Produce a Chinese restaurant r k with unlimited number of tables
(b)
The first customer in c sits at the first table
(c)
The customer c k in c sits at
i.
The table k * with probability n k * α 0 * + n k 1
ii.
A new table with probability α 0 * α 0 * + n k 1
Following this procedure, tables are sampled by customers. r means the Chinese restaurant for all customers. n k means the number of customers of table k. n means the number of total customers in the restaurant. α 0 is the discrete parameter that determines the dispersion of customers. c means a subset of customers. r k means the Chinese restaurant for customers in table k. n k * means the number of customers sit at table k * in restaurant r k . α 0 * is the discrete parameter that determines the dispersion of customers c .
A nested Chinese restaurant process is a two-step Chinese restaurant process. In the first step, a Chinese restaurant process assigns customers to K tables. Then, for each table k in this table set, we can get a subset of customers c that sits at the table k. So in the second step, for each customer c K c , a new Chinese restaurant process assigns customers to K * tables. Finally, for all tables generated from the first step, we generate Kth new Chinese restaurant processes in the second step. In these two steps, the first step samples a distribution G 0 . The second step incorporates G 0 as the prior and generates a new distribution G. As the expectation of G is G 0 , customers will generate tables in the second step following the distribution G 0 . In our model, short texts are customers and long documents are tables. In the next section, we will introduce how document embeddings are incorporated into our model.

3.3. Incorporating Document Embeddings

Document embedding information is calculated from short texts accord to the PV-DM method [23]. Each short text is mapped to a vector. Then we calculate the cosine distance of every two short texts. Suppose the embeddings vector of the short text a is A and the short text b is B . The cosine distance D ( a , b ) between these two short texts is calculated as follows:
D ( a , b ) = A · B | | A | | | | B | |
The cosine distance D ( a , b ) is also the similarity distance of short texts a and b. Then we incorporate the cosine distance of every two short texts into the nested Chinese restaurant process in our model.
In the first step of the nested Chinese restaurant process, we incorporate all cosine distances. Then in the second step of the nested Chinese restaurant process, we incorporate distances between similar short texts. By designing a threshold, distances of short texts that are not similar are discarded. The filtering process is as follows:
  • If D ( a , b ) η , then D s ( a , b ) = D ( a , b )
  • Else D ( a , b ) η , then D s ( a , b ) = 0
Here, η is the threshold. D ( a , b ) is the cosine distance. Because the more similar short text a and b are, the greater the value of D ( a , b ) will be. So threshold η can filter the cosine distances and retains distances of similar short texts that are greater than the value of η . Finally, we generate the filtered distance D s ( a , b )
Then, we generate the nested Chinese restaurant process according to the cosine distances. The generation process is as follows:
  • The first short texts samples the first long document
  • For each short text i
    (a)
    For each long document l *
    D ( i , l * ) = D ( i , i l * )
    (a)
    Sampling long document l * with probability n l * + D ( i , l * ) α * + n 1
    (b)
    Sampling a new document with probability α * α * + n 1
  • For each long document l * with short texts set I l *
    (a)
    The first short texts in I l * samples the first long document
    (b)
    For each short text i I l *
    i.
    For each long document l
    D s ( i , l ) = D s ( i , i l )
    i.
    Sampling long document l with probability n l + D s ( i , l ) α + n l * 1
    ii.
    Sampling a new document with probability α α + n l * 1
In the first step of the nested Chinese restaurant process, D ( i , i l * ) means the cosine distance between short text i and short text i l * in long document l * . Then, D ( i , l * ) is the summary of cosine distances between short text i and every short text in long document l * . n l * is the number of short texts in long document l * . α * is the discrete parameter that determines the dispersion of short texts. In the second step of the nested Chinese restaurant process, D s ( i , i l ) is the cosine distance between similar short texts. D s ( i , l ) is the summary of distances between short text i and all similar short texts in document l. n l is the number of short texts in long document l and α is the discrete parameter that determines the dispersion of short texts in l * .
After this nested Chinese restaurant process, we finally sample a long document l for each short text. Each long document is generated according to the global and local semantic information. The global semantic information is the distribution of the similarities of short texts generated through the first step of the nested Chinese restaurant process. The local semantic information is the cosine distances between similar short texts. Then the global semantic information becomes the prior of the second step of the nested Chinese restaurant process. Every long document l is sampled according to these two kinds of semantic information.

4. Inference

We train our model according to the collapsed Gibbs sampling method. The joint probability distribution of our model is p ( l * , l , z , w , α , β , γ , δ ) , where l * is the long document variable generated from the first step of the nested Chinese restaurant process, l is the long document variable generated from the second step of the nested Chinese restaurant process, z is the variable of the topic, w is the variable of the word, α is the dispersion prior of short texts that sample l * , β is the dispersion prior of short texts that sample l, γ is the prior of multi-nominal distribution between z and l, and δ is the prior of multi-nominal distribution between z and w. By integrating out parameters in this distribution, we sample three latent variables: the long document l * generated in the first step of the nested Chinese restaurant process, the long document l in the second step, and latent topics.

4.1. Sampling Long Documents Assignments l *

Samples of l * are generated from a Chinese restaurant process. For the ith short text in short texts corpus, the probability ϕ i of sampling long document l * is as follows:
ϕ i | ϕ 1 , . . . , ϕ i 1 , α , H l * = 1 L * n l * α + n 1 δ θ l * + α α + n 1 H
Here H is the prior distribution. L * is the number of long document samples of l * . n l * is the number of short texts in long document l * . n is the total number of short texts in the dataset. δ θ l * means the atom of long document l * . The sequence of short text is exchangeable in the Chinese restaurant process. So when sampling long document l * from any short text i, we suggest that i is the last short text in corpus. For each short text i, the probability function of sampling long document l * is designed as follows:
p ( l i * = l * | l i * , l , w , z , α , β , γ , δ ) p ( l i * | α ) p ( l i * | α )
n l * + d ( i , l * ) , l * < = L * α , l * = L * + 1
Here, l i * means the long document that short text i sampled. l i * means the long document excluding short text i. n k * means the number of short texts in long document k. d ( i , k * ) is the summary of cosine distances. We calculate d ( i , l * ) as d ( i , l * ) = d ( i , j ) . j is the short text in long document k and d ( i , j ) is the cosine distance between short text i and j.
Finally, we generate the probability of sampling l * as:
p ( l i * = l * | l i * , l , w , z , α , β , γ , δ ) n l * + d ( i , l * ) , l * < = L * α , l * = L * + 1

4.2. Sampling Long Documents Assignments l

After sampling long document l * , we will sample long documents l. Samples of l are also generated from a Chinese restaurant process. Each long document l * generates a short texts set I l * . For each short text i I l * , we aggregate short texts according to another Chinese restaurant process. The probability ϕ i of sampling long document l is as follows:
ϕ i | ϕ 1 , . . . , ϕ i 1 , β , H l * l = 1 L n l β + n l * 1 δ θ l + β β + n l * 1 H l *
Here H l * is the prior distribution generated by l * . L is the number of long document samples of l. n l is the number of short texts assigned to l. δ θ l means the atom of long document l. For each short text i I l * , we also suggest that i is the last short text in I l * . The probability function of sampling long document l is designed as follows:
p ( l i = l | l i , l * , w , z , α , β , γ , δ ) p ( l i | β ) p ( l i | β ) p ( w , z | l i , β , γ , δ ) p ( w , z | l i , γ , δ )
Here l i means the long document sample of short text i. l i means the long document excluding short text i. The equation contains two probabilities, we will generate them separately.
Firstly, we generate the probability p ( l i | β ) p ( l i | β ) as follows:
p ( l i | β ) p ( l i | β ) n l , l < = L β , l = L + 1
Secondly, we generate the probability p ( w , z | l i , β , γ , δ ) p ( w , z | l i , γ , δ ) . This probability can be simplified as follows:
p ( w , z | l i , γ , δ ) p ( w , z | l i , γ , δ ) p ( z | l i , γ ) p ( z i | l i , γ )
Here z i means the topics that excluding topics in short text i.
If long document l is sampled from an existing long document, we generate the probability p ( z | l i , γ ) p ( z i | l i , γ ) as follows:
p ( z | l i , γ ) p ( z i | l i , γ ) t i j = 1 n i , t ( n l , t i + γ + j 1 ) g = 1 n i ( n l i + T γ + g 1 )
Here t i means topics sampled from short text i. n i , t means the number of topic t in short text i. n l , t i means the number of topic t in document l excluding counts of topic t in short text i. n i means the length of short text i. n l i means the length of long document l excluding short text i. T means the total number of short texts.
If long document l is sampled from a new document, we generate the probability p ( z | l i , γ ) p ( z i | l i , γ ) as follows:
p ( z | l i , γ ) p ( z i | l i , γ ) t i j = 1 n i , t ( γ + j 1 ) g = 1 n i ( T γ + g 1 )
Finally, we generate the sampling probability of long document l as follows:
p ( l i = l | l i , l * , w , z , α , β , γ , δ ) n l t i j = 1 n i , t ( n l , t i + γ j ) g = 1 n i ( n l i + T γ g ) , l < = L β t i j = 1 n i , t ( γ + j 1 ) g = 1 n i ( T γ + g 1 ) , l = L + 1

4.3. Sampling Topics Assignments z

After sampling long documents l, we generate samples of topic z. For each word w in each short text i, we generate the probability of sampling topic z as follows:
p ( z l = z | z w , w , l , l * , α , γ , δ ) p ( z l | l , β ) p ( z w | l , β ) p ( w | z , γ ) p ( w | z w , γ )
Here z l means topic z belongs to long document l. z w means topics in corpus excluding the topic of word w. For this equation, we first generate the probability of p ( z l | l , β ) p ( z w | l , β ) as follows:
p ( z l | l , γ ) p ( z l | l , γ ) n l , z + γ 1
n l , z means the number of topics in long document l.
Then we generate the probability of p ( w | z , γ ) p ( w | z w , γ ) as follows:
p ( w | z , δ ) p ( w | z l , δ ) n w , z w + δ 1 n z w + V δ 1
Here n w , z w means the number of the same word w of topic z excluding the topic of word w. n z w means the number of total words of topic z excluding word w. V means the length of vocabulary. Finally, we obtain the probability of sampling topic z as follows:
p ( z l = z | z l , w , l , l * , β , γ , δ ) ( n l , z + γ 1 ) n w , z w + δ 1 n z w + V δ 1

4.4. DESTM Gibbs Sampling Process

After generating probabilities of sampling, we show the detailed Gibbs sampling process in Algorithm 1. The conventional iteration of the Gibbs sampling method is more than one thousand. In one iteration, for each short text, we first sample long document l * as the prior and long document l as the aggregation of short texts. Then for each short text with a sample of l, we sample the topic z of each word in the topic.
Algorithm 1: DESTM Sampling Process
Applsci 11 08708 i001

5. Experimental Results

In this section, we firstly introduce the experimental datasets, the state-of-the-art methods for comparison, evaluation measures, and parameter settings. Then we show the experimental results compared to these methods.

5.1. Experimental Setup

5.1.1. Datasets

We adopt four real-world short texts corpus from different domains. A brief description is as follows:
News: This dataset contains the titles of news collected from RSS feed of three websites (nyt.com, accessed on 5 September 2021, usatoday.com, accessed on 5 September 2021, reuters.com accessed on 5 September 2021). These short texts are categorized into seven classes: Sport, Business, U.S, Health, Sci&Tech, World, and Entertainment.
Tweets: This dataset contains tweets collected from Twitter. Each tweet has a hashtag and tweets with the same hashtag are considered to be in the same category. This corpus contains a total of 100 categories.
Snippets: This dataset contains the results of web search transaction [32]. Short texts are classified into eight categories: Business, Computers, Culture Arts, Education-Science, Engineering, Health, Politics-Society, and Sports.
Captions: This dataset contains captions solicited from Mechanical Turkers for photographs from Flickr and Pascal [33]. Captions are divided into 20 categories.
We pre-process these datasets by discarding stop-words and words that only occur once. The statistic information of four short texts datasets is shown in Table 2.
Here column D means the number of short texts in datasets. Column V means the scale of the vocabulary. Column Len means the average length of short texts. Column C means the number of categories in datasets.

5.1.2. Methods

In this section, we introduce the state-of-the-art methods implemented for comparison.
LDA [2]: This model is one of the conventional topic models. We implement this method as the base method.
DSTM [18]: This method incorporates the Spike and Slab prior that restricts the number of topics a short text corresponds to. Through this strategy, DSTM alleviates the problem of the sparsity of local word co-occurrence information.
SATM [21]: This method is the first self-aggregated topic model that aggregates short texts into latent long documents. This model is a two-phase model. The first phase generates latent long documents, and the second phase generates topics. Latent long documents can bring a lot of additional word co-occurrence information.
PTM [22]: This method is the state-of-the-art self-aggregated topic model. It generates latent long documents along with topics, which is simpler than model SATM. Additionally, aggregating short texts into long documents can overcome the problem of the sparsity of word co-occurrence information.
SPTM [22]: This method incorporates the Spike and Slab prior that restricts the number of topics a long document corresponds to. When the number of long documents is defined as too small, this method can avoid long documents becoming ambiguous.

5.1.3. Evaluation Measures

In this section, we introduce two measures: PMI score and classification accuracy.

PMI Score

We calculate the point-wise mutual information (PMI) [34] score to evaluate the coherence of topics. We use the latest dump of Wikipedia articles as the auxiliary dataset to calculate the PMI score of words. This auxiliary dataset contains 5 million documents and 14 million words of vocabulary. We build a sliding window of 10 words and calculate the PMI score between any two words that appear in the sliding window. The equation of the PMI score is as follows:
P M I ( w i , w j ) = log p ( w i , w j ) p ( w i ) p ( w j )
Here w i and w j are two words that appear in one sliding window. p ( w i , w j ) is the joint probability of the word pair. p ( w i ) and p ( w j ) are marginal probabilities of words. Then, we generate top-N words with the highest probabilities correspond to a topic. Word pairs of top N words are generated and the average PMI score is calculated according to the PMI score from the Wikipedia dataset. Finally, we calculate the average PMI score of all topics.

Classification Accuracy

We calculate the classification accuracy to evaluate the distributions between short texts and topics. Topics of a short text are suggested as features and the probabilities between a short text and its topics are suggested as feature values. Then we use the Support Vector Machine [35] as the classifier and calculate the classification accuracy according to these features of short texts through the bootstrap validation method [36]. The equation of the bootstrap accuracy is as follows:
A C C b o o t = 1 b 1 b ( 0.632 * A C C b + 0.368 * A C C t o t a l )
For a short text set with N short texts, we sample N instances with replacement. Then we generate a SVM classifier from these instances and calculate the accuracy A C C b on the rest of the short texts that are not sampled. A C C t o t a l means the accuracy calculated on the total short texts set according to the classifier. This procedure will be repeated b times and finally, the average accuracy A C C b o o t is calculated.

5.1.4. Parameter Settings

The Gibbs sampling method samples topics through thousands of iterations. For every model, We execute 3000 iterations and ignore first 1000 iterations to skip the convergence phase. For model LDA, we set α = 0.1 and β = 0.01 according to paper [10]. For DSTM, we set π = 0.1 and γ = 0.01 according to paper [22]. Researchers in this paper find these settings outperform settings in paper [18]. For SATM, we set the number of long documents as 300, α = 50 / T and β = 0.1 according to settings in paper [21]. For PTM, we set the number of long documents as 1000, α = 0.1 and β = 0.01 according to paper [22]. For SPTM, we set the number of long documents as 1000, α = 0.1 , β = 0.01 , γ 0 = 0.1 and a ¯ = 10 12 according to paper [22]. The number of long documents is set to 1000 because this number is the appropriate number according to the scale of our four datasets. For our model (DESTM), we set α = 15 , β = 15 , γ = 0.1 , δ = 0.01 , and η = 0.0 .

5.2. Topic Evaluation by Topic Coherence

In this section, we evaluate the topic coherence of the state-of-the-art models by calculating the PMI score. We generate 5, 10, and 15 topics for each model and choose the top 10 words for each topic according to the probability. The results are shown in Figure 2.
The results show that the LDA method achieves very poor performances on short texts. As a conventional method, LDA cannot overcome the problem caused by the sparsity of word co-occurrence information. The DSTM model achieves a little improvement compared to LDA. Although DSTM restricts the number of topics a short text corresponds to, this method cannot provide enough additional word co-occurrence information. So DSTM still suffers from the sparsity of word co-occurrence information. Surprisingly, the SATM model is worse than LDA in most cases. That is because SATM easily causes overfitting. The results of PTM and SPTM are the second or third best performances. The SPTM method can only improve the performances when the number of long documents is defined too short. However, we set this number to be 1000 which is an appropriate number for the scale of our datasets. So the performances of PTM and SPTM are almost the same. The results of these two models show that aggregating short texts into latent long documents can overcome the sparsity of word co-occurrence information. However, non-semantic word co-occurrence information restricts their performances. Our method (DESTM) successfully aggregates semantically related short texts together according to document embeddings and produces sufficient and semantically related additional word co-occurrence information. So the results show that DESTM achieves a significant improvement and outperforms the other state-of-the-art methods.

5.3. Topic Evaluation by Classification Accuracy

In this section, we evaluate the probability relations between short texts and topics. We also generate 5, 10, and 15 topics on 4 datasets. The results are shown in Figure 3.
In these results, the LDA model achieves poor performances as it cannot overcome the sparsity of word co-occurrence information. DSTM performs better than LDA. Restricting the number of topics a short text corresponds to can relieve the sparsity of local word co-occurrence information for each topic. However, without sufficient word co-occurrence information, the DSTM method cannot achieve a great improvement. SATM also performs the worst because of the overfitting problem. PTM and SPTM achieve the second or third best results in most cases. Latent long documents in these two models provide sufficient word co-occurrence information and achieve an improvement on short texts. However, PTM and SPTM only achieve little improvement. In these two models, the probabilities between short texts and topics are generated from the distribution P ( z | d ) . This distribution is generated according to the equation P ( z | d ) = P ( z | l ) P ( l | d ) . z means variables of topics, d means variables of short texts, and l means variables of the long documents. PTM and SPTM cannot avoid aggregating semantically irrelevant short texts together and these short texts will bring noisy information into long documents. So the noisy information in long documents will directly influence the probability between topics and short texts and leads to poor performances on Tweets and Snippets. For our model, the performances are superior compare to the other state-of-the-art models on four datasets. The results show that long documents generated by our model are more reasonable and contain less noisy information.

5.4. Experimental Results for Complete and Partial Document Embeddings

In this section, we discuss the impact of noisy information of document embeddings in our model. We compared several DESTM models with different parameter settings on dataset News. Firstly we use the PTM model to represent the condition without any document embeddings. Settings of PTM follow experiments above. Secondly, we set parameters of DESTM as α = 0 and η = 1 . The model with these settings means aggregating short texts according to the complete document embeddings and of cause including all noisy information. Thirdly, we set α = 15 and η = 1 . The model with these settings means aggregating short texts according to the similarity distribution and distances of short texts without discarding any distances. Then we vary the threshold η to be 0.5 , 0.0 , and 0.5 . We get models incorporating partial document embeddings with different threshold values. Finally, we set α = 15 and η = 1 . The model with these parameters means aggregating short texts only according to the similarity distribution. For other parameters of DESTM, we set β = 2 , γ = 0.1 , and δ = 0.01 . The evaluation results of PMI score and classification accuracy are shown in Figure 4.
In these results, the PTM model achieves the worst PMI score and the second-worst classification accuracy. Without any additional semantic information, the performances of the PTM model are very poor. The DESTM model with α = 0 and η = 1 incorporates the complete document embeddings without discarding noisy information. The model of these settings achieves the third-worst PMI score and the worst classification accuracy. Results show that aggregating short texts directly according to document embeddings will incorporate a lot of noisy information and leads to poor performances. For models with α = 15 and η varies from −1 to 1, the number of distances between short texts that are discarded varies from zero to all distances. The results show that the model with η = 0.0 achieves the best performance. The model with η = 0.5 achieves the second-best performance. The model with η = 1 is the third best and the model with η = 0.5 is the fourth one. Finally, the model with η = 1 achieves the second-worst PMI score and the third-worst classification accuracy. The results show that similarity distributions and distances of short texts are all beneficial to our model and both information should be retained. The threshold of distances should be carefully designed, discarding too many distances or retaining too many distances will all reduce the performances.

5.5. Semantic Explanations of Topic Demonstrations

Topics are constructed by words and should be understandable by human beings. In this section, we analyze the semantic meanings of topics. We generated 10 topics with the top 10 words on the News dataset. Results are shown in Table 3.
We can interpret these topics into semantic meanings. Topic1 is about sports news of basketball. Topic2 is about financial news. Topic3 is about economic news of oil prices. Topic4 is about disaster news in the world. Topic5 is news about international politics. Topic6 is also sports news but corresponds to tennis. Topic7 is political news of the election. Topic8 is about technological news of multimedia service. Topic9 is political news about terrorists. Topic10 is about entertainment news. Results show that we can understand the semantic meanings of topics generated by our model.

5.6. Efficiency Analysis

In this section, we evaluate the efficiency of these models. All topic models are implemented by java with the java 8 virtual machine and are calculated on the same computer. The hardware environment of the experimental computer is a CPU of i5 8400 and an 8G memory. We generate 10 topics on Snippets dataset and the results of initialization time and average time of each iteration are shown in Table 4. The unit of time is milliseconds.
In this result, the conventional model LDA is the fastest because this model is the simplest model. The DSTM model is the most inefficient model because it needs to sample many prior variables that restrict the number of topics. SATM is the second most inefficient model because this model has two phases to calculate. The efficiency of PTM and SPTM is similar. PTM only needs sample long document variables and topics. SPTM is slower than PTM because it needs to sample additional topic selector variables. Our model, DESTM, is faster than DSTM and SATM, but slower compared to PTM and SPTM. That is because we need to sample additional long documents in the first step of the nested Chinese restaurant process. For each iteration of Gibbs sampling, our model will sample three latent variables. For each short text, we sample two latent long document variables, and for each word, we sample the latent topic variable. The time complexity of our model is O ( K * + K + T ) . Here K * is the size of the total long document samples space in the first step of the nested Chinese restaurant process. We calculate K * as K * = i = 1 N K i * . Here N means the number of short texts. K i * means the size of the long document samples space in the first step of short text i. K is the size of the total long document samples space in the second step of the nested Chinese restaurant process. We calculate K as K = i = 1 N K i . K i means the size of the long document samples space in the second step of short text i. T is the size of the total topic samples space. T is calculated as T = N T * . Here T * is the size of the topic samples space of each short text.

6. Conclusions

In this paper, we propose a self-aggregated topic model incorporating document embeddings (DESTM). We calculate document embeddings from short texts and incorporate this information to aggregate semantically related short texts together and to provide sufficient additional semantically related word co-occurrence information. To discard noisy information generated by the sparse word co-occurrence information, document embeddings are separated into global and local semantic information. The global semantic information is the similarity distribution on the entire dataset, and the local semantic information is the distances between similar short texts. Then we use the nested Chinese restaurant process to incorporate these two kinds of semantic information. The global information is incorporated through the first step of the nested Chinese restaurant and the local information is incorporated through the second step. The experimental results show that our model outperforms the other state-of-the-art models in terms of PMI score and classification accuracy. However, our model still has a limitation. When document embeddings that contain noisy information are discarded, we cannot avoid discarding other useful information in these document embeddings like context semantic information. So semantic information incorporated into our model is not sufficient. In the future, we will try to incorporate other kinds of semantic information such as the global context semantic information to provide more sufficient semantic information.

Author Contributions

Conceptualization, Y.N.; methodology, Y.N.; software, Y.N.; validation, Y.N.; formal analysis, Y.N.; investigation, Y.N.; resources, Y.N.; data curation, Y.N.; writing—original draft preparation, Y.N.; writing—review and editing, H.Z.; visualization, Y.N.; supervision, J.L.; project administration, J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Datasets can be obtained from https://github.com/overlook2021/short-texts.git accessed on 5 September 2021.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Blei, D.M. Probabilistic topic models. Commun. ACM 2012, 55, 77–84. [Google Scholar] [CrossRef] [Green Version]
  2. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent dirichlet allocation. J. Mach. Learn. Res. 2003, 3, 993–1022. [Google Scholar]
  3. Teh, Y.W.; Jordan, M.I.; Beal, M.J.; Blei, D.M. Hierarchical Dirichlet Processes. J. Am. Stat. Assoc. 2006, 101, 1566–1581. [Google Scholar] [CrossRef]
  4. Wang, X.; McCallum, A. Topics over time: A non-markov continuous-time model of topical trends. In Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 20–23 August 2006; pp. 424–433. [Google Scholar]
  5. Zhao, W.X.; Jiang, J.; Weng, J.; He, J.; Lim, E.P.; Yan, H.; Li, X. Comparing twitter and traditional media using topic models. In European Conference on Information Retrieval; Springer: Berlin, Germany, 2011; pp. 338–349. [Google Scholar]
  6. Likhitha, S.; Harish, B.; Kumar, H.K. A detailed survey on topic modeling for document and short text data. Int. J. Comput. Appl. 2019, 178, 1–9. [Google Scholar] [CrossRef]
  7. Hong, L.; Davison, B.D. Empirical study of topic modeling in twitter. In Proceedings of the First Workshop on Social Media Analytics, Washington DC, USA, 25–28 July 2010; pp. 80–88. [Google Scholar]
  8. Mehrotra, R.; Sanner, S.; Buntine, W.; Xie, L. Improving LDA Topic Models for Microblogs via Tweet Pooling and Automatic Labeling; Association for Computing Machinery: Stroudsburg, PA, USA, 2013; pp. 889–892. [Google Scholar]
  9. Tang, J.; Zhang, M.; Mei, Q. One theme in all views: Modeling consensus topics in multiple contexts. In Proceedings of the 19th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Chicago, IL, USA, 11–14 August 2013; Association for Computing Machinery: Stroudsburg, PA, USA, 2013; pp. 5–13. [Google Scholar]
  10. Nguyen, D.Q.; Billingsley, R.; Du, L.; Johnson, M. Improving Topic Models with Latent Feature Word Representations. Trans. Assoc. Comput. Linguist. 2015, 3, 299–313. [Google Scholar] [CrossRef]
  11. Das, R.; Zaheer, M.; Dyer, C. Gaussian lda for topic models with word embeddings. In Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Beijing, China, 26–31 July 2015; Association for Computational Linguistics: Stroudsburg, PA, USA, 2015; pp. 795–804. [Google Scholar]
  12. Li, C.; Wang, H.; Zhang, Z.; Sun, A.; Ma, Z. Topic Modeling for Short Texts with Auxiliary Word Embeddings; Association for Computing Machinery: Stroudsburg, PA, USA, 2016; pp. 165–174. [Google Scholar]
  13. Liang, W.; Feng, R.; Liu, X.; Li, Y.; Zhang, X. GLTM: A global and local word embedding-based topic model for short texts. IEEE Access 2018, 6, 43612–43621. [Google Scholar] [CrossRef]
  14. Zuo, Y.; Li, C.; Lin, H.; Wu, J. Topic Modeling of Short Texts: A Pseudo-Document View with Word Embedding Enhancement. IEEE Trans. Knowl. Data Eng. 2021. [Google Scholar] [CrossRef]
  15. Shi, T.; Kang, K.; Choo, J.; Reddy, C.K. Short-text topic modeling via non-negative matrix factorization enriched with local word-context correlations. In Proceedings of the 2018 World Wide Web Conference, Lyon, France, 23–27 April 2018; pp. 1105–1114. [Google Scholar]
  16. Yin, J.; Wang, J. A dirichlet multinomial mixture model-based approach for short text clustering. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 233–242. [Google Scholar]
  17. Li, C.; Duan, Y.; Wang, H.; Zhang, Z.; Sun, A.; Ma, Z. Enhancing topic modeling for short texts with auxiliary word embeddings. ACM Trans. Inf. Syst. (TOIS) 2017, 36, 1–30. [Google Scholar] [CrossRef]
  18. Lin, T.; Tian, W.; Mei, Q.; Cheng, H. The dual-sparse topic model: Mining focused topics and focused terms in short text. In Proceedings of the 23rd International Conference on World Wide Web, Seoul, Korea, 7–11 April 2014; pp. 539–550. [Google Scholar]
  19. Cheng, X.; Yan, X.; Lan, Y.; Guo, J. BTM: Topic Modeling over Short Texts. IEEE Trans. Knowl. Data Eng. 2014, 26, 2928–2941. [Google Scholar] [CrossRef]
  20. Zuo, Y.; Zhao, J.; Xu, K. Word network topic model: A simple but general solution for short and imbalanced texts. Knowl. Inf. Syst. 2016, 48, 379–398. [Google Scholar] [CrossRef] [Green Version]
  21. Quan, X.; Kit, C.; Ge, Y.; Pan, S.J. Short and Sparse Text Topic Modeling via Self-Aggregation. 2015, pp. 2270–2276. Available online: https://www.aaai.org/ocs/index.php/IJCAI/IJCAI15/paper/viewFile/10847/10978 (accessed on 5 September 2021).
  22. Zuo, Y.; Wu, J.; Zhang, H.; Lin, H.; Wang, F.; Xu, K.; Xiong, H. Topic Modeling of Short Texts: A Pseudo-Document View. 2016, pp. 2105–2114. Available online: https://dl.acm.org/doi/10.1145/2939672.2939880 (accessed on 5 September 2021).
  23. Le, Q.; Mikolov, T. Distributed representations of sentences and documents. In Proceedings of the International Conference on Machine Learning, PMLR. 2014, pp. 1188–1196. Available online: http://proceedings.mlr.press/v32/le14.html (accessed on 5 September 2021).
  24. Hu, Y.; John, A.; Wang, F.; Kambhampati, S. Et-lda: Joint topic modeling for aligning events and their twitter feedback. In Proceedings of the AAAI Conference on Artificial Intelligence, Toronto, ON, Canada, 22–26 July 2012; Volume 26. [Google Scholar]
  25. Zhao, W.X.; Wang, J.; He, Y.; Nie, J.Y.; Wen, J.R.; Li, X. Incorporating social role theory into topic models for social media content analysis. IEEE Trans. Knowl. Data Eng. 2014, 27, 1032–1044. [Google Scholar] [CrossRef] [Green Version]
  26. Yang, Y.; Wang, F. Author topic model for co-occurring normal documents and short texts to explore individual user preferences. Inf. Sci. 2021, 570, 185–199. [Google Scholar] [CrossRef]
  27. Mikolov, T.; Sutskever, I.; Chen, K.; Corrado, G.; Dean, J. Distributed representations of words and phrases and their compositionality. arXiv 2013, arXiv:1310.4546. [Google Scholar]
  28. Mikolov, T.; Chen, K.; Corrado, G.; Dean, J. Efficient estimation of word representations in vector space. arXiv 2013, arXiv:1301.3781. [Google Scholar]
  29. Yi, F.; Jiang, B.; Wu, J. Topic modeling for short texts via word embedding and document correlation. IEEE Access 2020, 8, 30692–30705. [Google Scholar] [CrossRef]
  30. Mai, C.; Qiu, X.; Luo, K.; Chen, M.; Zhao, B.; Huang, Y. TSSE-DMM: Topic Modeling for Short Texts Based on Topic Subdivision and Semantic Enhancement; PAKDD (2); Springer: Berlin, Germany, 2021; pp. 640–651. [Google Scholar]
  31. Griffiths, T.L.; Steyvers, M. Finding scientific topics. Proc. Natl. Acad. Sci. USA 2004, 101, 5228–5235. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Phan, X.H.; Nguyen, L.M.; Horiguchi, S. Learning to classify short and sparse text & web with hidden topics from large-scale data collections. In Proceedings of the 17th International Conference on World Wide Web, Beijing China, 21–25 April 2008; pp. 91–100. [Google Scholar]
  33. Finegan-Dollak, C.; Coke, R.; Zhang, R.; Ye, X.; Radev, D. Effects of creativity and cluster tightness on short text clustering performance. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Berlin, Germany, 7–12 August 2016; pp. 654–665. [Google Scholar]
  34. Newman, D.; Lau, J.H.; Grieser, K.; Baldwin, T. Automatic evaluation of topic coherence. In Proceedings of the Human Language Technologies: The 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Los Angeles, CA, USA, 2–4 June 2010; pp. 100–108. [Google Scholar]
  35. Hearst, M.A.; Dumais, S.T.; Osuna, E.; Platt, J.; Scholkopf, B. Support vector machines. IEEE Intell. Syst. Their Appl. 1998, 13, 18–28. [Google Scholar] [CrossRef] [Green Version]
  36. Kohavi, R. A Study of Cross-Validation and Bootstrap for Accuracy Estimation and Model Selection; IJCAI: Montreal, QC, Canada, 1995; Volume 14, pp. 1137–1145. [Google Scholar]
Figure 1. DESTM graphical model.
Figure 1. DESTM graphical model.
Applsci 11 08708 g001
Figure 2. Average PMI score of all models on 4 datasets.
Figure 2. Average PMI score of all models on 4 datasets.
Applsci 11 08708 g002
Figure 3. Classification accuracy of all models on 4 datasets.
Figure 3. Classification accuracy of all models on 4 datasets.
Applsci 11 08708 g003
Figure 4. Comparison results between models with complete and partial document embeddings.
Figure 4. Comparison results between models with complete and partial document embeddings.
Applsci 11 08708 g004
Table 1. Classification of topic models for short texts.
Table 1. Classification of topic models for short texts.
StrategyModels
Models with auxiliary dataIncorporating metadata [7,8,9,24,25,26]
Incorporating a long document set [10,11,12,13,14,29,30]
Models without auxiliary dataLimiting topics correspond to each text [16,17,18]
Using global word co-occurrence information [19,20]
Aggregating short texts automatically [21,22]
Table 2. Statistics of short texts datasets.
Table 2. Statistics of short texts datasets.
DataSetDVLenC
News300010,906127
Tweets250049818100
Snippets30004705148
Captions48343165520
Table 3. Top 10 words for the topic News.
Table 3. Top 10 words for the topic News.
TopicTop-10 Words
Topic1game coach season year scored night team big points boston
Topic2billion group bank financial deal company international chief fund companies
Topic3oil prices year percent high rose growth economic demand fell
Topic4japan nuclear power health study earthquake plant government crisis world
Topic5air officials libyan city nato muammar country people military killed
Topic6open world final french year champion championship week players national
Topic7president court state obama federal minister judge prime year election
Topic8american long online video apple weekend internet music million year
Topic9president al killed security country forces government people bin laden
Topic10show life years news tv star play broadway wedding making
Table 4. Time of initialization and average time of each iteration.
Table 4. Time of initialization and average time of each iteration.
ModelsTime of Initialization (ms)Average Time per Iteration (ms)
LDA1515
DSTM712631
SATM16815
PTM15193
SPTM16237
DESTM310442
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Niu, Y.; Zhang, H.; Li, J. A Nested Chinese Restaurant Topic Model for Short Texts with Document Embeddings. Appl. Sci. 2021, 11, 8708. https://doi.org/10.3390/app11188708

AMA Style

Niu Y, Zhang H, Li J. A Nested Chinese Restaurant Topic Model for Short Texts with Document Embeddings. Applied Sciences. 2021; 11(18):8708. https://doi.org/10.3390/app11188708

Chicago/Turabian Style

Niu, Yue, Hongjie Zhang, and Jing Li. 2021. "A Nested Chinese Restaurant Topic Model for Short Texts with Document Embeddings" Applied Sciences 11, no. 18: 8708. https://doi.org/10.3390/app11188708

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop