Next Article in Journal
Optimizing the Energy Efficiency of Unreliable Memories for Quantized Kalman Filtering
Next Article in Special Issue
An Architecture for Service Integration to Fully Support Novel Personalized Smart Tourism Offerings
Previous Article in Journal
Optimal Design of PV Systems in Electrical Distribution Networks by Minimizing the Annual Equivalent Operative Costs through the Discrete-Continuous Vortex Search Algorithm
Previous Article in Special Issue
Perturbed-Location Mechanism for Increased User-Location Privacy in Proximity Detection and Digital Contact-Tracing Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Investigating the Efficient Use of Word Embedding with Neural-Topic Models for Interpretable Topics from Short Texts

by
Riki Murakami
1 and
Basabi Chakraborty
2,*
1
Graduate School of Software and Information Science, Iwate Prefectural University, 152-52 Sugo, Takizawa 020-0693, Iwate, Japan
2
Faculty of Software and Information Science, Iwate Prefectural University, 152-52 Sugo, Takizawa 020-0693, Iwate, Japan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 852; https://doi.org/10.3390/s22030852
Submission received: 26 December 2021 / Revised: 14 January 2022 / Accepted: 16 January 2022 / Published: 23 January 2022
(This article belongs to the Special Issue Sensor Networks: Physical and Social Sensing in the IoT)

Abstract

:
With the rapid proliferation of social networking sites (SNS), automatic topic extraction from various text messages posted on SNS are becoming an important source of information for understanding current social trends or needs. Latent Dirichlet Allocation (LDA), a probabilistic generative model, is one of the popular topic models in the area of Natural Language Processing (NLP) and has been widely used in information retrieval, topic extraction, and document analysis. Unlike long texts from formal documents, messages on SNS are generally short. Traditional topic models such as LDA or pLSA (probabilistic latent semantic analysis) suffer performance degradation for short-text analysis due to a lack of word co-occurrence information in each short text. To cope with this problem, various techniques are evolving for interpretable topic modeling for short texts, pretrained word embedding with an external corpus combined with topic models is one of them. Due to recent developments of deep neural networks (DNN) and deep generative models, neural-topic models (NTM) are emerging to achieve flexibility and high performance in topic modeling. However, there are very few research works on neural-topic models with pretrained word embedding for generating high-quality topics from short texts. In this work, in addition to pretrained word embedding, a fine-tuning stage with an original corpus is proposed for training neural-topic models in order to generate semantically coherent, corpus-specific topics. An extensive study with eight neural-topic models has been completed to check the effectiveness of additional fine-tuning and pretrained word embedding in generating interpretable topics by simulation experiments with several benchmark datasets. The extracted topics are evaluated by different metrics of topic coherence and topic diversity. We have also studied the performance of the models in classification and clustering tasks. Our study concludes that though auxiliary word embedding with a large external corpus improves the topic coherency of short texts, an additional fine-tuning stage is needed for generating more corpus-specific topics from short-text data.

1. Introduction

Due to the rapid developments of computing and communication technologies and the widespread use of internet, people are gradually becoming accustomed to communicating through various online social platforms, such as microblogs, Twitter, webpages, Facebook, etc. These messages over web and social networking sites contain important information regarding current social situations and trends, people’s opinions on different products and services, advertisements, and announcements of government policies, etc. An efficient text-processing technique is needed to automatically analyze these huge amounts of messages for extracting information. In the area of traditional natural language processing, a topic-modeling algorithm is considered an effective technique for the semantic understanding of text documents. Conventional topic models, such as pLSA [1] or LDA [2] and their various variants, are considerably good at extracting latent semantic structures from a text corpus without prior annotations and are widely used in emerging topic detection, document classification, comment summarizing, or event tracking. In these models, documents are viewed as a mixture of topics, while each topic is viewed as a particular distribution over all the words. Statistical tools are used to determine the latent topic distribution of each document, while higher-order word co-occurrence patterns are used to characterize each topic [3]. The efficient capture of document-level word co-occurrence patterns leads to the success of topic modeling.
The messages posted on various social network sites are generally short compared to the length of relatively formal documents such as newspapers or scientific articles. The main characteristics of these short texts are: (1) a limited number of words in one document, (2) the use of new and informal words, (3) meanings and usages of words that may change greatly depending on the posting, (4) spam posts, and (5) the restricted length of posts, such as API restrictions on Twitter. The direct application of traditional topic models for short-text analysis results in poor performance due to lack of word co-occurrence information in each short text document, originating from the above characteristics of short texts [4]. Earlier research on topic-modeling for short texts with traditional topic models used external, large-scale datasets such as Wikipedia, or related long-text datasets for a better estimation of word co-occurrences across short texts [5,6]. However, these methods work well only when the external dataset closely matches the original short-text data.
To cope with the problems of short-text topic-modeling by traditional topic models, three main categories of algorithms are found in the literature [7]. A simple solution is to aggregate a number of short texts into a long pseudo-document before training a standard topic model to improve word co-occurrence information. In [8], the tweets of an individual user are aggregated in one document. In [9,10], a short text is viewed as sampled from unobserved, long pseudo-documents, and topics are inferred from them. However, the performance of these methods depends on efficient aggregation and data type. When short texts of different semantic contents are aggregated to a long document, non-semantic word co-occurrence information can produce incoherent topics. In the second category, each short text document is assumed to consist of only a single topic. Based on this assumption, Direchlet Multinomial Mixture (DMM) model-based topic-modeling methods have been developed for short texts in [11,12,13]. Although this simple assumption eliminates data-sparsity problems to some extent, they fail to capture multiple topic elements in a document, which makes the model prone to over-fitting. Moreover, “shortness” is subjective and data-dependent; a single-topic assumption might be too strong for some datasets. A poisson-based DMM model (PDMM) [14] considers a small number of topics associated with each short text instead of only one. The third category of algorithms consider global word co-occurrence patterns for inferring latent topics. According to the usage, two types of models are developed. In [15], global word co-occurrence is directly used, while in [16], a word co-occurrence network is first constructed using global word co-occurrence, and then latent topics are inferred from this network. In the present work, we explored methods of exploiting this category for further improvement in the development of algorithms for extracting interpretable topics from short texts.
Another limitation of the above models for short-text analysis is that the context or background information is not used, resulting in the generation of not-so-coherent topics. The statistical information of words in the text document cannot fully capture words that are semantically correlated but that rarely co-occur. Recent advances in word embedding [17] provides an effective way of learning semantic word relations from a large corpus, which can help to develop models for generating more interpretable and coherent topics. Word embedding uses one-hot representation of words with vocabulary length vectors of zeroes with a single one, and words that are similar in semantics are close in a lower-dimensional vector space. An embedded topic model (ETM), a combination of LDA and word embedding which enjoys both the advantages of topic model and word embedding, has been proposed in [18]. Traditional topic models with word embedding for documents are explored in several other research works, cited in [19]. In [20], word embedding is combined with LDA to accelerate the inference process, resulting in the enhanced interpretability of topics. For short texts, models incorporating word embedding into DMM are proposed in [21,22]. In [23,24], short texts are merged into long pseudo-documents using word embedding. Word embedding in conjunction with conventional topic models seems to be a better technique for generating coherent topics.
The increasing complexity of inference processes in conventional topic models on large-text data, along with the recent developments of deep neural networks, has led to the emergence of neural-topic models (NTM). These models combine the performance, efficacy, scalability, and ease of leveraging parallel computing facilities, such as GPU, to probabilistic topic-modeling [25]. Neural-topic models are considered to be computationally simpler and easier for implementation, compared to traditional LDA models, and are increasingly used in various natural language processing tasks in which conventional topic models are difficult to use. A systematic study on the performances of several neural-topic models has been reported in [26]. Although various neural-topic models have been proposed, and although the reported experimental results on topic generation seem to be better than conventional topic models for long and formal texts, little research has been conducted on neural-topic models for effective analysis of short texts [27]. Most of the research works of topic modeling on short texts are based on extensions of Bayesian probabilistic topic models (BPTM) such as LDA.
The objective of the present research is to explore computationally easy and efficient techniques for improving the interpretability of generated topics from real-world short texts using neural-topic models. However, learning context information is the most challenging issue of topic-modeling for short texts, and incorporating pretrained word embedding into a topic model seems to be one of the most efficient ways of explicitly enriching the content information. Neural-topic models with pretrained word embedding for short-text analysis has not been extensively explored yet, compared to its long-text counterparts. In [28], we presented our preliminary analysis of short-text data (benchmark and real-world) with neural-topic models using pretrained word embedding. We found that although pretrained word embedding enhances the topic coherence of short texts that are similar to long and formal texts, the generated topics were often comprised of words having common meanings (which are found in the large external corpus used for pretraining) instead of the particular short-text-specific semantics of the word, which is especially important for real-world datasets. In other words, the learning of topic centroid vectors is influenced by pretraining the text corpus and fails to discover the important words of the particular short text. Our proposal is that this gap can be filled by adding a fine-tuning stage to the training of the topic model with the particular short-text corpus to be analysed. In this work, we have completed an extensive study to investigate the performance of recent neural-topic models with and without word embedding, and also with the proposed fine-tuning stage, for generating interpretable topics from short texts in terms of a number of performance metrics by simulation experiments on several datasets. We have also studied the performance of the NTM with pretrained word embedding added, with a fine-tuning stage for classification and clustering tasks. As a result of our experiments, we can confirm that the addition of a fine-tuning stage indeed enhances the topic quality of short texts in general, and generates topics with corpus-specific semantics.
In summary, our contributions in this paper are as follows:
  • A proposal for fine-tuning with the original short-text corpus, along with the pretrained word embedding with the large external corpus, for generating more interpretable and coherent corpus-specific topics from short texts;
  • An extensive evaluation of the performance of several neural-topic models, with and without pretrained word embedding and with an added fine-tuning stage, in terms of topic quality and measured by several metrics of topic coherence and topic diversity;
  • A performance evaluation of the proposed fine-tuned neural-topic models for classification and clustering tasks.
In the next section, neural-topic models are introduced in brief, followed by a short description of related works on neural-topic models (NTM), especially NTMs for short texts. The following sections contain our proposal, followed by simulation experiments and results. The final section presents the conclusion.

2. Neural-Topic Models and Related Works

The most popular neural-topic models (NTMs) are based on a variational autoencoder (VAE) [29], a deep generative model, and amortised variational inferences (AVI) [30]. The basic framework of VAE-based NTMs is described in the next section, in which generative and inference processes are modeled by a neural network-based decoder and encoder, respectively. Compared to the traditional Bayesian probabilistic topic models (BPTM), inference in neural-topic models is computationally simpler, their implementation is easier due to many existing deep learning frameworks, and NTMs are easy to be integrated with pretrained word embeddings for prior-knowledge acquisition. Several categories of VAE-based NTMs have been proposed. To name a few, there is the Neural Variational Document Model (NVDM) [31], Neural Variational Latent Dirichlet Allocation (NVLDA) [32], the Dirichlet Variational Autoencoder topic model (DVAE) [33], the Dirichlet Variational Autoencoder (DirVAE) [34], the Gaussian Softmax Model (GSM) [35], and iTM-VAE [36]. This list is not exhaustive and is still growing.
In addition to VAE-based NTMs, there are a few other frameworks for NTMs. In [37], an autoregressive NTM, named DocNade, has been proposed. Consequently, some extensions of DocNADE are found in the literature. Recently, some attempts have been made to use a GAN (Generative Adversarial Network) framework for topic-modeling [38,39]. Instead of considering a document as a sequence or a bag of words, a graph representation of a corpus of documents can be considered. In [40], a bipartite graph, with documents and words as two separate partitions and connected by word occurrences in documents as the weights, is used. Ref. [41] uses the framework of Wasserstein auto-encoders (WAEs), which minimizes the Wasserstein distance between reconstructed documents from the decoder and the real documents, similar to a VAE-based NTM. In [42], a NTM based on optimal transport that directly minimizes the optimal transport distance between the topic distribution learned by an encoder and the word distribution of a document has been introduced.

Neural-Topic Models for Short-Text Analysis

In order to generate coherent, meaningful, and interpretable topics from short texts by incorporating semantic and contextual information, a few researchers used NTMs in lieu of conventional topic models. In [43,44], a combination of NTM and either a recurrent neural network (RNN) or a memory network has been used, in which topics learned by the NTM are utilized for classification by a RNN or a memory network. In both works, the NTM shows better performance than conventional topic models in terms of topic coherence and a classification task. To enhance the discreteness of multiple topic distributions in a short text, in [27], the authors used Archimedean copulas. In [45], the authors introduced a new NTM with a topic-distribution quantization approach, producing peakier distributions, and also proposed a negative sampling decode, learning to minimize repetitive topics. As a result, the proposed model outperforms conventional topic models. In [46], the authors aggregated short texts into long documents and incorporated document embedding to provide word co-occurrence information. In [47], a variational autoencoder topic model (VAETM) and a supervised version (SVAETM) of it have been proposed by combining embedded representations of words and entities by employing an external corpus. To enhance contextual information, the authors in [48] proposed a graph neural network as the encoder of NTM, which accepts a bi-term graph of the words as inputs and produces the topic distribution of the corpus as the output. Ref. [49] proposed a context-reinforced neural-topic model with the assumption of a few salient topics for each short text, informing the word distributions of the topics using pretrained word embedding.

3. Proposal for Fine-Tuning of Neural-Topic Models for Short-Text Analysis

From the analysis of present research works on neural-topic models on short-text analysis, it seems that incorporating auxiliary information from an external corpus is one of the most popular and effective techniques for dealing with sparsity in short texts. As mentioned in the introduction, in our previous work [28], we found that although pretrained word embedding with a large external corpus helps with generating coherent topics from a short-text corpus, the generated topics lack the semantics expressed by the corpus-specific meaning of words. If the domain of the short-text corpus and the external corpus vary too much, the topic semantics become poor. This fact is also noted by other researchers [25].
In this work, we propose an additional fine-tuning stage, using the original short-text corpus, along with the pretrained word embedding and a large external corpus. For pretrained word embedding, we decided to use GloVe [50] after some preliminary experiments with two other techniques, namely, Word2Vec and Fast Text, as GloVe provided consistent results. Here, we have completed an extensive comparative study to evaluate the effect of pretrained embedding with our proposed additional fine-tuning stage using several short-text corpora and neural-topic models. The proposed study setting is represented in Figure 1.
Here, pretrained word embedding is denoted as PWE. We have performed three sets of experiments for topic extraction, using only neural-topic models (NTM), neural-topic models with pretrained word embedding (NTM-PWE), and neural-topic models with pretrained embedding and the proposed fine-tuning step (NTM-PWE/fine-tuning). In all the cases, the data corpus is first pre-processed, and in NTM-PWE, word embedding vectors are replaced by PWE after the model parameters are initialized; the weights of PWE are not updated during the training step. In our proposed PWE/fine-tuning or simple fine-tuning (as mentioned in the text), the weights are gradually updated in the training step after replacing the word embedding vectors, as in PWE. In this case, it is possible to update the parameters at the same learning rate that is set to update the entire model, but experiments have shown that updating the PWE values at a large learning rate can easily over-fit the training data. Therefore, in the simulation experiments, we have set the learning rate of the word embedding vectors to a smaller value than the learning rate of the entire model.
We have used popular VAE-based neural-topic models with a few similar WAE (Wasserstein autoencoder)-based models, and ten popular benchmark datasets, for our simulation experiments. The performance of each neural-topic model with no word embedding, pretrained word embedding, and additional fine-tuning has been evaluated by the generated topic quality using different evaluation metrics of topic coherence and topic diversity. The neural-topic models, datasets, and evaluation metrics used in this study are described below.

3.1. Neural-Topic Models for Evaluation

In this section, the neural-topic models used in this study are described briefly. Table 1 describes the meaning of the notations used for description of the models.
Figure 2 and Figure 3 describe the generalized architecture of the Variational autoencoder (VAE)- and Wasserstein autoencoder (WAE)-based neural-topic models, respectively. In both the models, the part of the network that generates θ is known as the encoder, which maps the input bag-of-words (BoW) to a latent document–topic vector, and the part that receives θ and outputs p ( x ) is called the decoder, which maps the document–topic vector to a discrete distribution over the words in the vocabulary. They are called autoencoders because the decoder aims to reconstruct the word distribution of the input. In VAE, h is sampled by Gaussian distribution, and θ is created by performing some transformation on it. WAE, on the other hand, uses the Softmax function directly to create θ , so no sampling is required. Evidence lower bound (ELBO), the objective function of VAE, is defined below [29]:
L d = E q ( θ d ) n = 1 N d log z n p w n β z n p z n θ D K L q ( θ x d ) p θ μ 0 , σ 0 2
It is empirically known that maximizing this ELBO alone will result in smaller (worse) topic diversity. In order to solve this problem, some NTMs use a regularization term to increase the topic diversity [51]:
a ( α i , α j ) = arccos α i · α j α i · α j ζ = 1 K 2 i K j K a α i , α j ν = 1 K 2 i K j K a α i , α j ζ 2 J = L + λ ( ζ ν )
where λ is a hyper-parameter that manipulates the influence of the regularization term; 10 was adopted here. This value was determined empirically. The VAE-based models in this paper use this regularization term.
The particular NTMs used in our study are mentioned in the next subsections.

3.1.1. Neural Variational Document Model (NVDM)

NVDM [31] is, to our knowledge, the first VAE-based document model proposed with the encoder implemented by a multilayer perceptron. This model uses the sample h from the Gaussian distribution as an input for the decoder, and variational inference is based on minimizing KL divergence. While most of the NTMs proposed after this one transform h to treat θ as a topic proportion vector, NVDM is a general VAE.

3.1.2. Neural Variational Latent Dirichlet Allocation (NVLDA)

NVLDA [32], another variant of NVDM, is a model that uses Neural Variational Inference to reproduce LDA. Here, the Softmax function is used to convert z to θ . The probability distribution that maps samples from a Gaussian distribution to the Softmax basis is called the Logisitic–Normal distribution, which is used as a surrogate for the Dirichlet distribution. Additionally, the decoder is p ( x ) = s o f t m a x ( β ) · θ . Unlike the NVDM, where both the topic proportions and the topic–word distribution are in the form of probability distributions, this model is a topic model. Logisitic–Normal distribution is defined as follows:
h N o r m a l ( μ , σ 2 )
θ = s o f t m a x ( h )

3.1.3. Product-of-Experts Latent Dirichlet Allocation (ProdLDA)

ProdLDA [32] is an extension of NVLDA in which the decoder is designed by following the product of the expert model, and the topics-word distribution are not normalized.

3.1.4. Gaussian Softmax Model (GSM)

GSM [35] converts h to θ using Gaussian Softmax, as defined below:
h N o r m a l ( μ , σ 2 )
θ = softmax ( W 1 T h )
where W 1 R K × K , a linear transformation, is the trainable parameters used as the connection weights.

3.1.5. Gaussian Stick-Breaking Model (GSB)

GSB [35] converts h to θ by Gaussian Stick-Breaking construction, which is defined as follows:
h N o r m a l ( μ , σ 2 )
η = s i g m o i d ( W 2 T h )
θ = f S B ( η )
where W 2 R K × K 1 is the trainable parameters used as connection weights, and the stick-breaking function f S B is described by Algorithm 1:
Algorithm 1 Stick-Breaking Process ( f S B )
Input: 
Return value from sigmoid function η
η R + K , where η k [ 0 , 1 ]
Output: 
Topic proportion vector θ
θ R + K , where k ( θ k ) = 1
1:
Assign η 1 to the first element of the topic proportion vector θ 1 .
θ 1 = η 1
2:
for k = 2 , K 1 do
θ k = η k i = 1 k 1 ( 1 η i )
3:
end for
4:
θ K = i = 1 K 1 ( 1 η i )

3.1.6. Recurrent Stick-Breaking Model (RSB)

RSB [35] converts h to θ by recurrent Stick-Breaking construction, as defined below. Here, the stick-breaking construction is considered as a sequential draw from a recurrent neural network (RNN).
h N o r m a l ( μ , σ 2 )
η = f RNN ( h )
θ = f SB ( η )
where f RNN is decomposed as:
h k = RNN SB ( h k 1 )
η k = sigmoid ( h k 1 T z )
and f S B ( η ) is the same as in GSB.

3.1.7. Wasserstein Latent Dirichlet Allocation (WLDA)

WLDA [41] is a topic model based on a Wasserstein autoencoder (WAE). Though various probability distributions can be used for the prior distribution of θ , in this paper, we use the Dirichlet distribution, which we believe is the most basic. In WAE, two training methods are available, GAN (Generative Adversarial Network)-based training and MMD (Maximum Mean Discrepancy)-based training, but in WLDA, MMD is used because of the ease of convergence of training loss. In VAE, the loss function is composed of the KL Divergence used as the regularization term for θ and the reconstruction error, while in WLDA, MMD is used as the regularization term.
If P Θ is a θ ’s prior distribution, and Q Θ is a fake samples‘s prior distribution, maximum mean discrepancy (MMD) is defined as:
MMD k Q Θ , P Θ = Θ k ( θ , · ) d P Θ ( θ ) Θ k ( θ , · ) d Q Θ ( θ ) H k
where H means the reproducing kernel Hilbert space (RKHS) of real-valued functions mapping Θ to R , and k is the kernel function.
k θ , θ = exp arccos 2 k = 1 K θ k θ k

3.1.8. Neural Sinkhorn Topic Model (NSTM)

NSTM [52] is trained using optimal transport [42], as in WLDA. Since we assume that θ encodes x into a low-dimensional latent space while preserving sufficient information about x, the optimal transport distance between θ and x is calculated by the Sinkhorn Algorithm. The sum of this optimal transport distance and the negative log likelihood is used as the loss function.

3.2. Datasets

Table 2 presents the details of the benchmark datasets used in this work. The first column represents the name of the dataset, followed by the number of documents ( | D | ), vocabulary size ( | V | ), the total number of tokens ( X ), average document length (ave dL), maximum document length (max dL), sparsity, number of classes (C,) and the source of the data in the respective columns. In the source column, 1 and 2 represent OCCITS and STTM, respectively.
1. OCTIS: https://aclanthology.org/2021.eacl-demos.31/ (accessed on 14 January 2022).
2. STTM: https://arxiv.org/pdf/1701.00185.pdf (accessed on 14 January 2022).
The first two datasets fall in the category of long documents, and the other eight datasets can be considered as the short-text corpus, as the average document length is quite short compared to the long documents.
The datasets shown in the table are pre-processed. HTML tags and other symbols have been removed from each dataset, and all words have been lowercased. Then, the stop-words were removed and lemmatized. From these datasets, 80% of the total documents was used as the training data and the rest as the test data. These pre-processed corpora are then converted into a BoW (Bag-of-Words), which basically has word frequency as an element, to be used as input data for the NTM. However, for the NSTM, the vector corresponding to each document in the BoW is divided by the total value of the vectors, as in the original paper.

3.3. Evaluation of Topic Quality

It is quite challenging to evaluate the performance of topic models, including NTMs, according to the quality of the generated topics. Topics generated by topic models can be considered as soft clusters of words. Under the constraints of the topic model, this is a probability distribution that collects the probability of word generation for each topic; the same is true for NTM, but this may not be in the form of a probability distribution for document models that impose even weaker constraints than the topic model. Either way, a topic here is a topic–word distribution, and each distribution has as many dimensions as the number of lexemes that occur in the corpus. It is very difficult to understand the goodness of a topic by directly comparing them with human topics. Therefore, in practice, analysts check a list of N words characteristic of a topic based on the values of the word distributions. In most cases, the list of the top-N words in terms of the large probability values in the word distribution is used.
Various metrics have been proposed to evaluate the quality of the top-N words with two main directions. One is to check whether the meaning of words belonging to the top-N words are consistent with each other, defined as topic coherence (TC). The other is to measure the diversity of the top-N words of each pair of topics, defined as topic diversity (TD) or topic uniqueness. Topics with high TC may have low TD. In this case, the top-N words of most topics will be nearly the same, which is not desirable. So, to evaluate the quality of a topic for human-like interpretability, it should have high TC as well as high TD.

3.3.1. Topic Coherence (TC)

For computing TC, general coherence between two sets of words are estimated based on word co-occurrence counts in a reference corpus [53]. The choices are (1) the training corpus for topic modeling; (2) a large external corpus (e.g., Wikipedia); (3) word embedding vectors trained on a large external corpus (e.g., Wikipedia). The scores may differ according to different computations. Choice 1 is easy, but the results are affected by the size of the training corpus. Choices 2 and 3 are more popular, although choice 2 is computationally costly. However, if the domain gap of the training corpus and the external corpus is high, the evaluation is not proper. In this work, we have used the following metrics for computation of topic coherence:
  • Normalized Point-Wise Mutual Information (NPMI) [54]: NPMI is a measure of the semantic coherence of a group of words. It is considered to have the largest correlations with human ratings, and is defined by the following equation:
    N P M I ( w ) = 1 N ( N 1 ) j = 2 N i = 1 j 1 l o g P ( w i , w j ) P ( w i ) P ( w j ) l o g P ( w i , w j )
    where w is the list of the top-N words for a topic. N is usually set to 10. For K topics, averages of NPMI over all topics are used for evaluation;
  • Word Embeddings Topic Coherence (WETC) [55]: WETC represents word embedding-based topic coherence, and pair-wise WETC for a particular topic is defined as:
    WETC P W ( E ( k ) ) = 1 N ( N 1 ) j = 2 N i = 1 j 1 E i , : ( k ) , E j , : ( k )
    where . , . denotes the inner product. For the calculation of the WETC score, pretrained weights of GloVe [50] have been used, and E ( k ) is the word embedding vector sequence of GloVe corresponding to the top-N words for topic k; E i ( k ) means and all vectors are normalized as follows: | | E i , : ( k ) | | = 1 , N is taken as 10.
    WETC c (centroid WETC) is defined as follows:
    WETC c ( E ( k ) ) = 1 N n = 1 N E ( k ) t t = α : , k | | α : , k | |

3.3.2. Topic Diversity

Topic diversity is defined here as the percentage of unique words in the top 25 words of all topics, according to [18]. A diversity close to 0 represents a redundant topic, and those close to 1 indicate more varied topics. Here, we have also used two other metrics, inverted rank-biased overlap (InvertedRBO) [56] and mean squared cosine deviation among topics (MSCD) [57], as a measure of diversity of the generated topics. InvertedRBO is a measure of disjointedness between topics weighted on word rankings, based on the top-N words. The higher these metrics are, the better. MSCD is the cosine similarity of the word distribution of each topic, so it should be lower for better topics. In general, NTM training updates parameters to maximize ELBO, but such a naive implementation can easily lead to poor TD. However, in this case, since we use the topic centroid vectors as trainable parameters, we regularize parameters of NTM to increase the angle formed by each topic centroid vector in order to increase the TD.

4. Simulation Experiments and Results

The simulation experiments have been performed with several benchmark datasets, and the performance of the topic models are evaluated by topic coherence and topic diversity measures.

4.1. Experimental Configuration

For the purpose of comparison and evaluation, the experimental setting should be similar for all the neural-topic models and all the datasets. At the beginning, we completed some trial experiments, and determined that the optimum topic size parameter should be set at K = 50 , based on topic coherence and perplexity, so that there are a sufficient number of topics without becoming very large, considering the length of short text. This value is also in accordance with the value used for related experiments in similar research works. The number of dimensions of the word embeddings was fixed at L = 300 . This is in accordance with the GloVe’s Common Crawl-based trained word embedding vectors, publicly available at https://nlp.stanford.edu/projects/glove/ (accessed on 14 January 2022), which cover largest number of vocabularies.
The other experimental parameters are set as follows: number of units of the encoder’s hidden layers: H ( 1 ) = 500 , H ( 2 ) = 500 ; Dropout rate: p dropout = 0.2 ; Minibatch size: 256; Max epochs: 200; Learning rate for the encoder network: 0.005; Learning rate for the decoder network: 0.001. We employ Adam as the optimizer and Softplus as the activation function of the encoder networks.
On WLDA, we employ Dirichlet as the prior distribution for topic proportion-generation, using MMD for this model’s training. On NSTM, Sinkhorn algorithm’s max number of updates is 2000, and the threshold value for updating termination condition is 0.05, constant value α Sinkhorn = 20 .

4.2. Results for Topic Coherence

Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11 and Table 12 represent the detail results of different topic coherence metrics (NPMI and WETC) for different neural models and different datasets, respectively. Values in bold faces indicate the best results. We have used two versions of GloVe, differing in terms of the size of the corpus.
For many datasets, NVLDA-PWE/fine-tuning has the highest TC. One of the challenges of using PWE without fine-tuning is that the high domain gap between the PWE training corpus and the corpus for topic modeling has a negative impact. In many cases, our proposal produces better results, but not for all the datasets or for all the models. The dataset “GoogleNews” often has the best TC with PWE, and does not show better performance with additional fine-tuning. This is probably because this corpus has a similar domain as the training data for PWE. In a few datasets, the best performance is noticed when no pretraining word embedding is used. It is verified that for those datasets, the original corpus contains sufficient word co-occurrence information.
However, we noted that the TC value changes significantly depending on the type of word embedding. This result suggests that the quality of the word embeddings may have a significant impact on the training of the topic model. In particular, whether or not the unique words in the training corpus are included in the unique words in the PWE has a significant impact. If the coverage of this word dictionary is large, the PWE can be used for evaluation, but if there are many missing words, the reliability of the evaluation value will be greatly compromised.
Figure 4 presents the summary of topic coherence over all the neural-topic models for the long-text corpus (2 datasets) and the short-text corpus (8 datasets), which shows the overall trend. In the case of long texts, the scores of the PWE/fine-tuning metrics for TC are either a little worse or the same as the others. One of the reasons for this is that the long-text corpus used in this study is composed of relatively formal documents, which is close to the domain of PWE. In contrast, the short-text corpus shows better performance in all metrics. The overall trend is none < PWE < PWE/finetuning.

4.3. Results for Topic Diversity

Table 13, Table 14, Table 15, Table 16, Table 17, Table 18, Table 19, Table 20, Table 21 and Table 22 represent the detailed results of different metrics (TopicDiversity, Inverted RBO, and MSCD) expressing a diversity of topics for different neural-topic models and different datasets, respectively. Values in bold indicate the best results. InvertedRBO focuses on the weight of the top-N words. It shows the highest values in almost all of the cases, from which it can be inferred that we were able to construct the topics with high diversity. This result shows that it was useful to add a regularization term that maximizes the distance between topic-centroid vectors, resulting in highly diverse topics.
Furthermore, WLDA and NSTM show similar results without this regularization term, indicating that these models are able to learn without compromising topic diversity in their raw form. To check if this regularization term is working well, we have added TopicCentroidDistance (TCD) in the tables. The larger this metric is, the better, but the values are almost the same for all cases. This metric was evaluated based on two PWEs, and since the values varied, we can infer that the quality of the embedding has a significant impact on the evaluation of the topic model.
Although the results of TopicDiversity varied greatly depending on model and dataset, when checked individually, the scores were sufficiently better in many cases. However, as in the case of Biomedical’s NVLDA-pwe/fine-tuning results, there were cases where the TC showed good scores but the TD showed bad scores. In this respect, InvertedRBO also shows a good score, but MSCD, which is an evaluation using the entire topic–word distribution, shows a relatively large value (i.e., a bad score), indicating that the topics are relatively tangled. Metrics such as TopicDiversity and InvertedRBO, which are based on the top-N words, are useful for evaluating topic diversity, but it is also important to evaluate the entire topic–word distribution.
Figure 5 presents the summary of topic diversity results over all the neural-topic models for the long-text corpus (2 datasets) and the short-text corpus (8 datasets), which shows the overall trend. Among the metrics related to TD, the InvertedRBO score is almost the highest in all cases. This indicates that there is sufficient diversity in all conditions. However, for the other scores, the performance is slightly worse for PWE and PWE/fine-tuning.

4.4. Classification and Clustering Performance

Table 23, Table 24, Table 25, Table 26, Table 27, Table 28, Table 29, Table 30, Table 31 and Table 32 represent the classification and clustering performance of all models and all datasets, respectively. Values in the bold face represent best results. For the TrecTweet dataset, the classification results could not be obtained, possibly due to some technical problem. Figure 6 presents the average classification and clustering performance of the models over long- and short-text datasets. Classification has been performed by a SVM (Support Vector Machine) with linear and rbf kernels. Classification accuracy, precision, recall, and F1 scores have been used for performance assessment and for supervised classification, and NMI (Normalized Mutual Information) and Purity have been used for unsupervised classification.
For classification, VAE-based models, such as NVDM and GSM, exhibit good performance, while WAE-based models, such as WLDA and NSTM, show relatively poor performance. NSTM shows good performance in TC and TD, especially in TD, without adding any regularization term. However, the application to downstream tasks using WAE variants remains a challenge. Considering the overall trend, for long texts, PWE with fine-tuning improves all the scores, but for short texts, the performance is the best for the cases without embedding. Although, after fine-tuning, the scores got better than those obtained with pretrained embedding only.
For clustering results, the large NMI and Purity scores for all models and all datasets for both long and short texts indicate that there is a concentration of documents with the same label around the topic-centroid vector, which proves that the proposal of PWE/fine-tuning improves topic cohesion. Therefore, we can see that our proposal of PWE/fine-tuning contributes to narrowing the domain gap between the training corpus and PWE.

5. Conclusions

Short-text data are now becoming ubiquitous in the real world through various social networking sites. The importance of analysing these short messages is also growing day by day. Unlike long texts or documents, short texts suffer from a lack of word co-occurrence information due to their restricted lengths, posing a difficulty in generating coherent and interpretable topics with popular topic-model techniques.
The use of pretrained word embedding in neural-topic models is a good choice to easily increase the generated topic quality as measured by topic coherence and topic diversity. This is effective for both long and short texts, and reduces the number of trainable parameters, thus shortening the training step time. However, to achieve better topic coherence, especially in short texts, or to make the top-N words of a topic more relevant to the real semantic contents of the training corpus, the additional fine-tuning stage proposed in this work is indeed necessary. The extensive study in this work with several neural-topic models and benchmark datasets justifies our proposal.
However, the use of pretrained word embedding (PWE) has its inherent limitations, which may affect the quality of the extracted topics from short texts. The short-text corpus to be analyzed may contain words that are not included in the vocabulary covered by the corpus used for pretrained word embedding. In this case, NTM-PWE uses a vector initialized with zero. As the vocabulary coverage increases, the performance is likely to deteriorate. Moreover, in the case of NTM-PWE/fine-tuning, there is a possibility that the number of parameter updates will increase until the loss function converges, resulting in an increase in training time. If the time difference between the corpus used for PWE training and the corpus to be analyzed is too large, the meanings of words may change with time, which may have a negative impact on the production of interpretable topics.
It is also seen that the improvement in topic quality after introducing a fine-tuning stage is not the same for all the datasets and all the models. It is difficult to define the correlation between the structure of neural-topic models and the inherent characteristics of the datasets, which poses a challenge to our study. In this work, we limited our study to benchmark datasets available on the internet. Currently, we are collecting data for the evaluation of our proposal with real-world datasets.
By incorporating the additional training with the original training corpus, along with pretrained word embedding with the external corpus, we can improve the purity and NMI of the topics evaluated using the class labels of the documents. Thus, we can construct topics that are more suitable for the training corpus. This method can also be expected to improve the performance of downstream tasks, such as classifications for long texts. Even for short texts, the performance of the downstream tasks is better than when using pretrained word embedding without fine-tuning.

Author Contributions

Conceptualization, R.M.; methodology, R.M.; software, R.M.; validation, R.M. and B.C.; investigation, R.M.; data curation, R.M.; writing—original draft preparation, R.M.; writing—review and editing, B.C.; visualization, R.M. and B.C.; supervision, B.C.; project administration, B.C.; funding acquisition, B.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The authors acknowledge the technical support of the Pattern Recognition and Machine Learning Laboratory, Department of Software, Iwate Prefectural University, Japan.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hofmann, T. Probabilistic latent semantic indexing. In Proceedings of the Twenty-Second Annual International SIGIR Conference, Berkeley, CA, USA, 15–19 August 1999; pp. 50–57. [Google Scholar]
  2. Blei, D.M.; Ng, A.Y.; Jordan, M.I. Latent Dirichlet Allocation. J. Mach. Learn. Res. JMLR 2003, 3, 993–1022. [Google Scholar]
  3. Blei, D.M. Probabilistic topic models. Commun. ACM 2012, 55, 77–84. [Google Scholar] [CrossRef] [Green Version]
  4. Hong, L.; Davison, B. Empirical study of topic modeling in Twitter. In Proceedings of the First Workshop on Social Media Analytics, Washington, DC, USA, 25–28 July 2020; ACM: New York, NY, USA, 2010; pp. 80–88. [Google Scholar]
  5. Phan, X.; Nguyen, L.; Horiguchi, S. Learning to classify short and sparse text & web with hidden topics from large scale data collections. In Proceedings of the 17th International Conference on World Wide Web, Beijing, China, 21–25 April 2008; pp. 91–100. [Google Scholar]
  6. Jin, O.; Liu, N.; Zhao, K.; Yu, Y.; Yang, Q. Transferring topical knowledge from auxiliary long texts for short text clustering. In Proceedings of the 20th International Conference on Information and Knowledge Management, Scotland, UK, 24–28 October 2011; pp. 775–784. [Google Scholar]
  7. Qiang, J.; Qian, Z.; Li, Y.; Yuan, Y.; Wu, X. Short Text Topic Modeling Techniques, Applications, and Performance: A Survey. IEEE Trans. Knowl. Data Eng. 2020. early access. [Google Scholar] [CrossRef]
  8. Weng, J.; Lim, E.; Jiang, J.; He, Q. Twitterrank: Finding topic-sensitive influential twitterers. In Proceedings of the Third ACM International Conference on Web Search and Data Mining WSDM, New York, NY, USA, 3–6 February 2010; pp. 261–270. [Google Scholar]
  9. Quan, X.; Kit, C.; Ge, Y.; Pan, S.J. Short and sparse text topic modeling via self-aggregation. In Proceedings of the 24th International Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 2270–2276. [Google Scholar]
  10. Zuo, Y.; Wu, J.; Zhang, H.; Lin, H.; Wang, F.; Xu, K.; Xiong, H. Topic modeling of short texts: A pseudo-document view. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 2105–2114. [Google Scholar]
  11. Zhao, W.X.; Jiang, J.; Weng, J.; He, J.; Lim, E.P.; Yan, H.; Li, X. Comparing twitter and traditional media using topic models. In Advances in Information Retrieval; Springer: Berlin/Heidelberg, Germany, 2011; pp. 338–349. [Google Scholar]
  12. Yin, J.; Wang, J. A dirichlet multinomial mixture model-based approach for short text clustering. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24–27 August 2014; pp. 233–242. [Google Scholar]
  13. Nigam, K.; McCallum, A.K.; Thrun, S.; Mitchell, T. Text classification from labeled and unlabeled documents using EM. Mach. Learn. 2000, 39, 103–134. [Google Scholar] [CrossRef] [Green Version]
  14. Li, C.; Duan, Y.; Wang, H.; Zhang, Z.; Sun, A.; Ma, Z. Enhancing topic modeling for short texts with auxiliary word embeddings. ACM Trans. Inf. Syst. (TOIS) 2017, 36, 11. [Google Scholar] [CrossRef]
  15. Cheng, X.; Yan, X.; Lan, Y.; Guo, J. BTM: Topic modeling over short texts. IEEE Trans. Knowl. Data Eng. 2014, 26, 2928–2941. [Google Scholar] [CrossRef]
  16. Zuo, Y.; Zhao, J.; Xu, K. Word network topic model: A simple but general solution for short and imbalanced texts. Knowl. Inf. Syst. 2016, 48, 379–398. [Google Scholar] [CrossRef] [Green Version]
  17. Almeida, F.; Xexeo, G. Word Embeddings: A Survey. arXiv 2019, arXiv:1901.09069. [Google Scholar]
  18. Dieng, A.B.; Ruiz, F.J.R.; Blei, D.M. Topic Modeling in Embedding Spaces. Trans. Assoc. Comput. Linguist. 2020, 8, 439–453. [Google Scholar] [CrossRef]
  19. Chauhan, U.; Shah, A. Topic Modeling Using Latent Dirichlet allocation: A Survey. ACM Comput. Surv. 2021, 54, 145. [Google Scholar] [CrossRef]
  20. Bunk, S.; Krestel, R. WELDA: Enhancing topic models by incorporating local word context. In Proceedings of the 18th ACM/IEEE on Joint Conference on Digital Libraries, Fort Worth, TX, USA, 3–7 June 2018; pp. 293–302. [Google Scholar]
  21. Nguyen, D.Q.; Billingsley, R.; Du, L.; Johnson, M. Improving topic models with latent feature word representations. Trans. Assoc. Comput. Linguist. 2015, 3, 299–313. [Google Scholar] [CrossRef]
  22. Li, C.; Wang, H.; Zhang, Z.; Sun, A.; Ma, Z. Topic modeling for short texts with auxiliary word embeddings. In Proceedings of the 39th International ACM SIGIR Conference on Research and Development in Information Retrieval, Pisa, Italy, 17–21 July 2016; pp. 165–174. [Google Scholar]
  23. Qiang, J.; Chen, P.; Wang, T.; Wu, X. Topic modeling over short texts by incorporating word embeddings. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Jeju, Korea, 23–26 May 2017; pp. 363–374. [Google Scholar]
  24. Bicalho, P.V.; Pita, M.; Pedrosa, G.; Lacerda, A.; Pappa, G.L. A general framework to expand short text for topic modeling. Inf. Sci. 2017, 393, 66–81. [Google Scholar] [CrossRef]
  25. Zhao, H.; Phung, D.; Huynh, V.; Jin, Y.; Du, L.; Buntine, W. Topic Modelling Meets Deep Neural Networks: A Survey. In Proceedings of the Thirtieth International Joint Conference on Artificial Intelligence (IJCAI-21), Montreal, QC, Canada, 19–27 August 2021; pp. 4713–4720. [Google Scholar]
  26. Doan, T.; Hoang, T. Benchmarking Neural Topic Models: An Empirical Study. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online Event, 1 August 2021; pp. 4363–4368. [Google Scholar]
  27. Lin, L.; Jiang, H.; Rao, Y. Copula Guided Neural Topic Modelling for Short Texts. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information SIGIR 20, Xi’an, China, 25–30 July 2020; pp. 1773–1776. [Google Scholar] [CrossRef]
  28. Murakami, R.; Chakraborty, B. Neural topic models for short text using pretrained embeddings and its application to real data. In Proceedings of the 2021 IEEE 4th International Conference on Knowledge Innovation and Invention (ICKII), Taichung, Taiwan, 23–25 July 2021; pp. 146–150. [Google Scholar] [CrossRef]
  29. Kingma, D.P.; Welling, M. Auto-encoding variational Bayes. In Proceedings of the ICML, Beijing, China, 21–26 June 2014. [Google Scholar]
  30. Rezende, D.J.; Mohamed, S.; Wierstra, D. Stochastic backpropagation and approximate inference in deep generative models. In Proceedings of the ICML, Beijing, China, 21–26 June 2014. [Google Scholar]
  31. Miao, Y.; Yu, L.; Blunsom, P. Neural Variational Inference for text processing. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 1727–1736. [Google Scholar]
  32. Srivastava, A.; Sutton, C.A. Autoencoding variational inference for topic models. In Proceedings of the International Conference on Learning Representations 2017, Toulon, France, 24–26 April 2017. [Google Scholar]
  33. Burkhardt, S.; Kramer, S. Decoupling sparsity and smoothness in the dirichlet variational autoencoder topic model. J. Mach. Learn. Res. 2019, 20, 1–27. [Google Scholar]
  34. Joo, W.; Lee, W.; Park, S.; Moon, I.C. Dirichlet Variational Autoencoder. Pattern Recognit. 2020, 107, 107514. [Google Scholar] [CrossRef]
  35. Miao, Y.; Grefenstette, E.; Blunsom, P. Discovering discrete latent topics with neural variational inference. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 2410–2419. [Google Scholar]
  36. Ning, X.; Zheng, Y.; Jiang, Z.; Wang, Y.; Yang, H.; Huang, J. Nonparametric Topic Modeling with Neural Inference. Neurocomputing 2020, 399, 296–306. [Google Scholar] [CrossRef] [Green Version]
  37. Larochelle, H.; Lauly, S. A neural Autoregressive topic model. Adv. Neural Inf. Process. Syst. 2012, 4, 2708–2716. [Google Scholar]
  38. Wang, R.; Zhou, D.; He, Y. ATM: Adversarial neural topic model. Inf. Process. Manag. 2019, 56, 102098. [Google Scholar] [CrossRef] [Green Version]
  39. Wang, R.; Hu, X.; Zhou, D.; He, Y.; Xiong, Y.; Ye, C.; Xu, H. Neural Topic Modeling with Bidirectional Adversarial Training. In Proceedings of the 58th Annual Meeting of Association for Computational Linguistics, Online Event, 5–10 July 2020; pp. 340–350. [Google Scholar]
  40. Yang, L.; Wu, F.; Gu, J.; Wang, C.; Cao, X.; Jin, D.; Guo, Y. Graph Attention Topic Modeling Network. In Proceedings of the WWW ’20: Proceedings of The Web Conference 2020, Ljubljana, Slovenia, 19–23 April 2020; pp. 144–154. [Google Scholar] [CrossRef]
  41. Nan, F.; Ding, R.; Nallapati, R.; Xiang, B. Topic Modeling with Wasserstein autoencoders. In Proceedings of the 2019 Meeting of the Association for Computational Linguistic, Florence, Italy, 28 July–2 August 2019; pp. 6345–6381. [Google Scholar] [CrossRef]
  42. Zhao, H.; Phung, D.; Huynh, V.; Le, T.; Buntine, W. Neural topic model via optimal transport. In Proceedings of the ICLR 2021, Vienna, Austria, 4 May 2021. [Google Scholar]
  43. Wang, X.; Yang, Y. Neural topic model with attention for supervised learning. In Proceedings of the 23rd International Conference on artificial Intelligence and Statistics (AISTSTS), Palermo, Italy, 26–28 August 2020; pp. 1147–1156. [Google Scholar]
  44. Zeng, J.; Li, J.; Song, Y.; Gao, C.; Lyu, M.R.; King, I. Topic memory networks for short text classification. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing EMNLP, Brussels, Belgium, 31 October–4 November 2018; pp. 3120–3131. [Google Scholar]
  45. Wu, X.; Li, C.; Zhu, Y.; Miao, Y. Short text topic modeling with topic distribution quantization and negative sampling decoder. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing EMNLP, Online Event, 13 November 2020; pp. 1772–1782. [Google Scholar]
  46. Niu, Y.; Zhang, H.; Li, J. A Nested Chinese Restaurant Topic Model for Short Texts with Document Embeddings. Appl. Sci. 2021, 11, 8708. [Google Scholar] [CrossRef]
  47. Zhao, X.; Wang, D.; Zhao, Z.; Liu, W.; Lu, C.; Zhuang, F. A neural topic model with word vectors and entity vectors for short texts. Inf. Process. Manag. 2021, 58, 102455. [Google Scholar] [CrossRef]
  48. Zhu, Q.; Feng, Z.; Li, X. Graphbtm: Graph enhanced autoencoded variational inference for biterm topic model. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing EMNLP, Brussels, Belgium, 31 October–4 November 2018; pp. 4663–4672. [Google Scholar]
  49. Feng, J.; Zhang, Z.; Ding, C.; Rao, Y.; Xie, H. Context reinforced neural topic modeling over short texts. arXiv 2020, arXiv:2008.04545. [Google Scholar]
  50. Pennington, J.; Socher, R.; Manning, C. Glove: Global vectors for word representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014. [Google Scholar]
  51. Xie, P.; Deng, Y.; Xing, E. Diversifying Restricted Boltzmann Machine for Document Modeling. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, Australia, 10–13 August 2015; pp. 1315–1324. [Google Scholar] [CrossRef]
  52. Zhao, H.; Phung, D.Q.; Huynh, V.; Le, T.; Buntine, W.L. Neural Topic Model via Optimal Transport. arXiv 2020, arXiv:2008.13537. [Google Scholar]
  53. Roder, M.; Both, A.; Hinneburg, A. Exploring the Space of Topic Coherence Measures. In Proceedings of the Eighth ACM International Conference on Web Search and Data Mining, Shanghai, China, 31 January–6 February 2015; pp. 399–408. [Google Scholar]
  54. Lau, J.H.; Newman, D.; Baldwin, T. Machine reading tea leaves: Automatically evaluating topic coherence and topic model quality. In Proceedings of the 14th Conference of the European Chapter of the Association for Computational Linguistics, Gothenburg, Sweden, 26–30 April 2014; pp. 530–539. [Google Scholar]
  55. Ding, R.; Nallapati, R.; Xiang, B. Coherence-Aware neural topic modeling. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; Association for Computational Linguistics: Stroudsburg, PA, USA, 2018; pp. 830–836. [Google Scholar]
  56. Carbone, G.; Sarti, G. ETC-NLG: End-to-end Topic-Conditioned Natural Language Generation. Ital. J. Comput. Linguist. 2020, 6, 61–77. [Google Scholar] [CrossRef]
  57. Chen, Y.; Zaki, M.J. KATE: K-Competitive Autoencoder for Text. In Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, USA, 13–17 August 2017; pp. 85–94. [Google Scholar]
Figure 1. Proposed Study.
Figure 1. Proposed Study.
Sensors 22 00852 g001
Figure 2. VAE-based model.
Figure 2. VAE-based model.
Sensors 22 00852 g002
Figure 3. WAE-based model.
Figure 3. WAE-based model.
Sensors 22 00852 g003
Figure 4. Summary of TC results.
Figure 4. Summary of TC results.
Sensors 22 00852 g004
Figure 5. Summary of TD Results.
Figure 5. Summary of TD Results.
Sensors 22 00852 g005
Figure 6. Classification and clustering performance.
Figure 6. Classification and clustering performance.
Sensors 22 00852 g006
Table 1. Table of notations used.
Table 1. Table of notations used.
Indices:
KNumber of topics k { 1 , , K }
LWord embedding vectors dimension l { 1 , , L }
CNumber of classes c { 1 , , C }
Decision Variables:
DSet of documents
VSet of lexicons, vocabularies
XBoW matrix of all documents, X R + | V | × | D |
x d d’s BoW representation vector, x d R + | V |
NNumber of words that occurred in document d
w n n-th word
Random Variables:
h ( i ) i-th hidden layer’s outputs
hGaussian random variables, h R K
z n latent topic for the n-th word
θ Topic proportion vector, θ R + K
β Topic–word distribution β R | V | × K
α Topic centroid vectors α R L × K
ρ Word embedding vectors ρ R L × | V |
Table 2. Details of Data set.
Table 2. Details of Data set.
Name | D | | V | X ave dLmax dLSparsityCSource
BBC_news22252949267,259120.1211760.02781151
20NewsGroup16,3091612783,15148.0230330.018855201
SearchSnippets12,2954720177,33814.42370.00216782
TrecTweet2472509821,1488.56200.001561892
Biomedical19,4483892144,6837.44280.001870202
GoogleNews11,108811069,2296.23140.0007641522
M108355169649,3855.91210.003411101
DBLP54,5951513294,7575.40210.00352741
PascalFlicker4834343125,9805.37190.001548202
StackOverflow16,407230382,3425.02170.002145202
Table 3. Topic Coherence on 20NewsGroups.
Table 3. Topic Coherence on 20NewsGroups.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.16−0.040.470.330.160.15
pwe−0.21−0.050.500.320.160.15
pwe/fine-tuning−0.12−0.030.480.340.170.17
GSMnone−0.18−0.050.500.320.150.15
pwe−0.21−0.050.500.320.160.15
pwe/fine-tuning−0.21−0.090.430.290.120.13
NSTMnone−0.17−0.050.530.330.130.16
pwe−0.19−0.050.530.330.150.15
pwe/fine-tuning−0.16−0.040.530.330.140.16
NVDMnone−0.27−0.050.470.300.140.14
pwe−0.22−0.050.510.320.150.15
pwe/fine-tuning−0.27−0.080.450.280.120.12
NVLDAnone−0.18−0.040.520.340.140.16
pwe−0.19−0.040.520.330.140.15
pwe/fine-tuning−0.10−0.040.550.360.160.18
ProdLDAnone−0.21−0.050.480.310.160.14
pwe−0.22−0.050.500.320.150.15
pwe/fine-tuning−0.21−0.090.460.290.120.12
RSBnone−0.19−0.060.480.310.150.14
pwe−0.18−0.040.520.330.150.16
pwe/fine-tuning−0.26−0.090.420.260.090.11
WLDAnone−0.28-0.440.290.130.12
pwe−0.17−0.040.520.330.150.16
pwe/fine-tuning−0.32−0.070.510.320.150.14
Table 4. Topic Coherence on BBCNews.
Table 4. Topic Coherence on BBCNews.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.36−0.090.440.280.130.13
pwe−0.42−0.110.430.260.140.11
pwe/fine-tuning−0.35−0.100.410.270.150.12
GSMnone−0.40−0.100.430.270.130.12
pwe−0.41−0.100.440.270.140.12
pwe/fine-tuning−0.32−0.090.450.290.160.14
NSTMnone−0.40−0.100.440.270.140.12
pwe−0.41−0.100.460.270.130.12
pwe/fine-tuning−0.41−0.090.450.270.140.12
NVDMnone−0.44−0.120.420.250.130.10
pwe−0.42−0.110.430.260.120.11
pwe/fine-tuning−0.45−0.170.380.220.120.08
NVLDAnone−0.38−0.090.460.280.150.13
pwe−0.39−0.090.450.270.130.12
pwe/fine-tuning−0.18−0.030.530.360.160.20
ProdLDAnone−0.43−0.130.410.250.120.10
pwe−0.42−0.100.430.260.130.11
pwe/fine-tuning−0.27−0.060.460.310.150.16
RSBnone−0.40−0.090.440.270.130.12
pwe−0.40−0.090.450.270.130.12
pwe/fine-tuning−0.43−0.180.340.210.140.09
WLDAnone−0.41-0.440.270.140.12
pwe−0.43−0.110.430.260.130.11
pwe/fine-tuning−0.39−0.100.430.270.150.12
Table 5. Topic Coherence on Biomedical.
Table 5. Topic Coherence on Biomedical.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.52−0.130.190.110.090.06
pwe−0.46−0.120.220.240.120.19
pwe/fine-tuning−0.51−0.140.190.230.090.18
GSMnone−0.52−0.120.190.100.090.06
pwe−0.47−0.120.210.240.110.19
pwe/fine-tuning−0.49−0.130.190.230.100.19
NSTMnone−0.52−0.130.210.120.100.06
pwe−0.49−0.130.210.120.100.06
pwe/fine-tuning−0.52−0.130.210.130.100.06
NVDMnone−0.53−0.120.180.090.080.05
pwe−0.51−0.130.200.120.090.06
pwe/fine-tuning−0.52−0.140.240.130.100.06
NVLDAnone−0.47−0.130.250.160.110.08
pwe−0.49−0.130.210.230.100.18
pwe/fine-tuning−0.15−0.010.290.360.330.27
ProdLDAnone−0.50−0.110.160.090.090.06
pwe−0.51−0.130.210.230.090.18
pwe/fine-tuning−0.44−0.120.250.250.140.19
RSBnone−0.49−0.140.190.120.100.07
pwe−0.48−0.130.220.240.110.18
pwe/fine-tuning−0.47−0.130.180.250.110.20
WLDAnone−0.51-0.250.140.100.07
pwe−0.48−0.130.200.130.110.07
pwe/fine-tuning−0.55−0.110.150.080.090.06
Table 6. Topic Coherence on DBLP.
Table 6. Topic Coherence on DBLP.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.52−0.140.340.210.110.09
pwe−0.35−0.100.370.250.130.11
pwe/fine-tuning−0.42−0.130.350.230.120.10
GSMnone−0.52−0.140.330.210.120.09
pwe−0.38−0.110.380.240.130.11
pwe/fine-tuning−0.43−0.130.360.220.120.10
NSTMnone−0.33−0.080.370.260.130.13
pwe−0.42−0.130.370.230.110.10
pwe/fine-tuning−0.31−0.070.400.270.150.14
NVDMnone−0.48−0.140.340.220.110.09
pwe−0.47−0.130.340.220.110.09
pwe/fine-tuning−0.45−0.130.350.230.130.10
NVLDAnone−0.41−0.110.370.240.130.11
pwe−0.41−0.120.370.240.130.10
pwe/fine-tuning−0.29−0.090.390.270.150.13
ProdLDAnone−0.43−0.140.340.210.120.09
pwe−0.41−0.120.370.240.130.11
pwe/fine-tuning−0.45−0.140.350.210.120.09
RSBnone−0.41−0.130.360.230.130.10
pwe−0.37−0.110.380.240.120.11
pwe/fine-tuning−0.37−0.130.330.220.130.10
WLDAnone−0.43-0.370.240.130.11
pwe−0.39−0.110.380.250.130.11
pwe/fine-tuning−0.35−0.060.380.260.130.12
Table 7. Topic Coherence on GoogleNews.
Table 7. Topic Coherence on GoogleNews.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.52−0.150.300.140.090.03
pwe−0.48−0.160.270.290.090.25
pwe/fine-tuning−0.51−0.170.270.220.080.15
GSMnone−0.53−0.160.280.130.090.03
pwe−0.48−0.160.280.270.090.21
pwe/fine-tuning−0.51−0.150.280.240.080.18
NSTMnone−0.54−0.160.320.170.100.05
pwe−0.48−0.160.290.130.080.03
pwe/fine-tuning−0.53−0.160.310.160.090.04
NVDMnone−0.53−0.150.250.110.070.02
pwe−0.50−0.160.280.130.080.03
pwe/fine-tuning−0.51−0.150.290.150.080.04
NVLDAnone−0.50−0.150.340.190.100.06
pwe−0.49−0.160.280.240.080.17
pwe/fine-tuning−0.55−0.140.390.270.120.19
ProdLDAnone−0.51−0.150.280.120.080.03
pwe−0.49−0.160.270.220.080.15
pwe/fine-tuning−0.53−0.180.290.220.090.14
RSBnone−0.53−0.180.300.150.090.04
pwe−0.48−0.160.290.250.090.18
pwe/fine-tuning−0.58−0.180.300.240.100.17
WLDAnone−0.53-0.250.110.080.02
pwe−0.50−0.160.280.130.080.03
pwe/fine-tuning−0.55−0.120.410.240.130.10
Table 8. Topic Coherence on M10.
Table 8. Topic Coherence on M10.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.54−0.130.370.230.110.10
pwe−0.51−0.110.410.250.130.11
pwe/fine-tuning−0.54−0.130.380.220.110.10
GSMnone−0.55−0.140.370.220.110.10
pwe−0.51−0.120.390.240.130.11
pwe/fine-tuning−0.54−0.150.350.200.100.08
NSTMnone−0.43−0.080.420.270.130.13
pwe−0.52−0.120.420.240.130.11
pwe/fine-tuning−0.45−0.080.420.270.130.13
NVDMnone−0.55−0.140.360.220.120.09
pwe−0.54−0.120.370.230.120.10
pwe/fine-tuning−0.53−0.130.380.240.120.11
NVLDAnone−0.51−0.110.390.240.130.11
pwe−0.51−0.120.390.240.120.11
pwe/fine-tuning−0.41−0.030.450.310.170.16
ProdLDAnone−0.52−0.140.370.220.120.09
pwe−0.53−0.120.380.240.120.10
pwe/fine-tuning−0.52−0.130.350.220.110.10
RSBnone−0.52−0.130.370.230.110.10
pwe−0.50−0.100.400.250.140.12
pwe/fine-tuning−0.53−0.140.360.220.120.10
WLDAnone−0.55-0.360.210.110.09
pwe−0.55−0.130.380.230.120.10
pwe/fine-tuning−0.54−0.110.400.250.130.11
Table 9. Topic Coherence on PascalFlicker.
Table 9. Topic Coherence on PascalFlicker.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.51−0.100.260.160.080.06
pwe−0.46−0.100.270.240.080.17
pwe/fine-tuning−0.49−0.090.270.230.100.15
GSMnone−0.52−0.100.260.160.080.06
pwe−0.46−0.100.260.240.090.17
pwe/fine-tuning−0.47−0.090.260.210.090.14
NSTMnone−0.52−0.090.240.150.080.05
pwe−0.48−0.090.240.150.090.05
pwe/fine-tuning−0.51−0.090.250.150.080.05
NVDMnone−0.52−0.090.250.140.090.05
pwe−0.51−0.090.250.140.080.05
pwe/fine-tuning−0.50−0.100.270.150.080.06
NVLDAnone−0.48−0.100.280.180.090.07
pwe−0.45−0.100.270.230.090.17
pwe/fine-tuning−0.40−0.070.450.340.160.25
ProdLDAnone−0.48−0.090.250.140.080.05
pwe−0.49−0.100.260.220.080.15
pwe/fine-tuning−0.40−0.070.390.300.120.21
RSBnone−0.48−0.090.310.200.090.08
pwe−0.45−0.100.260.240.080.18
pwe/fine-tuning−0.48−0.100.320.270.110.19
WLDAnone−0.49-0.260.150.080.05
pwe−0.52−0.090.260.140.080.05
pwe/fine-tuning−0.54−0.110.300.190.100.08
Table 10. Topic Coherence on SearchSnippets.
Table 10. Topic Coherence on SearchSnippets.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.49−0.150.270.140.090.04
pwe−0.46−0.140.320.280.120.20
pwe/fine-tuning−0.49−0.170.260.220.100.14
GSMnone−0.50−0.160.250.130.090.03
pwe−0.46−0.140.310.270.120.18
pwe/fine-tuning−0.48−0.160.270.210.100.13
NSTMnone−0.49−0.130.320.190.110.06
pwe−0.47−0.140.300.180.110.06
pwe/fine-tuning−0.49−0.130.310.180.100.06
NVDMnone−0.50−0.150.260.130.090.03
pwe−0.48−0.140.310.170.100.05
pwe/fine-tuning−0.48−0.130.320.200.110.06
NVLDAnone−0.46−0.120.360.220.130.09
pwe−0.45−0.130.300.260.110.17
pwe/fine-tuning−0.26−0.060.480.360.180.24
ProdLDAnone−0.48−0.150.280.160.110.05
pwe−0.48−0.140.310.240.110.14
pwe/fine-tuning−0.42−0.140.300.240.140.15
RSBnone−0.47−0.150.270.150.100.05
pwe−0.46−0.130.310.280.120.19
pwe/fine-tuning−0.48−0.140.240.200.090.13
WLDAnone−0.49-0.290.170.110.05
pwe−0.47−0.130.320.180.110.06
pwe/fine-tuning−0.31−0.070.440.330.170.17
Table 11. Topic Coherence on StackOverflow.
Table 11. Topic Coherence on StackOverflow.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.53−0.100.310.220.150.08
pwe−0.48−0.100.300.310.170.21
pwe/fine-tuning−0.50−0.120.270.290.140.17
GSMnone−0.53−0.100.300.210.140.08
pwe−0.48−0.110.300.310.160.21
pwe/fine-tuning−0.49−0.100.240.260.140.15
NSTMnone−0.48−0.100.280.230.160.09
pwe−0.50−0.100.310.230.140.09
pwe/fine-tuning−0.47−0.100.300.240.150.09
NVDMnone−0.55−0.100.250.170.120.07
pwe−0.52−0.100.300.220.140.08
pwe/fine-tuning−0.53−0.090.300.220.140.08
NVLDAnone−0.48−0.090.320.250.160.10
pwe−0.50−0.100.290.300.150.20
pwe/fine-tuning−0.26−0.050.370.390.250.24
ProdLDAnone−0.51−0.100.250.190.120.07
pwe−0.52−0.100.290.280.150.17
pwe/fine-tuning−0.44−0.120.260.290.170.16
RSBnone−0.49−0.110.260.220.170.08
pwe−0.48−0.090.300.310.150.21
pwe/fine-tuning−0.48−0.130.190.280.170.16
WLDAnone−0.53-0.310.230.150.09
pwe−0.53−0.100.290.220.150.08
pwe/fine-tuning−0.51−0.090.340.280.180.11
Table 12. Topic Coherence on TrecTweet.
Table 12. Topic Coherence on TrecTweet.
ModelResourcesTrainingCorpusExternalCorpusglove.42B.300dglove.6B.300d
MethodnpminpmiWETC c WETC pw WETC c WETC pw
PWE
GSBnone−0.53−0.140.290.140.080.04
pwe−0.50−0.140.290.240.090.17
pwe/fine-tuning−0.54−0.150.290.230.080.15
GSMnone−0.53−0.140.280.140.080.03
pwe−0.50−0.140.270.240.080.17
pwe/fine-tuning−0.53−0.140.280.220.090.14
NSTMnone−0.53−0.130.290.150.090.04
pwe−0.51−0.140.290.140.080.03
pwe/fine-tuning−0.53−0.140.290.150.090.04
NVDMnone−0.51−0.150.280.140.080.04
pwe−0.51−0.140.280.140.080.03
pwe/fine-tuning−0.52−0.140.290.150.090.04
NVLDAnone−0.51−0.140.320.180.090.05
pwe−0.50−0.140.290.240.090.17
pwe/fine-tuning−0.55−0.120.350.270.110.17
ProdLDAnone−0.53−0.140.280.130.080.03
pwe−0.51−0.140.290.210.090.13
pwe/fine-tuning−0.55−0.140.320.250.100.16
RSBnone−0.53−0.150.300.160.090.04
pwe−0.50−0.130.300.250.090.17
pwe/fine-tuning−0.57−0.150.280.270.100.17
WLDAnone−0.50-0.290.140.090.04
pwe−0.51−0.140.270.140.080.03
pwe/fine-tuning−0.54−0.150.340.180.100.06
Table 13. Topic Diversity on 20NewsGroups.
Table 13. Topic Diversity on 20NewsGroups.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.780.990.080.170.60
pwe0.790.990.20−0.180.67
pwe/fine-tuning0.680.980.310.150.52
GSMnone0.841.000.090.170.66
pwe0.851.000.210.180.71
pwe/fine-tuning0.720.990.390.190.77
NSTMnone0.830.990.930.170.74
pwe0.851.000.910.170.68
pwe/fine-tuning0.820.990.910.170.75
NVDMnone0.700.990.180.190.73
pwe0.851.000.210.180.71
pwe/fine-tuning0.831.000.550.200.84
NVLDAnone0.911.000.070.170.70
pwe0.820.990.190.170.83
pwe/fine-tuning0.640.990.350.150.65
ProdLDAnone0.770.990.110.180.68
pwe0.841.000.220.180.73
pwe/fine-tuning0.640.990.390.200.84
RSBnone0.881.000.070.180.65
pwe0.780.990.210.170.69
pwe/fine-tuning0.620.990.440.210.94
WLDAnone0.780.990.550.200.77
0pwe0.831.000.260.170.72
pwe/fine-tuning0.560.980.000.180.73
Table 14. Topic Diversity on BBCNews.
Table 14. Topic Diversity on BBCNews.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.850.990.070.210.80
pwe0.901.000.150.220.75
pwe/fine-tuning0.871.000.250.200.71
GSMnone0.921.000.070.220.79
pwe0.911.000.150.220.80
pwe/fine-tuning0.841.000.300.190.68
NSTMnone1.001.000.970.210.74
pwe0.901.000.860.220.80
pwe/fine-tuning0.951.000.760.210.74
NVDMnone0.921.000.050.230.80
pwe0.921.000.180.230.78
pwe/fine-tuning0.881.000.650.270.81
NVLDAnone0.971.000.060.210.74
pwe0.911.000.170.220.78
pwe/fine-tuning0.700.990.430.150.69
ProdLDAnone0.810.990.050.240.85
pwe0.911.000.160.230.75
pwe/fine-tuning0.690.990.390.180.71
RSBnone0.951.000.060.210.78
pwe0.881.000.190.210.80
pwe/fine-tuning0.730.990.400.270.72
WLDAnone0.941.000.040.220.75
pwe0.901.000.180.220.79
pwe/fine-tuning0.851.000.650.210.71
Table 15. Topic Diversity on Biomedical.
Table 15. Topic Diversity on Biomedical.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.961.000.080.560.93
pwe0.921.000.190.490.85
pwe/fine-tuning0.951.000.350.540.89
GSMnone0.951.000.080.550.93
pwe0.931.000.170.500.86
pwe/fine-tuning0.911.000.300.530.89
NSTMnone1.001.000.790.500.90
pwe0.951.000.840.490.86
pwe/fine-tuning1.001.000.800.530.91
NVDMnone0.851.000.190.590.97
pwe0.931.000.110.510.94
pwe/fine-tuning0.951.000.170.460.90
NVLDAnone0.971.000.060.450.85
pwe0.921.000.130.540.89
pwe/fine-tuning0.350.960.490.260.37
ProdLDAnone0.780.990.120.670.92
pwe0.941.000.140.490.90
pwe/fine-tuning0.850.990.430.390.78
RSBnone0.911.000.080.560.89
pwe0.881.000.190.490.85
pwe/fine-tuning0.620.980.440.620.87
WLDAnone0.911.000.260.450.88
pwe0.921.000.220.520.87
pwe/fine-tuning0.730.990.000.710.97
Table 16. Topic Diversity on DBLP.
Table 16. Topic Diversity on DBLP.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.861.000.090.270.88
pwe0.831.000.290.240.81
pwe/fine-tuning0.891.000.270.270.83
GSMnone0.861.000.090.280.86
pwe0.861.000.280.250.85
pwe/fine-tuning0.800.990.180.280.85
NSTMnone0.710.980.760.270.89
pwe0.871.000.910.260.89
pwe/fine-tuning0.700.980.770.260.82
NVDMnone0.881.000.580.270.87
pwe0.871.000.210.260.85
pwe/fine-tuning0.901.000.120.270.85
NVLDAnone0.901.000.060.250.81
pwe0.871.000.290.250.80
pwe/fine-tuning0.740.990.110.220.72
ProdLDAnone0.820.990.120.310.85
pwe0.871.000.260.250.79
pwe/fine-tuning0.821.000.100.290.85
RSBnone0.861.000.180.270.81
pwe0.650.980.340.250.84
pwe/fine-tuning0.680.990.400.270.81
WLDAnone0.851.000.380.250.79
pwe0.841.000.260.240.81
pwe/fine-tuning0.780.990.000.240.80
Table 17. Topic Diversity on GoogleNews.
Table 17. Topic Diversity on GoogleNews.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.981.000.090.380.96
pwe0.961.000.130.430.95
pwe/fine-tuning0.961.000.340.450.98
GSMnone0.981.000.090.410.97
pwe0.971.000.120.420.97
pwe/fine-tuning0.911.000.430.410.96
NSTMnone1.001.000.840.340.93
pwe0.991.000.690.400.98
pwe/fine-tuning1.001.000.810.360.94
NVDMnone0.961.000.070.500.99
pwe0.971.000.060.430.99
pwe/fine-tuning0.961.000.100.390.95
NVLDAnone0.991.000.060.320.90
pwe0.971.000.100.440.97
pwe/fine-tuning0.720.990.450.270.86
ProdLDAnone0.901.000.110.430.97
pwe0.971.000.090.450.98
pwe/fine-tuning0.911.000.490.380.97
RSBnone0.891.000.140.390.93
pwe0.871.000.200.430.96
pwe/fine-tuning0.710.990.400.370.92
WLDAnone0.931.000.210.470.98
pwe0.921.000.110.410.97
pwe/fine-tuning0.820.990.240.250.82
Table 18. Topic Diversity on M10.
Table 18. Topic Diversity on M10.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.911.000.080.260.84
pwe0.861.000.220.230.81
pwe/fine-tuning0.891.000.280.250.88
GSMnone0.891.000.080.270.85
pwe0.861.000.200.250.83
pwe/fine-tuning0.770.990.320.290.90
NSTMnone0.750.980.770.230.87
pwe0.840.990.900.240.82
pwe/fine-tuning0.820.990.750.240.86
NVDMnone0.851.000.040.280.86
pwe0.871.000.210.260.84
pwe/fine-tuning0.891.000.320.250.86
NVLDAnone0.911.000.070.250.83
pwe0.871.000.220.240.83
pwe/fine-tuning0.640.990.200.190.68
ProdLDAnone0.790.990.050.280.88
pwe0.871.000.190.260.84
pwe/fine-tuning0.810.990.200.270.87
RSBnone0.871.000.150.250.83
pwe0.670.980.240.240.80
pwe/fine-tuning0.690.990.410.260.82
WLDAnone0.871.000.160.290.90
pwe0.790.990.330.260.82
pwe/fine-tuning0.851.000.440.230.80
Table 19. Topic Diversity on PascalFlicker.
Table 19. Topic Diversity on PascalFlicker.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.951.000.070.410.98
pwe0.911.000.170.400.97
pwe/fine-tuning0.931.000.550.380.95
GSMnone0.941.000.070.420.97
pwe0.931.000.170.420.96
pwe/fine-tuning0.891.000.610.390.97
NSTMnone1.001.000.790.480.99
pwe0.951.000.850.440.95
pwe/fine-tuning1.001.000.780.450.98
NVDMnone0.931.000.050.460.98
pwe0.931.000.170.440.99
pwe/fine-tuning0.921.000.240.400.98
NVLDAnone0.951.000.060.380.96
pwe0.911.000.180.400.97
pwe/fine-tuning0.490.980.750.180.71
ProdLDAnone0.770.990.070.470.98
pwe0.931.000.160.430.98
pwe/fine-tuning0.590.970.820.220.84
RSBnone0.911.000.090.320.94
pwe0.861.000.180.420.98
pwe/fine-tuning0.620.990.530.290.90
WLDAnone0.941.000.030.430.98
pwe0.901.000.170.430.99
pwe/fine-tuning0.830.990.400.340.95
Table 20. Topic Diversity on SearchSnippets.
Table 20. Topic Diversity on SearchSnippets.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.971.000.070.410.94
pwe0.941.000.200.320.81
pwe/fine-tuning0.961.000.340.420.91
GSMnone0.951.000.070.450.93
pwe0.941.000.160.340.83
pwe/fine-tuning0.931.000.330.410.92
NSTMnone1.001.000.860.320.87
pwe0.961.000.810.340.85
pwe/fine-tuning1.001.000.830.330.89
NVDMnone0.901.000.070.450.95
pwe0.941.000.100.340.87
pwe/fine-tuning0.961.000.140.320.86
NVLDAnone0.981.000.060.280.78
pwe0.951.000.130.330.85
pwe/fine-tuning0.510.970.640.170.68
ProdLDAnone0.881.000.070.370.87
pwe0.951.000.120.340.89
pwe/fine-tuning0.780.990.560.320.82
RSBnone0.921.000.080.430.91
pwe0.860.990.230.330.82
pwe/fine-tuning0.780.990.390.480.94
WLDAnone0.921.000.250.360.89
pwe0.931.000.150.340.86
pwe/fine-tuning0.340.960.720.180.64
Table 21. Topic Diversity on StackOverflow.
Table 21. Topic Diversity on StackOverflow.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.941.000.090.290.70
pwe0.891.000.190.280.69
pwe/fine-tuning0.931.000.340.340.71
GSMnone0.931.000.090.320.74
pwe0.901.000.200.340.68
pwe/fine-tuning0.790.990.380.370.69
NSTMnone0.951.000.770.320.71
pwe0.901.000.890.280.74
pwe/fine-tuning0.961.000.780.320.71
NVDMnone0.820.990.120.430.82
pwe0.891.000.180.340.72
pwe/fine-tuning0.921.000.250.320.75
NVLDAnone0.941.000.060.290.69
pwe0.881.000.200.340.69
pwe/fine-tuning0.460.970.230.220.47
ProdLDAnone0.750.990.060.380.79
pwe0.901.000.190.300.70
pwe/fine-tuning0.770.990.320.340.63
RSBnone0.881.000.130.370.64
pwe0.710.990.250.330.68
pwe/fine-tuning0.650.990.390.490.70
WLDAnone0.881.000.400.290.71
pwe0.841.000.300.340.71
pwe/fine-tuning0.831.000.160.240.59
Table 22. Topic Diversity on TrecTweet.
Table 22. Topic Diversity on TrecTweet.
ModelPWETopic DiversityInverted RBOMSCDTCD (42B)TCD (6B)
GSBnone0.971.000.060.410.97
pwe0.951.000.150.420.96
pwe/fine-tuning0.941.000.430.360.96
GSMnone0.971.000.060.420.96
pwe0.961.000.140.430.96
pwe/fine-tuning0.951.000.530.400.96
NSTMnone1.001.000.550.420.96
pwe0.971.000.790.410.97
pwe/fine-tuning1.001.000.770.430.96
NVDMnone0.961.000.050.410.96
pwe0.951.000.140.440.98
pwe/fine-tuning0.971.000.250.390.96
NVLDAnone0.981.000.060.360.94
pwe0.951.000.140.400.96
pwe/fine-tuning0.830.990.410.260.87
ProdLDAnone0.921.000.050.420.97
pwe0.961.000.140.430.96
pwe/fine-tuning0.760.990.640.310.92
RSBnone0.921.000.080.360.95
pwe0.881.000.190.390.95
pwe/fine-tuning0.520.980.570.350.96
WLDAnone0.961.000.050.400.96
pwe0.941.000.120.440.97
pwe/fine-tuning0.871.000.440.330.94
Table 23. Classification and Clustering performance on 20NewsGroups.
Table 23. Classification and Clustering performance on 20NewsGroups.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.340.320.320.290.300.290.280.270.180.14
pwe0.340.330.330.310.340.340.320.310.260.25
pwe/fine-tuning0.480.460.470.450.490.480.480.470.330.33
GSMnone0.300.280.280.250.250.230.240.210.200.14
pwe0.310.280.300.270.300.290.280.270.240.21
pwe/fine-tuning0.450.420.440.410.450.430.430.420.350.33
NSTMnone0.220.230.210.180.070.150.070.030.020.07
pwe0.140.180.130.110.080.290.080.060.040.08
pwe/fine-tuning0.190.190.180.150.070.050.070.020.010.06
NVDMnone0.360.340.350.320.230.220.220.200.020.07
pwe0.380.360.370.340.330.330.320.320.110.13
pwe/fine-tuning0.540.530.520.500.550.530.530.530.210.20
NVLDAnone0.310.300.300.270.260.290.250.250.060.11
pwe0.200.200.190.150.120.150.110.080.070.09
pwe/fine-tuning0.360.340.350.310.270.340.260.230.330.29
ProdLDAnone0.180.150.170.120.110.080.100.060.050.07
pwe0.180.170.170.130.100.140.100.060.050.08
pwe/fine-tuning0.360.340.340.300.270.330.260.230.320.27
RSBnone0.150.090.140.070.140.090.130.060.180.13
pwe0.110.050.100.050.100.030.090.030.080.09
pwe/fine-tuning0.300.260.290.240.300.290.290.250.210.15
WLDAnone0.070.100.070.030.070.130.070.030.000.06
pwe0.070.110.070.040.070.120.070.040.000.06
pwe/fine-tuning0.070.100.070.040.070.100.070.040.000.06
Table 24. Classification and Clustering performance on BBCNews.
Table 24. Classification and Clustering performance on BBCNews.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.940.940.940.940.950.950.940.940.490.79
pwe0.890.890.890.890.890.890.880.880.420.81
pwe/fine-tuning0.920.920.920.920.910.920.910.910.510.86
GSMnone0.940.940.940.940.940.940.940.940.510.78
pwe0.890.890.880.880.880.880.880.880.430.81
pwe/fine-tuning0.940.940.930.930.920.920.920.920.560.91
NSTMnone0.570.690.560.570.520.660.510.510.160.51
pwe0.630.690.620.630.510.650.500.500.170.46
pwe/fine-tuning0.600.630.580.580.520.600.500.490.160.42
NVDMnone0.920.920.910.910.830.830.830.830.210.41
pwe0.920.930.920.920.880.880.880.880.320.57
pwe/fine-tuning0.950.950.950.950.950.950.950.950.440.78
NVLDAnone0.810.810.800.800.830.830.820.820.140.47
pwe0.870.880.870.870.780.820.770.780.400.72
pwe/fine-tuning0.920.930.920.920.900.910.900.900.610.92
ProdLDAnone0.920.930.920.920.880.880.870.870.440.62
pwe0.850.870.850.850.740.810.730.740.330.59
pwe/fine-tuning0.930.930.930.930.910.910.910.910.640.92
RSBnone0.430.290.390.270.430.280.390.270.300.41
pwe0.660.660.650.640.660.660.650.640.220.46
pwe/fine-tuning0.930.930.930.930.930.930.930.930.420.58
WLDAnone0.340.360.320.270.280.270.250.160.010.24
pwe0.300.290.280.250.290.270.260.210.000.23
pwe/fine-tuning0.310.330.290.260.290.310.270.220.000.23
Table 25. Classification and Clustering performance on Biomedical.
Table 25. Classification and Clustering performance on Biomedical.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.390.390.390.350.490.510.490.490.110.19
pwe0.220.200.220.180.250.250.250.230.070.14
pwe/fine-tuning0.310.300.310.280.360.360.360.350.130.21
GSMnone0.400.380.400.360.490.510.490.490.110.20
pwe0.250.230.250.220.260.270.260.250.100.18
pwe/fine-tuning0.320.300.320.280.350.350.350.340.150.23
NSTMnone0.250.230.240.210.290.280.290.280.000.05
pwe0.280.260.280.240.320.320.320.310.010.06
pwe/fine-tuning0.230.220.230.190.280.270.270.260.000.05
NVDMnone0.420.410.420.380.550.570.550.550.060.14
pwe0.380.360.380.340.510.520.510.510.050.12
pwe/fine-tuning0.450.440.440.410.570.590.570.570.050.12
NVLDAnone0.280.250.280.230.260.250.260.240.060.15
pwe0.060.020.060.010.060.020.060.010.000.05
pwe/fine-tuning0.100.070.100.060.070.040.070.030.030.08
ProdLDAnone0.140.130.140.100.090.100.090.060.050.10
pwe0.060.010.050.010.060.010.050.010.000.05
pwe/fine-tuning0.110.080.110.070.070.050.070.030.030.08
RSBnone0.180.140.180.130.200.170.190.160.040.09
pwe0.080.030.080.030.080.030.080.030.010.06
pwe/fine-tuning0.160.130.160.130.170.150.170.140.040.10
WLDAnone0.080.060.080.040.080.060.080.040.000.05
pwe0.080.050.080.040.080.050.080.040.000.05
pwe/fine-tuning0.070.050.070.030.080.060.070.040.000.05
Table 26. Classification and Clustering performance on DBLP.
Table 26. Classification and Clustering performance on DBLP.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.690.650.580.580.740.700.650.670.040.46
pwe0.620.600.520.530.650.630.550.570.090.55
pwe/fine-tuning0.720.680.620.630.750.710.660.680.110.56
GSMnone0.680.640.570.570.740.700.650.670.040.44
pwe0.640.610.550.560.660.640.560.580.110.59
pwe/fine-tuning0.720.690.620.640.740.710.660.670.140.61
NSTMnone0.480.460.360.350.520.510.410.410.010.39
pwe0.500.470.400.390.560.550.460.470.000.38
pwe/fine-tuning0.460.460.340.320.500.500.400.400.000.38
NVDMnone0.710.670.610.620.760.720.680.700.040.46
pwe0.640.620.540.540.720.690.630.650.020.42
pwe/fine-tuning0.710.670.610.610.760.720.680.700.040.45
NVLDAnone0.530.500.420.410.540.530.430.430.030.45
pwe0.380.150.250.150.380.130.250.150.000.38
pwe/fine-tuning0.570.540.450.440.540.520.420.400.150.55
ProdLDAnone0.660.650.540.560.630.660.520.540.200.64
pwe0.380.130.250.140.380.110.250.140.000.38
pwe/fine-tuning0.570.480.440.420.540.470.410.380.140.53
RSBnone0.710.670.610.610.750.710.660.680.040.43
pwe0.410.230.310.240.400.190.300.220.020.39
pwe/fine-tuning0.690.640.580.580.700.660.610.620.110.50
WLDAnone0.390.230.260.180.390.240.260.180.000.38
pwe0.390.250.260.180.390.280.260.180.000.38
pwe/fine-tuning0.390.240.260.170.390.270.260.170.000.38
Table 27. Classification and Clustering performance on GoogleNews.
Table 27. Classification and Clustering performance on GoogleNews.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.850.810.780.780.850.820.740.770.460.25
pwe0.550.370.360.330.520.300.300.270.630.41
pwe/fine-tuning0.850.770.740.740.830.760.700.710.540.27
GSMnone0.840.820.750.770.850.820.740.770.470.25
pwe0.530.300.310.270.510.240.270.230.680.46
pwe/fine-tuning0.800.650.600.600.820.680.640.650.670.40
NSTMnone0.150.040.040.030.090.020.020.020.000.04
pwe0.520.460.400.410.490.440.320.350.010.04
pwe/fine-tuning0.150.040.050.030.110.030.030.020.010.04
NVDMnone0.870.850.830.830.890.870.820.840.300.15
pwe0.850.830.810.810.880.870.800.830.320.17
pwe/fine-tuning0.880.830.830.820.890.870.810.830.270.13
NVLDAnone0.560.450.410.400.480.370.300.300.330.19
pwe0.140.030.040.030.140.030.040.030.170.12
pwe/fine-tuning0.320.080.120.080.290.070.100.070.530.30
ProdLDAnone0.410.120.170.120.380.120.150.110.610.36
pwe0.110.020.030.020.100.020.030.020.110.08
pwe/fine-tuning0.350.080.130.090.330.080.120.080.590.32
RSBnone0.590.490.480.470.590.480.470.470.340.12
pwe0.070.000.010.000.070.000.010.000.060.05
pwe/fine-tuning0.650.430.440.410.640.390.410.380.370.14
WLDAnone0.050.010.010.010.060.010.010.010.000.04
pwe0.060.010.020.010.060.010.020.010.000.04
pwe/fine-tuning0.060.010.020.010.060.010.020.010.000.04
Table 28. Classification and Clustering performance on M10.
Table 28. Classification and Clustering performance on M10.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.660.620.590.590.730.740.670.680.140.34
pwe0.520.460.460.450.530.500.470.470.220.46
pwe/fine-tuning0.660.620.590.590.680.670.630.630.260.50
GSMnone0.660.620.590.590.720.740.670.680.140.34
pwe0.540.490.480.480.540.500.480.480.240.50
pwe/fine-tuning0.660.630.600.600.680.680.620.630.300.54
NSTMnone0.400.360.350.350.500.450.450.450.040.17
pwe0.410.390.360.360.490.490.440.440.000.14
pwe/fine-tuning0.360.330.320.310.440.420.400.400.010.14
NVDMnone0.620.600.560.550.720.740.660.670.080.26
pwe0.570.530.500.500.700.730.640.650.070.26
pwe/fine-tuning0.690.670.640.640.760.770.710.730.100.28
NVLDAnone0.410.370.360.350.430.400.380.380.080.28
pwe0.190.150.150.100.200.180.160.110.080.20
pwe/fine-tuning0.520.530.450.450.500.570.430.440.330.51
ProdLDAnone0.640.580.570.560.560.630.500.520.340.59
pwe0.160.110.120.060.180.160.140.080.050.18
pwe/fine-tuning0.540.570.480.480.520.580.450.460.330.54
RSBnone0.640.590.570.570.690.700.630.640.110.26
pwe0.220.130.180.120.210.130.180.110.050.19
pwe/fine-tuning0.640.570.580.570.660.590.590.590.210.32
WLDAnone0.160.110.130.090.160.110.130.090.000.13
pwe0.160.110.130.080.160.130.130.090.000.13
pwe/fine-tuning0.160.120.130.080.160.120.130.080.000.13
Table 29. Classification and Clustering performance on PascalFlicker.
Table 29. Classification and Clustering performance on PascalFlicker.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.340.340.340.310.410.450.420.410.140.19
pwe0.220.210.220.190.230.220.230.210.140.18
pwe/fine-tuning0.300.280.300.270.320.330.320.310.180.21
GSMnone0.340.320.340.310.400.440.400.400.140.20
pwe0.220.210.220.190.220.220.220.200.150.19
pwe/fine-tuning0.310.310.320.290.310.320.310.290.190.21
NSTMnone0.190.230.190.160.110.120.110.090.000.05
pwe0.140.130.150.110.130.130.130.100.010.06
pwe/fine-tuning0.110.100.110.060.090.070.090.040.000.05
NVDMnone0.340.320.340.310.460.490.460.460.060.12
pwe0.350.330.350.320.460.500.460.470.090.14
pwe/fine-tuning0.400.380.400.370.500.530.500.500.100.15
NVLDAnone0.250.240.260.220.260.260.260.250.110.15
pwe0.060.020.060.020.060.040.060.020.000.06
pwe/fine-tuning0.080.050.080.040.080.050.080.040.070.10
ProdLDAnone0.300.290.300.270.230.270.230.210.130.16
pwe0.050.010.050.010.050.010.050.010.000.05
pwe/fine-tuning0.090.060.090.050.090.060.080.050.080.11
RSBnone0.230.200.230.190.260.250.260.250.070.11
pwe0.100.050.100.040.100.030.100.030.080.10
pwe/fine-tuning0.170.130.170.130.170.140.170.130.100.11
WLDAnone0.070.060.070.040.070.050.070.040.000.05
pwe0.060.050.060.040.070.050.070.040.000.05
pwe/fine-tuning0.060.040.060.040.060.040.060.030.000.05
Table 30. Classification and Clustering performance on SearchSnippets.
Table 30. Classification and Clustering performance on SearchSnippets.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.510.540.480.490.780.800.760.780.080.33
pwe0.410.390.340.340.480.490.410.420.090.35
pwe/fine-tuning0.600.590.530.540.710.730.670.680.160.41
GSMnone0.500.550.450.470.780.810.750.770.080.33
pwe0.450.420.380.380.480.530.410.420.130.40
pwe/fine-tuning0.610.610.540.540.700.720.650.670.180.45
NSTMnone0.250.320.160.110.270.410.190.160.010.22
pwe0.290.430.200.160.260.590.170.130.020.23
pwe/fine-tuning0.240.290.150.100.250.320.170.130.010.22
NVDMnone0.440.490.410.430.800.830.770.790.030.25
pwe0.440.500.400.410.780.820.760.780.030.25
pwe/fine-tuning0.520.550.510.520.830.850.810.820.040.26
NVLDAnone0.370.360.310.310.400.420.330.340.050.28
pwe0.270.230.180.140.240.200.150.100.030.24
pwe/fine-tuning0.270.190.190.140.250.120.160.110.030.26
ProdLDAnone0.330.360.250.240.460.560.380.400.060.30
pwe0.230.110.140.080.220.110.130.060.010.22
pwe/fine-tuning0.270.170.180.130.250.140.160.110.030.25
RSBnone0.300.220.210.180.420.380.350.330.020.23
pwe0.220.090.130.070.220.050.130.060.010.22
pwe/fine-tuning0.320.280.240.210.380.350.300.300.030.25
WLDAnone0.220.160.140.080.220.160.140.080.000.22
pwe0.220.140.140.080.230.170.140.080.000.22
pwe/fine-tuning0.230.150.140.090.230.160.140.090.000.22
Table 31. Classification and Clustering performance on StackOverflow.
Table 31. Classification and Clustering performance on StackOverflow.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone0.710.680.700.680.720.720.710.710.160.23
pwe0.700.710.700.700.720.780.720.740.500.60
pwe/fine-tuning0.680.660.670.660.690.690.680.680.350.44
GSMnone0.710.690.700.690.720.710.710.710.170.24
pwe0.730.760.730.730.740.810.730.760.590.70
pwe/fine-tuning0.710.700.710.700.710.720.710.710.470.59
NSTMnone0.220.200.210.190.250.240.240.240.000.06
pwe0.450.420.440.410.470.450.460.450.000.06
pwe/fine-tuning0.210.190.200.180.220.220.220.220.000.06
NVDMnone0.740.720.740.720.760.760.750.760.130.19
pwe0.710.680.700.680.750.750.740.750.100.19
pwe/fine-tuning0.780.770.770.770.790.800.790.790.150.22
NVLDAnone0.500.470.490.460.480.470.470.460.190.30
pwe0.070.020.060.020.070.010.060.020.000.06
pwe/fine-tuning0.220.150.210.150.150.120.140.100.150.16
ProdLDAnone0.440.430.430.390.320.380.310.300.350.34
pwe0.060.010.060.010.060.020.060.010.000.06
pwe/fine-tuning0.290.250.280.230.180.170.170.140.210.19
RSBnone0.640.610.630.600.660.640.650.640.180.17
pwe0.100.040.100.040.100.040.090.030.050.08
pwe/fine-tuning0.660.630.650.630.670.660.660.650.290.22
WLDAnone0.090.070.090.050.090.080.090.050.000.06
pwe0.100.080.100.060.100.080.100.060.000.06
pwe/fine-tuning0.110.080.100.070.100.090.100.070.000.06
Table 32. Classification and Clustering performance on TrecTweet.
Table 32. Classification and Clustering performance on TrecTweet.
ModelClassifierSVM (Linear)SVM (rbm)-
MethodACCPrecisionRecallF1ACCPrecisionRecallF1NMIPurity
PWE
GSBnone--------0.500.39
pwe--------0.620.54
pwe/fine-tuning--------0.660.49
GSMnone--------0.520.40
pwe--------0.650.58
pwe/fine-tuning--------0.740.56
NSTMnone--------0.040.12
pwe--------0.030.13
pwe/fine-tuning--------0.000.10
NVDMnone--------0.390.30
pwe--------0.370.27
pwe/fine-tuning--------0.430.33
NVLDAnone--------0.440.36
pwe--------0.390.34
pwe/fine-tuning--------0.700.59
ProdLDAnone--------0.600.50
pwe--------0.160.18
pwe/fine-tuning--------0.620.49
RSBnone--------0.380.26
pwe--------0.170.18
pwe/fine-tuning--------0.410.28
WLDAnone--------0.000.10
pwe--------0.010.10
pwe/fine-tuning--------0.000.10
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Murakami, R.; Chakraborty, B. Investigating the Efficient Use of Word Embedding with Neural-Topic Models for Interpretable Topics from Short Texts. Sensors 2022, 22, 852. https://doi.org/10.3390/s22030852

AMA Style

Murakami R, Chakraborty B. Investigating the Efficient Use of Word Embedding with Neural-Topic Models for Interpretable Topics from Short Texts. Sensors. 2022; 22(3):852. https://doi.org/10.3390/s22030852

Chicago/Turabian Style

Murakami, Riki, and Basabi Chakraborty. 2022. "Investigating the Efficient Use of Word Embedding with Neural-Topic Models for Interpretable Topics from Short Texts" Sensors 22, no. 3: 852. https://doi.org/10.3390/s22030852

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop