Next Article in Journal
Ball Tracking Based on Multiscale Feature Enhancement and Cooperative Trajectory Matching
Next Article in Special Issue
Improving Performance of Massive Text Real-Time Classification for Document Confidentiality Management
Previous Article in Journal
Status Recognition of Marine Centrifugal Pumps Based on a Stacked Sparse Auto-Encoder
Previous Article in Special Issue
A Combined Semantic Dependency and Lexical Embedding RoBERTa Model for Grid Field Relational Extraction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Distillation and Pinyin Character Prediction for Chinese Spelling Correction Based on Multimodality

1
School of Information Science and Technology, North China University of Technology, Beijing 100144, China
2
CNONIX National Standard Application and Promotion Lab, Beijing 100144, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(4), 1375; https://doi.org/10.3390/app14041375
Submission received: 20 January 2024 / Revised: 4 February 2024 / Accepted: 6 February 2024 / Published: 7 February 2024
(This article belongs to the Special Issue Cross-Applications of Natural Language Processing and Text Mining)

Abstract

:
Chinese spelling correction (CSC) constitutes a pivotal and enduring goal in natural language processing, serving as a foundational element for various language-related tasks by detecting and rectifying spelling errors in textual content. Numerous methods for Chinese spelling correction leverage multimodal information, including character, character sound, and character shape, to establish connections between incorrect and correct characters. Research indicates that a majority of spelling errors stem from pinyin similarity, with character similarity accounting for half of the errors. Consequently, effectively modeling character pinyin and character relationships emerges as a key challenge in the CSC task. In this study, we propose enhancing the CSC task by introducing the pinyin character prediction task. We employ an adaptive weighting method in the pinyin character prediction task to address predictions in a more granular manner, achieving a balance between the two prediction tasks. The proposed model, SPMSpell, utilizes ChineseBERT as an encoder to capture multimodal feature information simultaneously. It incorporates three parallel decoders for character prediction, pinyin prediction, and self-distillation modules. To mitigate potential overfitting concerning pinyin, a self-distillation method is introduced to prioritize character information in predictions. Extensive experiments conducted on three SIGHAN benchmark tests showcase that the model introduced in this paper attains a superior level of performance. This substantiates the correctness and superiority of the adaptive weighted pinyin character prediction task and underscores the effectiveness of the self-distillation module.

1. Introduction

Chinese spelling correction (CSC) is dedicated to identifying and rectifying spelling errors in text, playing a pivotal role in natural language processing (NLP) with far-reaching implications in downstream applications like search engines [1], OCR [2], and ASR. In contrast to languages like English, Chinese spelling correction poses a more formidable challenge owing to the intricacies of Chinese spelling rules. The Chinese language is characterized by a profusion of homophones, morphologically similar characters, and intricate phonetics, creating a fertile ground for spelling errors. Additionally, the complexity is amplified as Chinese spelling correction must consider contextual nuances, word-matching intricacies, and other linguistic features, further elevating the difficulty of error correction. Spelling errors in Chinese typically fall into two primary categories: pinyin errors and glyph errors, arising from the misuse of characters with similar pinyin pronunciations or visual resemblances, respectively. Figure 1 illustrates instances of these errors, such as the confusion between the pinyin-similar characters “稍” (meaning “little”) and “烧” (meaning “burn”), or the misinterpretation of visually similar shapes “人” (meaning “human”) and “入” (meaning “enter”). The origins of spelling errors predominantly stem from human writing mistakes and machine recognition errors, particularly those induced by Automatic Speech Recognition (ASR) and Optical Character Recognition (OCR) systems [3].
In recent years, leveraging pre-trained language models (PLM) has emerged as a dominant approach for Chinese spelling correction tasks. Prominent examples include FASpell [4], Softmasked-BERT [5], SpellGCN [6], and PLOME [7]. Concurrently, certain researchers have directed their focus towards the phonological and glyph features of Chinese characters, aiming to enhance the model’s error correction capabilities by amalgamating phonetic and visual information [7,8]. Previous studies underscore the significance of pinyin and glyph similarities, revealing that 83% of spelling errors are attributed to pinyin similarity and 48% to glyph similarity [9]. Effectively modeling Chinese pronunciation in the context of the CSC task remains a pivotal challenge, and nearly all contemporary state-of-the-art CSC methods explicitly or implicitly incorporate Chinese pronunciation.
Explicit methods delve into the Chinese pronunciation of entire characters, such as encoding the pinyin of a Chinese character and integrating it into the character representation using specific strategies [10]. Alternatively, Chinese pronunciation prediction models the relationship between similar characters in pinyin [7]. Implicit methods primarily focus on pinyin glyph similarity between Chinese character pairs, either by increasing the decoding probability of characters with akin pronunciations [6] or by incorporating pinyin glyph similarity into the encoding process via graph convolutional networks (GCN) [6]. Despite the notable performance gains achieved, these approaches confront two potential challenges: during training, pinyin information may either be overlooked or overshadowed by textual information. For instance, a specific BERT model that exclusively considers pinyin sequences, excluding Chinese characters, can still detect and rectify erroneous characters, while models like REALISE [8] encode and merge textual and pinyin information through a gating mechanism that overlooks an error. Second, the introduction of pinyin features may compromise the representation of normal text. Using Figure 2 as an illustration, a standard BERT model can correct the misspelled character “的” in the input, whereas REALISE fails to do so. This issue is attributed to REALISE’s tendency to over-rely or over-fit on pinyin information.
Currently, only a fraction of the available multimodal information is harnessed in the existing Chinese spelling correction (CSC) methods, highlighting the underutilization of the inherent value in Chinese character multimodality. This underscores the unexplored potential for enhancement within the field of multimodality for the CSC task. Building upon this realization, we present a novel approach named SPMSpell (Self-Distillation and Character Pinyin Prediction Based on Multimodality). SPMSpell introduces a finely tuned pinyin prediction task, an adaptive weighting mechanism, and a self-distillation module, aiming to refine the performance of the CSC task.
Concretely, when provided with a sentence containing spelling errors, ChineseBERT [11] serves as the encoder backbone to fuse three pivotal feature dimensions: semantic, phonological, and graphemic. Three distinct decoders are then constructed based on this fusion: one for accurate character prediction, another for predicting the consonant-rhyme of each target character (i.e., pinyin prediction), and the last one devoted to the self-distillation module, ensuring the model’s prediction consistency when presented with the original text input. Throughout the training phase, the pinyin prediction task serves as an auxiliary task, with its predictions discarded during inference.
The primary contributions of this paper can be succinctly summarized as follows: (1) Introduction of the pinyin prediction task to enhance the efficacy of the CSC task, adopting a fine-grained prediction approach for pinyin. (2) Proposition of a self-distillation module to mitigate the risk of the model overly fixating on pinyin features. (3) Introduction of an adaptive weighting method to harmonize subtasks within the pinyin prediction task, encompassing consonant, rhyme, and tone predictions while simultaneously balancing character and pinyin prediction tasks. (4) Achievement of an elevated performance benchmark across three prominent CSC datasets.

2. Related Work

2.1. Chinese Spelling Correction

Chinese spelling correction is a foundational task within the realm of natural language processing. Simultaneously, the inherent complexities of the Chinese language, such as polyphonic characters and morphologically similar characters, render this task highly challenging. The increasing interest among natural language processing researchers underscores the significance of Chinese spelling correction. Presently, neural network-based models, particularly pre-trained language models, dominate this domain and can be broadly classified into two research directions.
One line of research centers on enhancing the semantic modeling of text features [4,12,13]. This approach treats Chinese spelling correction as a sequence annotation task, utilizing pre-trained language models to derive contextual representations. For instance, Soft-Masked BERT [5] integrates a detection network to predict the correctness of individual characters. It subsequently generates soft-masked embeddings for the correction network to rectify errors. MDCSpell [14] introduces a multitasking detector-correction framework, merging the representations of detection and correction networks.
Another avenue of research involves incorporating phonological information into the correction process. This stems from the observation that homophones contribute significantly to usage errors [9]. Models like MLM-phonetics [15] and PLOME [7] adopt a word replacement strategy during pre-training, substituting randomly selected characters with phonetically or visually similar ones. REALISE [8] and PHMOSpell [3] employ multiple encoders to capture textual, phonological, and visual features, utilizing a selective gating mechanism for fusion. SCOPE [16] introduces an assisted pronunciation prediction task and devises an iterative inference strategy to enhance performance. Nonetheless, these methods often amalgamate textual and phonological features without facilitating direct and deep interaction between them, potentially resulting in the underutilization of pinyin information.

2.2. Multimodal Learning

Numerous studies have explored the integration of information from diverse modalities to enhance overall performance. Notable advancements include multimodal sentiment analysis [17,18], visual quizzing [19,20], and multimodal machine translation [21,22]. Recent developments involve the introduction of multimodal pre-training models, such as VL-BERT [23], Unicoder-VL [24], and LXMERT [25]. To incorporate visual information from Chinese characters into language models, Meng et al. [26] innovatively designed a Tianzige-CNN to enhance various NLP tasks, including named entity recognition and sentence classification.
Subsequently, researchers have endeavored to embed multimodal character information into error correction models to facilitate more accurate corrections. Xu et al. [8] meticulously model the semantic, phonetic, and visual information of input characters, introducing a gating mechanism to selectively blend information from these modalities for the final correction prediction. Guo et al. [12] adopt a strategy of pre-training BERT with artificially constructed obfuscated sets containing phonological and graphemic features of similar characters. This approach enables BERT to fully leverage character speech and visual features for error correction.
Wang et al. [10] enhance error correction by utilizing a pinyin-enhanced candidate based on character pronunciation features. This method, combined with an attentional mechanism to model neighboring token dependencies, contributes to more accurate predictions. Zhang et al. [15] integrate a pre-trained detection module and an error correction module based on speech features, effectively predicting the final correction. Guo et al. [12] combine a pre-trained set of character speech features with an error correction module for final corrections. Zhang et al. [15] undertake joint fine-tuning error correction based on pre-trained detection and error correction modules with phonological features.
Liu et al. [7] employ a GRU to extract phonological and visual features of characters, predicting the pronunciation of the target character in a coarse-grained, non-adaptive manner. Li et al. [13] introduce a speech prediction assistance task, combining fine-grained phonological features of characters to achieve adaptive weighting. Wei et al. [27] establish two new pre-training targets for the error corrector, capturing the phonetic and shape information of characters. These features are later fused with semantic information to achieve effective error correction.

2.3. Self-Distillation

Knowledge distillation [28] functions as a methodology to extract concise student models from their larger teacher counterparts. As a specialized distillation strategy, deep mutual learning [29] facilitates collaborative knowledge acquisition among various student models, enabling them to guide and refine each other throughout the training process. Specifically, when these student models share identical parameters, this methodology is termed self-distillation [30].
The implementation of self-distillation in the context of Chinese spelling correction (CSC) has resulted in significant performance enhancements. SDCL [31] autonomously encodes both the original sentence and its correct counterpart, employing contrast loss to enhance contextual representations. On the other hand, CRASpell [32] generates a noisy sample for each input, applying KL dispersion to both outputs to augment the performance in handling multiple misspelled sentences. In this study, the self-distillation module is intricately designed to alleviate the risk of overfitting associated with pinyin information during the model training phase.

3. Methodology

In this section, we present the SPMSpell model designed for the Chinese spelling correction task. We begin by providing a clear problem definition and subsequently delve into the intricate details of the proposed model’s implementation.

3.1. Problem Definition

The Chinese Spell-Checking (CSC) task is to detect and correct spelling errors in Chinese text. Given a source sentence X = { x 1 , x 2 , , x n } , with n characters containing spelling errors, the CSC model takes X as input, detects the errors and corrects them at the character level, and outputs the correct target sentence Y = { y 1 , y 2 , , y n } . The lengths of X and Y are equal, and thus, the task can be regarded as a sequence labeling task, i.e., modeling the p ( Y | X ) probability. We further give the fine-grained pinyin of each character y i in the correct sentence Y, denoted as a ternary of the form ( α i , β i , γ i ) , where α i , β i and γ i denote the vowel, rhyme, and tone, respectively. Please note that this transformation of output utterances is required and is only provided during training. Usually, there are no or only a small number of misspelled characters in the sentence. Most of the characters should be copied, and the misspelled character x i X bears some similarity to its correct character y i Y .

3.1.1. Architecture

The fundamental concept driving SPMSpell is to employ self-distillation, fine-grained pinyin character prediction tasks, and an adaptive task weighting mechanism to augment the effectiveness of the Chinese spelling correction (CSC) task. To realize this objective, SPMSpell is constructed around a shared encoder featuring three parallel decoders. The first is primarily dedicated to character prediction, addressing the core CSC task. The second decoder focuses on the fine-grained pinyin prediction task, while the third serves the self-distillation module. The equilibrium between the two prediction tasks, as well as within the pinyin prediction task itself, is dynamically determined based on the pinyin character and vowel-rhyme-tone similarities discerned between the input and target sentences. Figure 3 provides a comprehensive overview of the overall architecture of SPMSpell.

3.1.2. Encoder

Chinese is hieroglyphic, and character shapes and character sound features contain important information. Similar to the recent CSC methods utilizing multimodal information [7,8], we use ChineseBERT [11] as an encoder, as shown in Figure 4 ChineseBERT is a pre-trained language model that incorporates both Chinese pinyin and glyph information in a pre-trained language model to extract semantic, phonological and morphological features for the CSC task. Since the multimodal information of characters is captured and fused by the same model without adding additional networks, the problem of ambiguity between different feature information can be largely mitigated; furthermore, using only ChineseBERT as a multimodal encoder further simplifies the architectural complexity of the overall model of the CSC; on the other hand, in terms of the degree of pinyin feature extraction, the model uses a fine-grained level, i.e., the consonants, rhymes and tones of characters are extracted as graphemic features instead of capturing the whole pinyin as in previous work. Fine-grained pronunciation feature capture allows for better modeling of the phonetic similarity between the target character and the error character. Specifically, for each character X i in the input sentence X, the encoder first generates all its character embeddings, pinyin embeddings, and grapheme embeddings with embedding size D. These three embeddings are then concatenated in series and mapped to the D-dimensional fusion embeddings via a fully connected layer. Afterward, as in most other pre-trained language models, the fusion embeddings are added to the positional embeddings and fed into the Transformer’s stack to generate a contextual representation h i R D of the input character x i . We denote the representation of the character after this encoding process as H = { h 1 , h 2 , h n } .

3.1.3. Character Prediction Decoder

This decoder is used to predict the character in the correct sentence Y based on the coded output H. Specifically, given each input character x i , we first project its coded output h i into the character-specific feature space (GeLu is a Gaussian error linear unit activation function, and its derivative is continuous, which makes it easier to propagate the gradient when training a deep neural network, avoiding the problem of discontinuity of the derivative at zero, thus reducing the problem of gradient vanishing during training):
h i c = G e L u ( W c h i + b c )
The corresponding correct character y ^ i is then predicted based on the projected output Softmax is an activation function that normalizes a numerical vector into a probability distribution vector, where the sum of all probabilities is equal to 1. It is commonly employed as the final layer in neural networks, particularly for the output in multi-class classification problems. The Softmax layer is often used in conjunction with the cross-entropy loss function:
p ( y ^ i | X ) = s o f t m a x ( W y h i c + b y )
where W c R D × D , b c R D is the learnable parameter of the character-specific feature projection layer; W y R V × D , b y R V is the learnable parameter of the character prediction layer; and V is the vocabulary size.

3.1.4. Pinyin Prediction Decoder

This decoder is used to predict fine-grained predictions, i.e., to determine the vowel, rhyme, and tone of each character in the correct sentence Y, based on the coded output H. Similarly, given each input character x i and its encoded output h i , we project h i into the feature space specific to pronunciation:
h i p = G e L u ( W p h i + b p )
Based on the projected output, we predict the consonant α ^ i , rhyme β ^ i , and tone γ ^ i of the corresponding correct character:
p ( α ^ i X ) = s o f t m a x ( W α h i p + b α )
p ( β ^ i X ) = s o f t m a x ( W β h i p + b β )
p ( γ ^ i X ) = s o f t m a x ( W γ h i p + b γ )
where W p R D × D , b p R D is the learnable parameter of the articulation-specific feature projection layer; W τ R U × D , b τ R U , where τ { α , β , γ } is the learnable parameter of the articulation prediction layer; and U is the total number of articulatory units (consonants, rhymes, and tones).

3.1.5. Self-Distillation Decoder

The decoder is used for the self-distillation module, i.e., based on the encoded output H, which is generated by taking the last layer of hidden states of the encoder. Then, h i is projected into a pronunciation-specific feature space, and the probability distribution s i for the i character is computed based on h i :
p ( s i X ) = s o f t m a x ( W s h i c + b s )
After obtaining the output distribution for each character, the model performs another forward pass with the original sequence X as input, generating for each character x i another output distribution q i R V . The two sets of distributions are then forced to be close together by applying a bidirectional KL divergence:
L k l = 1 n i = 1 n 1 2 ( D k l ( s i | | q i ) + D k l ( q i | | s i ) )

3.1.6. Adaptive Weighting

We design an adaptive weighting method to balance the character prediction and pinyin prediction tasks during training, as well as the balance of consonant, rhyme, and tone prediction within pinyin prediction. Given an input sentence X, the character prediction task aims to match the predicted character { y ^ i } i = 1 n with the truth value { y i } i = 1 n , while the pinyin prediction task aims to match the predicted fine-grained character { ( α ^ i , β ^ i , γ ^ i ) } i = 1 n with the ground truth value { ( α i , β i , γ i ) } i = 1 n match. Where the character prediction task loss function is defined as:
L i c = l o g p ( y ^ i = y i X )
Then, as we discussed in the previous introduction, the phonetic prediction task may provide different levels of benefit with different input characters. The more similar the pronunciation of the input and target characters, the more likely it is that spelling errors caused by phonetic similarity will occur. In this case, the pinyin prediction task may provide greater benefits and should be assigned greater weights. To compute such adaptive weights, we feed the target correct sentence Y to the encoder and subsequent pronunciation-specific projection layers. Then, we compute the cosine similarity c o s ( h x i p , h y i p ) for each input character x i and its target character y i based on their pronunciation-specific feature representations h x i p , h y i p , and accordingly define the adaptive weights at the i position as:
w i p = s i n ( ( c o s ( h x i p , h y i p ) + 1 2 ) 2 )
The higher the cosine similarity, c o s ( h x i p , h y i p ) , the greater the weight w i will be. Similarly, for the pinyin prediction task internally, consonants, rhymes, and tones provide different levels of benefit for different input characters. However, empirically, tones should be weaker than the previous two in comparison to consonants and rhymes, so their weight values should be smaller than the previous two when weighting. Therefore, the fine-grained pinyin prediction task loss function is defined as follows:
L i p = k 1 w α l o g p ( α ^ i = α i | X ) k 2 w β l o g p ( β ^ i = β i | X ) k 3 w γ l o g p ( γ ^ i = γ i | X )
where, w α , w β , w γ re the weights of consonants, rhymes, and tones computed based on cosine similarity, and k 1 , k 2 , k 3 hyperparameters. Finally, the total loss function is defined as the sum of the character prediction loss and the pinyin prediction loss, as well as the self-distillation loss:
L = 1 n i = 1 n ( w i c L i c + w i p L i p + L k l )
where w i c is the character prediction, and the target correct sentence Y is computed based on the cosine similarity. Two points are worth noting here: (1) The branch that encodes and maps the target sentence Y is used to compute the adaptive weights only in the forward passes, while it will be separated in the backward passes. (2) The pinyin prediction, the self-distillation module, and the adaptive weighting scheme are introduced only during training. During inference, we use the character prediction decoder alone for prediction.

3.2. Inference

Similar to the models proposed by Liu et al. [32] and Devlin et al. [33], addressing the challenges of error correction in multi-error texts remains a substantial hurdle, leading to suboptimal effects. In an effort to overcome these challenges and mitigate the tendency toward excessive error correction, this paper introduces a novel polling error correction strategy during the inference phase.
Specifically, during inference, each input sentence undergoes an iterative process of character detection and correction, deviating from the conventional all-at-once correction approach. In each iteration, only erroneous characters within a specified window size around each correction position from the preceding iteration are eligible for correction. If a position undergoes modification in every iteration, it is reverted to its original character without any correction. Drawing from experimental insights, the paper opts for two iterations, with a window size set to 5.
This iterative correction process involves counting a specified number of low-probability words as the targets for the subsequent error correction after each prediction. The corrected information from the previous round is then incorporated into the next round, enriching contextual information and facilitating more nuanced error correction. This iterative strategy proves conducive to elevating the overall quality of error correction.

4. Experiments

In this section, we provide a comprehensive overview of the experiments conducted to assess the efficacy of the proposed models.

4.1. Datasets and Metrics

As with the previous work [6,7,8], our training data consists of two parts. One part is manually annotated training samples from SIGHAN 13 [34], SIGHAN 14 [35], and SIGHAN 15 [36]. The other part comprises 271k training samples generated by Wang et al. [37] using methods based on Automatic Speech Recognition (ASR) and Optical Character Recognition (OCR). We use the test sets of SIGHAN 13, SIGHAN 14, and SIGHAN 15 for evaluation. We follow previous work [6,8] to convert them to Simplified Chinese using OpenCC. We further used pypinyin (pypinyin is a very popular Python library for converting Chinese characters to pinyin. It can output the corresponding pinyin according to the Chinese characters and supports polyphonic words and tone selection. Pinyin link address: https://pypi.org/project/pypinyin, 15 November 2023) to obtain the pronunciation of each character and segment them into consonants, rhymes, and tones using a predefined vocabulary of consonants and rhymes provided by Xu et al. [8].
We use the metrics of sentence-level precision, recall, and F1 to evaluate our model for detection and correction. Sentence-level metrics are more stringent than character-level metrics because a sentence is considered correct when and only when all errors in the sentence are successfully detected and corrected. Errors were reported on the Detect and Correct subtask. In addition to the game-level evaluation, we also considered the character-level evaluation and the official SIGHAN evaluation.

4.2. Baselines

We compare SPMSpell to the following baseline methods. All these methods use character speech information in some way and represent the current state of the art on the SIGHAN baseline.
  • GAD [12]: This method learns the global relationship between potentially correct input characters and candidates for potentially incorrect characters.
  • DCN [10]: Through a unique dynamic connection network, K n paths (K denotes the number of candidate words, n denotes the length of the sentence) are generated in the output stage of the model, and then an optimal path is selected by scoring through the dynamic connection network; an attentional mechanism is introduced to model the dependency relationship between neighboring characters.
  • REALISE [8]: The method predicts the output by encoding phonological and graphemic information based on semantic information and finally introduces a gating mechanism to selectively fuse semantic, phonological, and graphemic information.
  • uChecker [38]: A masked pre-trained language model is proposed as an unsupervised Chinese spell checker. The method uses the pre-trained model to learn contextual information and uses a masked language modeling task to predict misspelled words for spell-checking.
  • MDCSpell [14]: The method designs a multitasking framework with BERT as a corrector that captures visual and phonetic features of characters and integrates the hidden state of the detector to minimize the impact of errors.
  • ECOPE [13]: This thesis proposes a method for Chinese spell correction using error-driven comparison probability optimization to improve the accuracy of future spell-checking by learning the comparison between past spelling errors and correct spellings.
  • UMRSpell [39]: It aims to unify the detection and correction parts of the pre-trained model to achieve Chinese missing, redundancy, and spelling correction. The method utilizes the contextual information of the pre-trained model and the Transformer structure to achieve spelling detection and correction by means of joint learning.

4.3. Experimental Parameter Setting

We use the pre-trained ChineseBERT as the encoder. In the specific training process, the dimension of the hidden layer feature vectors was set to 768 based on the model’s representational capacity and computational efficiency. The learning rate, determining the magnitude of weight adjustments during each update, was set to 5 ×   10 5 with linear decay, as determined by experimental considerations. Dropout was generally set to 0.1. The Batch Size, determining the number of samples used for each weight update, was set to 32 based on experimental conditions and device capabilities. The number of epochs set to 30 based on fitting results during the experimental process represented the training iterations. The AdamW optimizer was employed and widely recognized for its effectiveness in various deep-learning tasks. All experiments are conducted on one GeForce RTX 3090. (The GeForce RTX 3090 is manufactured by NVIDIA from Santa Clara, CA, USA).

4.4. Overall Results

The comprehensive results presented in Table 1 illustrate the evaluation outcomes of the proposed SPMSpell model alongside seven baseline models. The F1 scores for sentence-level detection and correction, as well as character-level detection and correction, are reported on the SIGHAN 13/14/15 test datasets, with bold font highlighting the best-performing results. The results showcase the superiority of the SPMSpell model across all test sets, both at the sentence and character levels. Particularly noteworthy is the exceptional performance on the SIGHAN 15 test set, where SPMSpell outperforms all baseline models, affirming the model’s efficacy. When compared to models incorporating speech and visual features, such as REALISE, SPMSpell achieves a notable improvement in detection/correction F1 by 0.9%, 0.9%, 4.3%, and 2.9% on SIGHAN 14/15, respectively. This underscores the effectiveness of ChineseBERT-based encoders in mitigating the mismatch phenomenon arising from the fusion of multiple feature information.
It is noteworthy that the observed improvement in recall and F1, while substantial, is accompanied by a disparity in accuracy compared to other baseline models. This discrepancy may stem from the limitation of the training data volume. To address this, the introduction of a new public dataset could augment the training volume and potentially bridge the existing gap. Nevertheless, it is essential to highlight that the model’s performance remains competitive, particularly in comparison to other methods incorporating speech information for this task.

4.5. Ablation Study

In this subsection, we explore the impact of the hyperparameters k 1 , k 2 , k 3 in the loss function on the model performance as well as the contribution of the pinyin prediction task, the self-distillation module, and the adaptive weighting method to the SPMSpell model. For this ablation experiment, we evaluate the model using the SIGHAN 2015 test set.
According to what was mentioned earlier, in the pinyin prediction task, the gap between the roles of tones and rhymes is small and theoretically larger than the roles provided by tones, so for exploring the effects of hyperparameters k 1 , k 2 , k 3 can be transformed into exploring the effects of hyperparameter k. The transformation is as follows:
k 1 , k 2 , k 3 ( 1 k ) 2 , ( 1 k ) 2 , k
The effect of hyperparameter k taking value on F1 value of model performance is shown in Figure 5. Based on the experimental results, we can find that the F1 scores are basically increasing with increasing values of k. However, after the value exceeds 0.3, the F1 scores begin to decrease. Therefore, setting k to 0.2 achieves the overall best corrected F1 score. We further explored the reasons for this and found that in Mandarin, pinyin is the official system used for phonetic transcription. It uses the three components of consonants, rhymes, and tones to express the pronunciation and spelling of Chinese characters. Since the similarity of pronunciation of Chinese characters is mainly determined by their consonants or rhymes rather than their tones, its proportion of tones should be relatively low, and its proportion of consonants and rhymes high, i.e., the value of k should be taken to be relatively low.
To further validate the effect of analyzing several constituent modules, this paper also performs ablation experiments on the SIGHAN 2015 test set in the following settings to explore the contribution of each module: (a) w/o PP refers to the removal of the pinyin prediction task from the SPMSpell model; (b) w/o SD refers to the removal of the self-distillation module from the SPMSpell model; and (c) w/o AW refers to the removal of the adaptive weighting method from the SPMSpell model.
The experimental results, as presented in Table 2, showcase precision (P), recall (R), and F1 scores for sentence-level detection/correction, along with the F1 scores for character-level detection/correction. The primary aim of this study is to leverage a multimodal language model to unveil the intricate relationships between characters, therefore addressing the challenge of uneven fusion of multimodal feature information. Notably, upon excluding the pinyin prediction task, there is a discernible decrease in detection/correction F1 scores by 2.6% and 1.9%, respectively. This underscores the effectiveness of the pinyin prediction task in capturing the nuances of Chinese character pronunciation. Similarly, upon removal of the self-distillation module, there is a reduction in detection/correction F1 scores by 2.4% and 2.2%, respectively, affirming the efficacy of the introduced self-distillation module in mitigating the risk of overfitting to pinyin features.
To further verify the effectiveness of different encoders, this study conducted ablation experiments with the following configurations on the SIGHAN13 test set to explore encoder selection: (a) SPMSpell (BERT) refers to using BERT as the encoder in the SPMSpell model; (b) SPMSpell (REALISE) refers to using the REALISE model as the encoder in the SPMSpell model; (c) SPMSpell (ChineseBERT) refers to using ChineseBERT as the encoder in the SPMSpell model, which is the proposed SPMSpell model in this paper.
The experimental results, as shown in Table 3, display the precision (P), recall (R), and F1 score of sentence-level detection/correction. Overall, compared to using BERT as the encoder, the REALISE and ChineseBERT models, possibly due to incorporating character shape and pinyin information into language model pre-training, achieve better performance. With the assistance of character shape and pinyin information, better performance is attained. When comparing REALISE with the ChineseBERT model, using ChineseBERT as the encoder yields superior results. This may be because the REALISE model is multimodal, emphasizing the integration of text, sound, and visual information, while ChineseBERT focuses on the features of the Chinese language, particularly the fusion of character shape and pinyin information. It can simultaneously capture semantic, character shape, and pronunciation features without encountering the issue of re-fusing different features obtained by different models, effectively addressing the problem of uneven feature information. On the other hand, using ChineseBERT alone as the multimodal encoder backbone further simplifies the overall architecture complexity of the CSC model.
In essence, the performance of SPMSpell experiences a decline when any individual component is removed. This provides compelling evidence for the effectiveness of each component within our approach, emphasizing the integral role played by each element in contributing to the overall efficacy of SPMSpell.

5. Conclusions

The paper presents SPMSpell, a Chinese spelling error correction model grounded in multimodal features. Although prior studies have emphasized the importance of character, character sound, and character shape information, our approach innovatively employs the multimodal language model ChineseBERT as an encoder. This enables the simultaneous capture of character, character sound, and character shape features. Feature fusion is then applied to model character and character sound effectively, addressing issues related to multimodal feature information mismatch. This not only reduces model complexity but also mitigates training costs. For the character prediction task, fine-grained pinyin prediction task, and the self-distillation module, we utilize three parallel decoders. To tackle potential overfitting in pinyin prediction, we introduce the self-distillation method, ensuring a predominant role for character information in predictions. Adaptive weighting methods are incorporated to balance the model between character prediction and pinyin prediction tasks, as well as among their respective subtasks. Experimental results on the SIGHAN public dataset demonstrate SPMSpell’s superior performance compared to other comparative models. Detailed analyses and studies reveal SPMSpell’s excellence in leveraging pinyin features within multimodal information, showcasing robust generalization capabilities in the realm of Chinese spelling error correction. Looking ahead, our contributions lie in advancing the state of the art by effectively integrating multimodal features for improved spelling error correction. Future work will explore further optimizations, new multimodal architectures, and incorporating large language models to enhance the overall performance and applicability of the model.

Author Contributions

Conceptualization, L.H., F.L., J.L., J.D. and H.W.; writing—original draft preparation, L.H. and F.L.; writing—review and editing, L.H.; data curation, F.L.; validation, J.L.; supervision, L.H., J.L., J.D. and H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Key Research and Development Program of China (2020AAA0109700), the National Natural Science Foundation of China (62076167), the National Natural Science Foundation of China (61972003), R&D Program of Beijing Municipal Education Commission (KM202210009002), and the Beijing Urban Governance Research Base of North China University of Technology (2023CSZL16).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

We would like to thank the anonymous reviewers for their helpful comments. We would like to thank the referees for their comments, which helped improve this paper considerably.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Martins, B.; Silva, M.J. Spelling Correction for Search Engine Queries. In Advances in Natural Language Processing, Proceedings of the International Conference on Natural Language Processing, EsTAL 2004, Alicante, Spain, 20–22 October 2004; Springer: Berlin/Heidelberg, Germany, 2004. [Google Scholar]
  2. Afli, H.; Qiu, Z.; Way, A.; Sheridan, P. Using SMT for OCR Error Correction of Historical Texts; Language Resources and Evaluation; European Language Resources Association (ELRA): Paris, France, 2016. [Google Scholar]
  3. Huang, L.; Li, J.; Jiang, W.; Zhang, Z.; Chen, M.; Wang, S.; Xiao, J. PHMOSpell: Phonological and Morphological Knowledge Guided Chinese Spelling Check. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, 1–6 August 2021; pp. 5958–5967. [Google Scholar]
  4. Hong, Y.; Yu, X.; He, N.; Liu, N.; Liu, J. FASPell: A Fast, Adaptable, Simple, Powerful Chinese Spell Checker Based on DAE-Decoder Paradigm. In Empirical Methods in Natural Language Processing; Association for Computational Linguistics: Toronto, ON, Canada, 2019. [Google Scholar]
  5. Zhang, S.; Huang, H.; Liu, J.; Li, H. Spelling Error Correction with Soft-Masked BERT. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020. [Google Scholar]
  6. Cheng, X.; Xu, W.; Chen, K.; Jiang, S.; Wang, F.; Wang, T.; Chu, W.; Qi, Y. SpellGCN: Incorporating Phonological and Visual Similarities into Language Models for Chinese Spelling Check. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020. [Google Scholar]
  7. Liu, S.; Yang, T.; Yue, T.; Zhang, F.; Wang, D. PLOME: Pre-training with Misspelled Knowledge for Chinese Spelling Correction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Online, 1–6 August 2021; pp. 2991–3000. [Google Scholar]
  8. Xu, H.-D.; Li, Z.; Zhou, Q.; Li, C.; Wang, Z.; Cao, Y.; Huang, H.; Mao, X.-L. Read, Listen, and See: Leveraging Multimodal Information Helps Chinese Spell Checking. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021; Association for Computational Linguistics: Toronto, ON, Canada, 2021. [Google Scholar]
  9. Liu, C.-L.; Lai, M.-H.; Chuang, Y.-H.; Lee, C.-Y. Visually and Phonologically Similar Characters in Incorrect Simplified Chinese Words. In Coling 2010: Posters; Coling 2010 Organizing Committee: Beijing, China, 2010; pp. 739–747. [Google Scholar]
  10. Wang, B.; Che, W.; Wu, D.; Wang, S.; Hu, G.; Liu, T. Dynamic Connected Networks for Chinese Spelling Check. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021; Association for Computational Linguistics: Toronto, ON, Canada, 2021. [Google Scholar]
  11. Sun, Z.; Li, X.; Sun, X.; Meng, Y.; Ao, X.; He, Q.; Wu, F.; Li, J. ChineseBERT: Chinese Pretraining Enhanced by Glyph and Pinyin Information. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Online, 1–6 August 2021. [Google Scholar]
  12. Guo, Z.; Ni, Y.; Wang, K.; Zhu, W.; Xie, G. Global Attention Decoder for Chinese Spelling Error Correction. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021; Association for Computational Linguistics: Toronto, ON, Canada, 2021. [Google Scholar]
  13. Li, Y.; Zhou, Q.; Li, Y.; Li, Z.; Liu, R.; Sun, R.; Wang, Z.; Li, C.; Cao, Y.; Zheng, H.-T. The Past Mistake is the Future Wisdom: Error-driven Contrastive Probability Optimization for Chinese Spell Checking. In Findings of the Association for Computational Linguistics: ACL 2022; Association for Computational Linguistics: Toronto, ON, Canada, 2022. [Google Scholar]
  14. Zhu, C.; Ying, Z.; Zhang, B.; Mao, F. MDCSpell: A multi-task detector-corrector framework for Chinese spelling correction. In Findings of the Association for Computational Linguistics: ACL 2022; Association for Computational Linguistics: Toronto, ON, Canada, 2022; pp. 1244–1253. [Google Scholar]
  15. Zhang, R.; Pang, C.; Zhang, C.; Wang, S.; He, Z.; Sun, Y.; Wu, H.; Wang, H. Correcting Chinese spelling errors with phonetic pre-training. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021; Association for Computational Linguistics: Toronto, ON, Canada, 2021; pp. 2250–2261. [Google Scholar]
  16. Li, J.; Wang, Q.; Mao, Z.; Guo, J.; Yang, Y.; Zhang, Y. Improving Chinese Spelling Check by Character Pronunciation Prediction: The Effects of Adaptivity and Granularity. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 4275–4286. [Google Scholar]
  17. Zadeh, A.; Zellers, R.; Pincus, E.; Morency, L.P. Multimodal sentiment intensity analysis in videos: Facial gestures and verbal messages. IEEE Intell. Syst. 2016, 31, 82–88. [Google Scholar] [CrossRef]
  18. Zhang, D.; Li, S.; Zhu, Q.; Zhou, G. Effective Sentiment-relevant Word Selection for Multi-modal Sentiment Analysis in Spoken Language. In Proceedings of the 27th ACM International Conference on Multimedia, Nice, France, 21–25 October 2019. [Google Scholar]
  19. Agrawal, A.; Lu, J.; Antol, S.; Mitchell, M.; Zitnick, L.C.; Batra, D.; Parikh, D. VQA: Visual Question Answering. Int. J. Comput. Vision 2016, 123, 2425–2433. [Google Scholar] [CrossRef]
  20. Chao, W.L.; Hu, H.; Sha, F. Being Negative but Constructively: Lessons Learnt from Creating Better Visual Question Answering Datasets. In North American Chapter of the Association for Computational Linguistics; Association for Computational Linguistics: Toronto, ON, Canada, 2018. [Google Scholar]
  21. Hitschler, J.; Riezler, S. Multimodal Pivots for Image Caption Translation. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics, Berlin, Germany, 7–12 August 2016. [Google Scholar]
  22. Barrault, L.; Bougares, F.; Specia, L.; Lala, C.; Elliott, D.; Frank, S. Findings of the Third Shared Task on Multimodal Machine Translation. In Proceedings of the Third Conference on Machine Translation: Shared Task Papers, Belgium, Brussels, 27 July 2018. [Google Scholar]
  23. Su, W.; Zhu, X.; Cao, Y.; Li, B.; Lu, L.; Wei, F.; Dai, J. VL-BERT: Pre-training of Generic Visual-Linguistic Representations. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
  24. Li, G.; Duan, N.; Fang, Y.; Gong, M.; Jiang, D.; Zhou, M. Unicoder-VL: A Universal Encoder for Vision and Language by Cross-modal Pre-training. arXiv 2019, arXiv:1908.06066. [Google Scholar]
  25. Tan, H.; Bansal, M. LXMERT: Learning Cross-Modality Encoder Representations from Transformers. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, Hong Kong, China, 3–7 November 2019. [Google Scholar]
  26. Meng, Y.; Wu, W.; Wang, F.; Li, X.; Nie, P.; Yin, F.; Li, M.; Han, Q.; Sun, X.; Li, J. Glyce: Glyph-vectors for Chinese Character Representations. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  27. Wei, X.; Huang, J.; Yu, H.; Liu, Q. PTCSpell: Pre-trained Corrector Based on Character Shape and Pinyin for Chinese Spelling Correction. In Findings of the Association for Computational Linguistics: ACL 2023; Association for Computational Linguistics: Toronto, ON, Canada, 2023; pp. 6330–6343. [Google Scholar]
  28. Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. Comput. Sci. 2015, 14, 38–39. [Google Scholar]
  29. Zhang, Y.; Xiang, T.; Hospedales, T.M.; Lu, H. Deep mutual learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  30. Mobahi, H.; Farajtabar, M.; Bartlett, P.L. Self-Distillation Amplifies Regularization in Hilbert Space. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, BC, Canada, 6–12 December 2020. [Google Scholar]
  31. Zhang, X.; Yan, H.; Yu, S.; Qiu, X. Sdcl: Self-distillation contrastive learning for Chinese spell checking. arXiv 2022, arXiv:2210.17168. [Google Scholar]
  32. Liu, S.; Song, S.; Yue, T.; Yang, T.; Cai, H.; Yu, T.; Sun, S. CRASpell: A contextual typo robust approach to improve Chinese spelling correction. In Findings of the Association for Computational Linguistics: ACL 2022; Association for Computational Linguistics: Toronto, ON, Canada, 2022; pp. 3008–3018. [Google Scholar]
  33. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  34. Wu, S.H.; Liu, C.L.; Lee, L.H. Chinese Spelling Check Evaluation at SIGHAN Bake-off 2013. In Proceedings of the International Joint Conference on Natural Language Processing. Asian Federation of Natural Language Processing, Nagoya, Japan, 14–19 October 2013. [Google Scholar]
  35. Yu, L.-C.; Lee, L.-H.; Tseng, Y.-H.; Chen, H.H. Overview of SIGHAN 2014 Bake-off for Chinese Spelling Check. In Proceedings of the Cips-sighan Joint Conference on Chinese Language Processing, Wuhan, China, 20–21 October 2014. [Google Scholar] [CrossRef]
  36. Tseng, Y.-H.; Lee, L.-H.; Chang, L.-P.; Chen, H.H. Introduction to SIGHAN 2015 Bake-off for Chinese Spelling Check. In Proceedings of the 8th SIGHAN Workshop on Chinese Language Processing (SIGHAN’15), Beijing, China, 30–31 July 2015. [Google Scholar]
  37. Wang, D.; Song, Y.; Li, J.; Han, J.; Zhang, H. A hybrid approach to automatic corpus generation for chinese spelling check. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October–4 November 2018; pp. 2517–2527. [Google Scholar]
  38. Li, P. uChecker: Masked Pretrained Language Models as Unsupervised Chinese Spelling Checkers. In Proceedings of the 29th International Conference on Computational Linguistics, Gyeongju, Republic of Korea, 12–17 October 2022; pp. 2812–2822. [Google Scholar]
  39. He, Z.; Zhu, Y.; Wang, L.; Xu, L. UMRSpell: Unifying the Detection and Correction Parts of Pre-trained Models towards Chinese Missing, Redundant, and Spelling Correction. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Toronto, ON, Canada, 9–14 July 2023; pp. 10238–10250. [Google Scholar]
Figure 1. Examples of pinyin error and glyph error, where red color indicates the wrong character.
Figure 1. Examples of pinyin error and glyph error, where red color indicates the wrong character.
Applsci 14 01375 g001
Figure 2. Two examples of Chinese spelling error correction and the predictions of different models.
Figure 2. Two examples of Chinese spelling error correction and the predictions of different models.
Applsci 14 01375 g002
Figure 3. SPMSpell overall framework.
Figure 3. SPMSpell overall framework.
Applsci 14 01375 g003
Figure 4. ChineseBERT framework.
Figure 4. ChineseBERT framework.
Applsci 14 01375 g004
Figure 5. Effect of hyperparameter k on F1 value.
Figure 5. Effect of hyperparameter k on F1 value.
Applsci 14 01375 g005
Table 1. Experimental results of each model on test sets.
Table 1. Experimental results of each model on test sets.
DatasetModelDetection-LevelCorrection-LevelChar-Level
D-P D-R D-F C-P C-R C-F D-F C-F
SIGHAN13GAD85.779.582.584.978.781.587.693.5
DCN86.879.683.084.777.781.085.286.4
REALISE88.682.585.487.281.284.186.188.4
uChecker75.473.474.472.670.871.7--
MDCSpell89.178.383.487.576.881.8--
ECOPE87.281.784.486.180.683.3--
UMRSpell83.073.678.080.071.075.284.996.4
SPMSpell (our)87.783.785.686.982.884.692.195.3
SIGHAN14GAD66.671.869.165.070.167.582.987.6
DCN67.470.468.965.968.767.278.986.2
REALISE67.871.569.666.370.068.178.580.1
uChecker61.761.561.657.657.557.6--
MDCSpell70.268.869.569.067.768.3--
ECOPE65.869.067.463.766.965.3--
UMRSpell69.056.662.263.957.260.473.293.3
SPMSpell (our)68.673.570.567.071.269.081.589.0
SIGHAN15GAD75.680.477.973.277.875.488.290.1
DCN77.180.979.074.578.276.385.084.9
REALISE77.381.379.375.979.977.887.486.2
uChecker75.472.073.770.667.368.9--
MDCSpell80.880.680.778.478.278.3--
ECOPE78.282.380.276.680.478.4--
UMRSpell77.272.275.069.364.867.083.091.5
SPMSpell (our)81.785.683.679.483.481.388.392.8
Note: The test sets are SIGHAN13, SIGHAN14, and SIGHAN15. In the table results, Detection-level and Correction-level represent sentence-level metrics, Char-level represents character-level metrics, D-P means detection module precision, D-R means detection module recall, D-F means detection module F1 value, C-P, C-R, and C-F are correction module corresponding metrics, and black bold in the table is the optimal result.
Table 2. Results of model ablation experiments.
Table 2. Results of model ablation experiments.
DatasetModelDetection-LevelCorrection-LevelChar-Level
D-P D-R D-F C-P C-R C-F D-F C-F
SIGHAN15SPMSpell81.785.683.679.483.481.388.392.8
w/o PP79.082.581.077.480.679.487.090.9
w/o PP80.182.381.278.080.279.187.191.2
w/o PP80.484.382.377.580.579.587.190.3
Note: The dataset used in the ablation experiment was SIGHAN15. In the table results, Detection-level and Correction-level represent sentence-level metrics, Char-level represents character-level metrics, D-P means detection module precision, D-R means detection module recall, D-F means detection module F1 value, C-P, C-R, and C-F are correction module corresponding metrics.
Table 3. Results of encoder ablation experiments.
Table 3. Results of encoder ablation experiments.
DatasetModelDetection-LevelCorrection-Level
D-P D-R D-F C-P C-R C-F
SIGHAN13SPMSpell (BERT)85.077.080.883.075.278.9
SPMSpell (REALISE)87.682.585.487.281.284.1
SPMSpell (ChineseBERT)87.7 83.785.686.982.884.6
Note: The dataset used in the ablation experiment was SIGHAN13. In the table results, Detection-level and Correction-level represent sentence-level metrics, D-P means detection module precision, D-R means detection module recall, D-F means detection module F1 value, C-P, C-R, and C-F are correction module corresponding metrics, and black bold in the table is the optimal result.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, L.; Liu, F.; Liu, J.; Duan, J.; Wang, H. Self-Distillation and Pinyin Character Prediction for Chinese Spelling Correction Based on Multimodality. Appl. Sci. 2024, 14, 1375. https://doi.org/10.3390/app14041375

AMA Style

He L, Liu F, Liu J, Duan J, Wang H. Self-Distillation and Pinyin Character Prediction for Chinese Spelling Correction Based on Multimodality. Applied Sciences. 2024; 14(4):1375. https://doi.org/10.3390/app14041375

Chicago/Turabian Style

He, Li, Feng Liu, Jie Liu, Jianyong Duan, and Hao Wang. 2024. "Self-Distillation and Pinyin Character Prediction for Chinese Spelling Correction Based on Multimodality" Applied Sciences 14, no. 4: 1375. https://doi.org/10.3390/app14041375

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop