Next Article in Journal
An Empirical Evaluation of Document Embeddings and Similarity Metrics for Scientific Articles
Previous Article in Journal
A Framework for Learning and Rapid Implementation of Real-Time Global Illumination Methods
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using a Deep Quantum Neural Network to Enhance the Fidelity of Quantum Convolutional Codes

Department of Communication Science and Engineering, Fudan University, Shanghai 200433, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(11), 5662; https://doi.org/10.3390/app12115662
Submission received: 29 April 2022 / Revised: 29 May 2022 / Accepted: 30 May 2022 / Published: 2 June 2022
(This article belongs to the Topic Quantum Information and Quantum Computing)

Abstract

:
The fidelity of quantum states is an important concept in quantum information. Improving quantum fidelity is very important for both quantum communication and quantum computation. In this paper, we use a quantum neural network (QNN) to enhance the fidelity of [6, 2, 2] quantum convolutional codes. Towards the circuit of quantum convolutional codes, the target quantum state | 0 or | 1 is turned into entangled quantum states, which can defend against quantum noise more effectively. As the quantum neural network works better for quantum states with low dimension, we divide the quantum circuits into two parts. Then we apply the quantum neural network to each part of the circuit. The results of the simulation show that the network performs well in enhancing the fidelity of the quantum states. Through the quantum neural network, the fidelity of the first part is enhanced from 95.2 % to 99.99 % , and the fidelity of the second part is enhanced from 93.88 % to 94.57 % .

1. Introduction

Quantum error correction codes are used to preserve coherent states against noise. They can effectively control operational errors and decoherence. The 9-bit quantum encoding proposed by Shor in 1995 can correct single-bit phase flip error or bit flip error [1]. On this basis, Calderbank, Shor and Steane proposed a basic quantum block code coding framework called Calderbank–Shor–Steane (CSS) code in 1996, using the idea of classical linear block error correction codes, and designed quantum error correction codes using two special classical binary error correction codes [2,3]. In the same year, Gottesman proposed the stabilizer code, which is a group-theoretical structure and associated subclass of quantum codes [4]. It provides a general framework which can generate several types of quantum error correcting codes instead of specific codes.
The stabilizer formalism is very similar to the linear codes in classical error correction. Using the concept of classical error correction, it is very useful in quantum communication by preventing quantum data from encountering noise in a quantum channel. As it plays an important role in quantum error correction, it has been developed over the decades. Quantum convolutional code is a type of error correction code that combines the stabilizer formalism and the concept of traditional convolutional codes. Chau proposed a type of quantum convolutional code based on classical convolutional codes in 1998 [5]. Then, based on the concept of traditional convolutional codes [6], quantum convolutional codes were constructed using the CSS code framework and methods of the stabilizer codes. Ollivier and Tillich first described quantum convolutional codes in 2003 using stabilizer forms and providing an explicit encoding circuit [7]. Forney proposed a tail-biting quantum convolutional code, which has a higher rate and lower decoding complexity [8,9].
A good circuit should have low complexity and high reliability. Fidelity is an important factor to evaluate the reliability of the output state. If the fidelity of the states after the decoding circuit is low, the result is not credible. Recently, a quantum neural network was proposed to enhance the fidelity of quantum states.
In 1996, G. Toth et al., constructed a quantum cellular neural network based on quantum metacellular automata [10] using physical nearest-neighbor connections [11]. In 1998, Dr. T. Menneer first explored quantum neural networks from a multi-universe quantum theoretical viewpoint [12], and later proposed a model for constructing superposed multi-universe quantum neural networks based on this [13], and compared to various conventional neural networks, he found that the performance of quantum neural networks is superior. Subsequently, various models and algorithms such as quantum-derived neural networks [14], quantum associative storage [15], quantum dot molecular models [16], and entangled neural networks [17] were proposed. Recent studies by Verdon et al. show that quantum neural networks can learn the unknown unitary operator to process the target quantum state [18]. Based on the study in [18], Kerstin et al. proposed a quantum QNN neural network framework [19], using quantum feedforward neural networks with the trained unitary operators as quantum perceptions to assist quantum computation and enhance the fidelity of quantum states.
Consider the case where all quantum gates are noisy. To resist the effect of noise from quantum gate operators, we use quantum neural networks to optimize the input coded states. The effect of resisting quantum gate noise is achieved by learning the unknown unitary operator and applying it to the coded quantum state, thus improving the fidelity of the coded quantum state before entering the quantum noise channel. We provide a simulation to show the effect of applying a quantum neural network on quantum convolutional codes.

2. Basic Concepts

We hope to enhance the fidelity of quantum convolutional codes with a quantum neural network in order to ensure the reliabiliy of the code. In this section, we will introduce the concepts of quantum convolutional code and a quantum neural network.

2.1. Quantum Convolutional Codes

Quantum convolutional codes use the stabilizer formalism which was first proposed by Gottesman [20,21]. The error syndrome is defined by the Pauli matrix.
X = 0 1 1 0 , Y = 0 i i 0 , Z = 1 0 0 1
Pauli matrix X refers to bit flipping, Z refers to phase flipping and Y refers to both. The code subspace C of any stabilizer code is defined as the largest subspace stabilized by an Abelian group S acting on the N physical qubits of the code. In practice, S is a subgroup of the multiplicative Pauli group G N = sp { I , X , Y , Z } N [22]. The code subspace C can be defined by a set of independent generators M i of S:
i , | ψ = M i | ψ , | ψ C
An n , k , m stabilizer code is possible to encode k qubits of quantum information in n qubits and correct at most (m − 1)/2 errors. Similarly, the stabilizer group S for an n , k , m -convolutional code can be given by:
S = sp M j , i = I j × n M 0 , i , 1 i n k , 0 j
where M 0 , i G n + m . Above, the M j , i s are required to be independent and to commute with each other. The encoding circuit maps the to-be-protected qubits c j , i onto the code subspace, written as follows:
c 0 , 1 , , c q 1 , k i , j 1 + M j , i 2 r , s X ¯ s , r c s , r | 0 , , 0
for c s , r { 0 , 1 } , 0 < i n k , 0 j < q + λ , 0 s < q and 1 r k . The X ¯ i operator is the encoded Pauli operator X. This operation can be decomposed in two steps. The first one, r , s X ¯ s , r c s , r applies the different flip operators depending on the value of the to-be-protected qubits in the computational basis. The second projects this state onto the code subspace.
Take the [5, 1, 2] convolutional code, for example [23]. The generator matrix of the stabilizer group can be written as:
M = 0 1 1 0 0 0 0 1 1 0 0 0 0 1 1 D 0 0 0 1 1 0 0 1 0 0 1 0 0 1 D 0 1 0 0 0 D 0 1 0
From the generator matrix, it is easy to obtain the encoding circuit as Figure 1 [23].
To estimate the error at the receiver, the quantum Viterbi algorithm was introduced. The algorithm examines the syndromes block by block and updates a list of error candidates, among which one of them coincides with the most likely error. Through the algorithm, the most likely error can be selected and then be corrected.

2.2. Quantum Neural Network

After decades of development, quantum neural networks have an important role in quantum pattern recognition, overcoming quantum noise, and improving the fidelity of quantum states.
The QNN quantum neural network implements a true quantum analogue of classical neurons, which form quantum feedforward neural networks capable of general quantum computation. The effective training of these networks is described in neural network models using fidelity as a cost function, providing both classical and efficient quantum implementations. This approach allows for fast optimization with reduced memory requirements: the number of quantum bits required is scaled only according to the width, allowing for deep network optimization. The neural network has significant generalization behavior for learning unknown unitary operators and is highly robust to noisy training data with low-dimensional quantum states.
The minimal building block of QNN is a quantum perceptron, which is a quantum analogue of the perceptron used in classical machine learning. A quantum perceptron is an arbitrary unitary operator with m input qubits and n output qubits. These ( m + n ) arbitrary qubits depend on 2 m + n 2 1 parameters. The input qubits are initialized in the possibly unknown mixed state ρ ln , and the output qubits are in the baseline product state | 0 0 out . Take the case of perceptrons acting on m input qubits and one output qubit, i.e., there are ( m + 1 ) unitary operators. A QNN is a quantum circuit composed of quantum perceptrons that are organized into L hidden layers that act on the initial input state ρ ln and usually produce a mixed output state ρ out , with the following expression:
ρ out tr in , hid U ρ in | 0 0 hid , out 0 0 | U
where U U out U L U L 1 U 1 is the QNN circuit and U l is the overall unitary operator of the l-th hidden layer, consisting of the product of quantum perceptrons acting on the ( l 1 ) -th layer and the l-th layer. Since the quantum perceptrons are independent unitary operators, they do not commute with each other and need to be operated in order. The overall QNN structure is shown in Figure 2; from left to right are the input layer, hidden layer and output layer, respectively.
The quantum neural network has an input layer, an output layer, and L hidden layers, where perceptron units are applied layer by layer from top to bottom. The input quantum states are expanded in the hidden layer to fit a sufficient number of perceptron units, and then we reduce the dimension to keep them consistent with the number of input qubits before reaching the output layer. A key feature of a QNN is that the network output can be represented as a combination of a series of fully positive definite layer-to-layer transition mappings E l , with the following expression:
ρ out = E out E L E 2 E 1 ρ in
E l X l 1 tr l 1 j = m i 1 U j l X l 1 | 0 0 l 0 0 | j = 1 m l U j l
where U j l is the j-th perceptron acting on the ( l 1 ) -th and l-th layers, and m l is the total number of perceptrons acting on the ( l 1 ) -th and l-th layers. The information is propagated from the input to the output, thus naturally implementing the quantum feedforward neural network.
The QNN is used to train N pairs of data pairs ϕ x in , ϕ x out , x = 1 , 2 , , N . For a particular x, we train it with multiple samples to overcome the effect of quantum noise in the evaluation of the cost function. In addition, the number of samples required for each training round grows rapidly with the number of neurons, which is linearly related to the number of network parameters, making it difficult to handle quantum states with high dimension.
To evaluate the performance of a QNN quantum neural network in learning training data, a cost function needs to be constructed. In quantum information, there is an intrinsically unique measure of fidelity for the proximity of different quantum states. Defining the cost function as the fidelity between the output of the QNN and the desired output on the training data:
C = 1 N x = 1 N ϕ x out ρ x out ϕ x out
We train the QNN by optimizing the cost function C. In each training step, the perceptron unitary operator is updated according to U e i ϵ K U , where K is a matrix containing all parameters of the corresponding perceptron unit, and ϵ is the choice step length. The matrix K is chosen such that the cost function grows the fastest. The change in cost function C is given by:
Δ C = ϵ n x = 1 n l = 1 L + 1 tr σ x l Δ ε l ρ x l 1
ρ x l = E l E 2 E 1 ρ x in
σ x l = F l + 1 F L F out ϕ x out ϕ x out

3. Methods

When we encode quantum states with quantum gates, the quantum gates themselves introduce noise. Quantum gates have been shown to be fault-tolerant in quantum coding. However, considering that the subsequent encoding will pass through a noisy channel, the impact of quantum gates cannot be ignored.
After we encode the quantum states with the quantum convolutional code, the fidelity of the encoded quantum states needs to be improved before entering the quantum channel due to the presence of noise in the quantum gates themselves and the fact that the quantum states involved in the encoding do not necessarily have a high purity. The QNN can then play its role.
We chose the [ 6 , 2 , 2 ] quantum convolutional codes. The generator matrix is as follows:
M = 1 0 0 1 0 0 0 1 1 0 0 0 D 0 0 1 0 0 0 0 0 0 1 1 0 1 1 0 0 0 1 1 1 1 0 0 0 0 0 0 1 1 D 0 0 1 1 1
The corresponding quantum circuit is shown in Figure 3.
As shown in Figure 3, | δ i refers to the target quantum bit that needs to be encoded. The ratio of the number of input auxiliary qubits to the number of target qubits is 2:1. The output 6-bit quantum state is controlled by 4 qubits at the current moment and 2 qubits in the previous coding module, so that the goal of associating the input information from different moments before and after together is achieved, and the noise affected by the coding states at different moments is no longer independent and more conducive to decoding. Moreover, each 2 target qubits are in fact controlled by 6 qubits at the same time, but the ratio of the number of auxiliary qubits to the number of encoding qubits is 2:1, which also saves some resources and achieves higher bit rates.
In order to reduce the effect caused by the noise of quantum gates, the encoded quantum state is trained with a quantum neural network (QNN). When encoding the target quantum state, the quantum state information is known, so the fidelity of the encoded quantum state information is also known before entering the quantum channel, and it can be improved by the quantum neural network. Firstly, according to the encoding circuit, we can derive the four encoded states corresponding to the four inputs, | 00 , | 01 , | 10 , | 11 , respectively, and train them with a certain volume of sample set to get the QNN for different coding states. The trained QNN is then applied to the noisy coding quantum states to obtain the high-fidelity encoded quantum states.
The complexity of the QNN increases exponentially with the increase in the number of qubits. To reduce the complexity and training difficulty of the QNN, we can perform block processing on the quantum convolutional code circuit, and only train 4 qubits once. The block method is as shown in Figure 4.
For each 6-bit quantum state, it is divided into two parts for the QNN training. As shown in Figure 4, the input of the 4-bit quantum state in the first part is | 0000 , and the first 4 qubits of the encoded state are obtained after passing through the encoding circuit, which is not affected by the input information, so 4 different input quantum states can all use the same QNN. The 4-bit quantum state in the second part is encoded by | 00 through the quantum gate operation, and then the input quantum state is controlled for encoding. Therefore, 4 different QNNs are required to be trained for 4 input quantum states.
Take the input quantum state | 00 as an example. We construct two QNNs with the expectation that the quantum states are trained to the standard quantum states after noiseless quantum gates. For the first part of the quantum state, the expected result is 1 4 ( 0000 + 0001 + 1000 + 1001 ) , and the expected result of the second quantum state is 1 4 ( 0000 + 0011 + 0100 + 0111 0001 0010 0101 0110 ) . In this way, the target output quantum state is known, so the unknown unitary operator can be trained by using the sample sets, and the resulting QNN can be trained layer by layer according to the input encoded quantum state to obtain the unitary operator that adapts the noisy encoded quantum state. Then, we use the unitary operators acting on the encoded quantum state to obtain a quantum state that approximates the target noiseless quantum state. The settings of the two QNNs are identical except that the sample set and the target quantum state are different.

4. Simulation and Analysis

Assuming that all quantum gates are noisy quantum gates, we provide the simulation of the whole circuit in Figure 5.
It can be seen through simulation that when the noise of the quantum gate is taken into account, the qubit error rate is greatly increased. As the SNR increases, in the case of a noisy quantum gate, the qubit error rate interval is [ 23.44 % , 0.45 % ] , and compared with the noiseless quantum gate, the qubit error rate interval is [ 14.96 % , 0.057 % ] .
The results of training the first part of the quantum state are shown in Figure 6. Without the QNN, its fidelity is relatively discrete. After applying the QNN on the quantum state, the fidelity of the quantum state is very close to 1, indicating that it is very close to the desired target quantum state.
However, the training effect of the second part is relatively poor. It can be seen from Figure 7 that the fidelity is closer to 1 than before training, but it is still relatively discrete. By observing the quantum states of the two parts, it is not difficult to see that the quantum states of the second part have a higher degree of entanglement and are more complicated to train, so the fitting may be poor.
We simulate the cost function of the two parts to observe the fit of the QNN in Figure 8.
As shown in the figure, with the increase of training samples, the cost functions of both parts rise rapidly, but the trends are different. The cost function of the first part converges quickly, the rise is linear, and the final convergence value is very close to 1, so the training effect is better, and the fidelity close to 1 can be obtained when applied to the encoding quantum state. The second part of the cost function is more tortuous in ascending, probably because the data is more complex, it is difficult to select a suitable direction, and the convergence is slower. After fitting, it finally converges to 0.92, which is still some distance from 1. The fitting effect is not ideal, but still useful for improving the fidelity of the quantum state.
After applying the QNN for all encoded quantum states, the trained encoded quantum states are used to enter the quantum noise channel. We simulate the final BER, and the result is shown in Figure 9.
As shown in the figure, the final BER after QNN is intermediate between the case of noiseless quantum gates and the case of noisy quantum gates. Since the convergence of the second part of the QNN is not sufficient, the encoded quantum state cannot reach the noise-free criterion. In addition, the quantum gate at the receiver used to decode the encoded quantum state is also noisy. Since the received quantum information cannot be obtained at the receiver side without measurement, the QNN cannot be applied to optimize the decoding circuit at the receiver.
In a word, the quantum neural network has a good correction effect on the encoding state of the quantum convolutional code. It can defend the noise of quantum gates by improving the fidelity of the encoded quantum state, which has a certain effect on reducing the BER.

5. Conclusions

This paper provides a method to reduce the error probability of the quantum convolutional code by enhancing the fidelity of the encoded states using quantum neural networks. Considering that the quantum gates used by the encoding and decoding circuits have noise, we use a quantum neural network to optimize the quantum state after the encoding circuit. A quantum neural network can convert a quantum state with lower fidelity into a target quantum state with higher fidelity through the unitary transformation obtained by learning. Through quantum neural networks, the fidelity of a large number of noisy quantum states is improved from a discrete state to a state whose fidelity is close to 1.
The increase in qubits to be processed will bring exponentially increased complexity to the quantum neural network. When too many qubits are input to the QNN once, the gradient will collapse, and the fidelity cannot be effectively improved. Therefore, we divide the encoding circuits that need to be optimized for processing, and divide the quantum state with more qubits into multiple quantum states with fewer qubits for quantum neural network optimization. After the quantum neural network optimization, the BER is greatly improved compared with the unoptimized case. After simulation, under the condition of 3 dB SNR, the BER of 0.45 % in the case of noisy quantum gates can be reduced to 0.19 % .
In the case of assuming no noise, the signal-to-noise ratio at 3 dB is 0.057 % . There is still a certain gap after quantum neural network optimization. The reasons include: (1) the output quantum state of the decoding circuit is unknown and cannot pass through the quantum neural network. However, the quantum gate of the decoding circuit also has noise; (2) the training results of the quantum neural network for more complex quantum states are poor, and the effect of completely improving the fidelity to 1 cannot be achieved. Thus, there is a certain gap with the ideal situation.

Author Contributions

Formal analysis, H.X.; Supervision, X.C.; Writing—review & editing, J.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shor, P.W. Scheme for reducing decoherence in quantum computer memory. Phys. Rev. A 1995, 52, R2493. [Google Scholar] [CrossRef] [PubMed]
  2. Calderbank, A.R.; Shor, P.W. Good quantum error-correcting codes exist. Phys. Rev. A 1996, 54, 1098–1105. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Steane, A. Multiple Particle Interference and Quantum Error Correction. Proc. R. Soc. Lond. Ser. A 1996, 452, 2551–2577. [Google Scholar]
  4. Gottesman, D. Class of quantum error-correcting codes saturating the quantum Hamming bound. Phys. Rev. A 1996, 54, 1862. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Chau, H.F. Quantum convolutional error-correcting codes. Phys. Rev. A 1998, 58, 905. [Google Scholar] [CrossRef] [Green Version]
  6. Johannesson, R.; Zigangirov, K.S. Fundamentals of Convolutional Coding; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  7. Ollivier, H.; Tillich, J.P. Description of a Quantum Convolutional Code. Phys. Rev. Lett. 2003, 91, 177902. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Forney, G.D.; Guha, S. Simple rate-1/3 convolutional and tail-biting quantum error-correcting codes. In Proceedings of the International Symposium on Information Theory, 2005. ISIT 2005, Adelaide, SA, Australia, 4–9 September 2015; IEEE: Piscataway, NJ, USA, 2005; pp. 1028–1032. [Google Scholar]
  9. Forney, G.D.; Grassl, M.; Guha, S. Convolutional and tail-biting quantum error-correcting codes. IEEE Trans. Inf. Theory 2005, 53, 865–880. [Google Scholar] [CrossRef] [Green Version]
  10. Grössing, G.; Zeilinger, A. Quantum cellular automata. Complex Syst. 1988, 2, 197–208. [Google Scholar]
  11. Narayanan, A.; Menneer, T. Quantum artificial neural network architectures and components. Inf. Sci. 2000, 128, 231–255. [Google Scholar] [CrossRef]
  12. Menneer, T. Quantum Artificial Neural Networks. Ph.D. Thesis, The University of Exeter, Exeter, UK, 1998. [Google Scholar]
  13. Toth, G.; Lent, C.S.; Tougaw, P.D.; Brazhnik, Y.; Weng, W.; Porod, W.; Liu, R.W.; Huang, Y.F. Quantum Cellular Neural Networks. Superlattices Microstruct. 2000, 20, 473–478. [Google Scholar]
  14. Menneer, T.; Narayanan, A. Quantum-inspired neural networks. In Proceedings of the NIPS’95, Denver, CO, USA, 27 November–2 December 1995. [Google Scholar]
  15. Ventura, D.; Martinez, T. Quantum associative memory with exponential capacity. In Proceedings of the IEEE World Congress on IEEE International Joint Conference on Neural Networks, Anchorage, AK, USA, 4–9 May 1998. [Google Scholar]
  16. Behrman, E.C.; Steck, J.E.; Skinner, S.R. A spatial quantum neural computer. In Proceedings of the IJCNN’99, International Joint Conference on Neural Networks, Washington, DC, USA, 10–16 July 1999. [Google Scholar]
  17. Weigang, L. Entangled Neural Networks; Department of Computer Science, University of Brasilia: Brasilia, Brazil, 2000; Volume 4466, pp. 70919–70970. [Google Scholar]
  18. Verdon, G.; Pye, J.; Broughton, M. A universal training algorithm for quantum deep learning. arXiv 2018, arXiv:1806.09729. [Google Scholar]
  19. Beer, K.; Bondarenko, D.; Farrelly, T.; Osborne, T.J.; Wolf, R. Training deep quantum neural networks. Nat. Commun. 2020, 11, 808. [Google Scholar] [CrossRef] [Green Version]
  20. Gottesman, D. Stabilizer Codes and Quantum Error Correction. Ph.D. Thesis, California Institute of Technology, Pasadena, CA, USA, 1997. [Google Scholar]
  21. Gottesman, D. An Introduction to Quantum Error Correction and Fault-Tolerant Quantum Computation. Quantum Inf. Sci. Contrib. Math. 2009, 91, 13–58. [Google Scholar]
  22. Djordjevic, I. Quantum Information Processing and Quantum Error Correction: An Engineering Approach; Academic Press: Cambridge, MA, USA, 2012. [Google Scholar]
  23. Ollivier, H.; Tillich, J.P. Quantum convolutional codes: Fundamentals. HAL—INRIA 2004, 54, 4053–4068. [Google Scholar]
Figure 1. Circuit for encoding the first three qubits of a stream of quantum information with the 5-qubit convolutional code.
Figure 1. Circuit for encoding the first three qubits of a stream of quantum information with the 5-qubit convolutional code.
Applsci 12 05662 g001
Figure 2. The structure of a QNN circuit.
Figure 2. The structure of a QNN circuit.
Applsci 12 05662 g002
Figure 3. The encoding circuit of the [6, 2, 2] quantum convolutional code.
Figure 3. The encoding circuit of the [6, 2, 2] quantum convolutional code.
Applsci 12 05662 g003
Figure 4. The block method of the [6, 2, 2] quantum convolutional code.
Figure 4. The block method of the [6, 2, 2] quantum convolutional code.
Applsci 12 05662 g004
Figure 5. Simulation of the [6, 2, 2] quantum convolutional code with noisy quantum gates.
Figure 5. Simulation of the [6, 2, 2] quantum convolutional code with noisy quantum gates.
Applsci 12 05662 g005
Figure 6. The fidelity of quantum states before and after training the QNN on the quantum state of the first part.
Figure 6. The fidelity of quantum states before and after training the QNN on the quantum state of the first part.
Applsci 12 05662 g006
Figure 7. The fidelity of quantum states before and after training the QNN on the quantum state of the second part.
Figure 7. The fidelity of quantum states before and after training the QNN on the quantum state of the second part.
Applsci 12 05662 g007
Figure 8. The cost function for the QNNs of two parts.
Figure 8. The cost function for the QNNs of two parts.
Applsci 12 05662 g008
Figure 9. The BER of the [6, 2, 2] quantum convolutional codes applying the QNN compared with BER before training and the noiseless quantum gate.
Figure 9. The BER of the [6, 2, 2] quantum convolutional codes applying the QNN compared with BER before training and the noiseless quantum gate.
Applsci 12 05662 g009
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiao, H.; Chen, X.; Xu, J. Using a Deep Quantum Neural Network to Enhance the Fidelity of Quantum Convolutional Codes. Appl. Sci. 2022, 12, 5662. https://doi.org/10.3390/app12115662

AMA Style

Xiao H, Chen X, Xu J. Using a Deep Quantum Neural Network to Enhance the Fidelity of Quantum Convolutional Codes. Applied Sciences. 2022; 12(11):5662. https://doi.org/10.3390/app12115662

Chicago/Turabian Style

Xiao, Hanwei, Xiaoguang Chen, and Jin Xu. 2022. "Using a Deep Quantum Neural Network to Enhance the Fidelity of Quantum Convolutional Codes" Applied Sciences 12, no. 11: 5662. https://doi.org/10.3390/app12115662

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop