Next Article in Journal
Assessing by Simulation the Effect of Process Variability in the SALB-1 Problem
Previous Article in Journal
Physics-Informed Neural Networks for Bingham Fluid Flow Simulation Coupled with an Augmented Lagrange Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

QPDE: Quantum Neural Network Based Stabilization Parameter Prediction for Numerical Solvers for Partial Differential Equations

Department of Computational and Data Sciences, Indian Institute of Science, Bengaluru 560012, Karnataka, India
AppliedMath 2023, 3(3), 552-562; https://doi.org/10.3390/appliedmath3030029
Submission received: 4 May 2023 / Revised: 11 June 2023 / Accepted: 20 June 2023 / Published: 13 July 2023
(This article belongs to the Special Issue Mathematical Perspectives on Quantum Computing and Communication)

Abstract

:
We propose a Quantum Neural Network (QNN) for predicting stabilization parameter for solving Singularly Perturbed Partial Differential Equations (SPDE) using the Streamline Upwind Petrov Galerkin (SUPG) stabilization technique. SPDE-Q-Net, a QNN, is proposed for approximating an optimal value of the stabilization parameter for SUPG for 2-dimensional convection-diffusion problems. Our motivation for this work stems from the recent progress made in quantum computing and the striking similarities observed between neural networks and quantum circuits. Just like how weight parameters are adjusted in traditional neural networks, the parameters of the quantum circuit, specifically the qubits’ degrees of freedom, can be fine-tuned to learn a nonlinear function. The performance of SPDE-Q-Net is found to be at par with SPDE-Net, a traditional neural network-based technique for stabilization parameter prediction in terms of the numerical error in the solution. Also, SPDE-Q-Net is found to be faster than SPDE-Net, which projects the future benefits which can be earned from the speed-up capabilities of quantum computing.

1. Introduction

Quantum computing has the potential to accelerate traditional algorithms in science and engineering, including the data-driven schemes for solving Partial Differential Equations (PDEs). Since this research area is quite new, we don’t see much literature available for speeding up traditional solvers with quantum computing. There are a lot of problems in the literature on scientific computing which can harness the advantages of quantum computing. One such possibility is to improve traditional fluid flow simulations as they are compute-intensive and scale with the size of the problem at hand. To address this, we propose the development of an approach that leverages the advantages of quantum computing and deep learning methods for solving PDEs. Our main objective is to develop a Quantum Neural Network (QNN) for predicting stabilization parameters when solving SPDEs using the Streamline Upwind Petrov Galerkin (SUPG) technique. We introduce SPDE-Q-Net, a QNN for approximating the value of the stabilization parameter for SUPG in 2-dimensional convection-dominated problems. Our results show that the performance of SPDE-Q-Net is comparable to that of SPDE-Net, a traditional neural network-based technique for stabilization parameter prediction. This research has the potential to address a significant challenge in fluid flow simulations by providing stabilized and efficient solutions. By combining quantum computing with neural networks, our approach could lead to significant improvements in a wide range of fields, including engineering and science. However, further research is needed to address the complex challenges in this area and to fully realize the potential of this approach.

1.1. Neural Networks for Solving PDEs

The development of data-driven PDE solvers has spawned a plethora of studies. Several data-driven numerical PDE solving techniques exist such as Physics Informed Neural Networks (PINNs) [1,2], Fourier Neural Operator (FNO) [3,4]. PINNs use a neural network to approximate the numerical solution by minimizing the strong form of the residual. Their performance, however, is limited for singularly perturbed partial differential equations (SPDEs). To solve this issue, Yadav and Ganesan [5] proposed to use a hybrid scheme to solve SPDEs. This hybrid scheme uses Artificial Neural Networks (ANN) to predict the stabilization parameter and stabilized FEM for solving the equation. Having a neural network in the solving pipeline makes the entire process slow, so we proposed to use a QNN so that we can show a quicker convergence of the proposed neural network. In this current work, we leverage Quantum Computing to predict the stabilization parameter for solving SPDEs.

1.2. Quantum Computing

As explained in [6], Quantum computing generally refers to the controlled evolution of a set of quantum two-level systems of a 2-dimensional Hilbert space ( H 2 ) with basis | 0 , 1 | , which are called qubits. This can better be explained with a wavefunction as
| ψ = α | 0 + β | 1 | α | 2 + | β | 2 = 1 , α , β C
Unitary quantum gates manipulate qubits in the same way logic gates affect conventional bits in a computer, allowing qubits to evolve through so-called quantum channels. Quantum computers have been proposed as linear algebra accelerators, paving the way for quantum machine learning (QML). The use of quantum computing devices to do QML is still in its early phases of development [7,8,9,10,11]. While completely quantum algorithms have been shown to have an asymptotic advantage, deep protocols are still out of reach of current QC capabilities. This necessitates new ways based on a hybrid quantum-classical process, which is primarily motivated by deep learning’s success in scientific computing. Data embedding circuits (feature maps) [6] and parametrized variational circuits are the building blocks of variable QML models, which may be trained to execute various unsupervised tasks [8]. These can very well be used for regression problems [12].

1.3. Quantum Neural Network

The primary idea behind introducing quantum features into classical Neural Networks is to replace the classical neurons z = { 1 , 1 } with ‘qurons’ | z with basis { | 0 , | 1 } [13,14]. It is a parametrized quantum circuit. Quantum feature maps are used to encode the inputs to the feature space. This feature map is a unitary circuit consisting of x-dependent gates. When this feature map is applied to any initial state, it transforms them into quantum states. A variational quantum circuit then processes the quantum states.
A QNN is a machine learning model or algorithm that combines concepts from quantum computing and ANNs. QNN is used to represent variational or parametrized quantum circuits. While mathematically somewhat different from the inner workings of neural networks, the analogy highlights the “modular” nature of quantum gates in a circuit and the wide use of tricks from training neural networks used in the optimization of quantum algorithms. We extend the network proposed by [15]. Every input data is stored in s register, | α i , the hidden states will be computing the weighted sum of the inputs and will represent them as | ψ i j for j t h hidden unit.

1.4. Numerical Methods for Solving SPDEs

Higher-order convection-diffusion equations do not have analytical, closed-form solutions, and numerical methods are often used to approximate their solution. Mesh-based methods, such as Finite Difference, FE, and FV, are famous for the numerical simulation of such equations. These techniques have a conservative formulation, higher-order accuracy, and the ability to accommodate complex geometries. Further, many of these methods are amenable to parallelization. However, solving SPDEs is challenging since it requires the numerical schemes to carefully treat high-gradient regions (boundary layer/interior layer) typically associated with such equations to avoid spurious oscillations. The numerical schemes often use a stabilization term to deal with the singularly perturbed equation. Several stabilization strategies exist, such as slope limiting, flux limiting, the introduction of artificial dissipation, etc. In the context of FEM, the SUPG method is a remarkably successful stabilization scheme that follows the philosophy of artificial dissipation. The scheme has been widely used to solve SPDEs for problems such as contaminant transport in shallow water, evolutionary convection-diffusion-reaction equations, arbitrary Lagrangian-Eulerian framework for time-dependent convection-diffusion equations, linear hyperbolic problems, and reduced order models among other areas. In addition to the SUPG technique, the local projection stabilization, edge stabilization, orthogonal subgrid scale, Galerkin least-squares [16] and continuous interior penalty method are some other popular techniques used to stabilize numerical methods for simulation of convection-dominated equations.

1.5. Streamline Upwind/Petrov Galerkin

Many techniques exist for solving SPDEs, most of them have some sort of stabilization in them. One such popular stabilization technique is the streamline upwind/Petrov Galerkin technique, generally called SUPG. It was introduced in [16] and is the focus of our current research work. It provides a nodally correct solution to the 1-d convection-diffusion problem. The accuracy of the SUPG technique for higher-dimensional problems is limited as it depends on the appropriate value of the stabilization parameter. This stabilization parameter is hard to find and is a topic of research for the last few decades.

1.6. On the Choice of Stabilization Parameter for SUPG

The value of the stabilization parameter has a significant impact on the correctness of the SUPG numerical solution. This parameter specifies how much artificial diffusion should be used in the equation to achieve a steady result. Artificial diffusion with a high value can cause the solution to smear, but under-diffusion will not eliminate the false oscillations. Many approaches for determining the ideal value of the stabilization parameter have been developed, but each has its own set of restrictions and is difficult to generalize to higher dimensions.

1.7. SPDE-Net

Many techniques have been proposed for estimating the value of the stabilization parameters. Yadav and Ganesan [5,17,18,19] employed a neural network to estimate the stabilization parameter for the SUPG methodology with finite element methods-based loss formulation for solving SPDEs. In [5], SPDE-Net is used as an optimizer to estimate the stabilization parameter ( τ ) for 1-dimensional problems. The performance of both supervised and unsupervised approaches is compared and shown to be comparable, implying that unsupervised training can be self-sufficient. Furthermore, the L 2 error from SPDE-Net was significantly lower than that from the state-of-the-art PINN networks. It suggests that we need to supplement PINN’s loss formulation with a stabilization term for higher-dimensional challenges.

1.8. Contributions

  • We propose a Quantum Neural Network for predicting stabilization parameters for solving SPDEs using the SUPG stabilization technique
  • Developed an unsupervised quantum model for SPDE-Net, based on equation coefficients and the local gradients based normalization
  • Compared the performance of classical neural network (SPDE-Net) and SPDE-Q-Net in terms of different errors such as L 2 , H 1 , L , relative l 2
  • SPDE-Q-Net has been tested for different mesh refinements to check its generalization ability

1.9. Organization of the Paper

In Section 1, we provide an introduction to the problem addressed in this research. Section 2 covers the necessary background and foundational information for the proposed work. We delve into the specifics of the proposed network architecture in Section 3. The experimental methodology, along with testing examples, is presented in detail in Section 4. Finally, we conclude the manuscript in Section 6, summarizing the key findings and their implications.

2. Preliminaries

In this paper, our focus is on solving the convection-diffusion equation, a fundamental equation in the field of computational science and engineering. The convection-diffusion equation arises in various domains, including fluid dynamics, heat transfer, and mass transport, and plays a crucial role in modeling physical phenomena. It combines the effects of advection (convection) and diffusion, making it a challenging problem to solve accurately and efficiently.

2.1. Convection-Diffusion Equation

ϵ Δ u + b · u = f in Ω and u = u b on Ω .
Here Ω R 2 , is a bounded domain with a polygonal Lipschitz-continuous boundary Ω , n is outward normal vector to the boundary Ω , ϵ > 0 is the diffusion coefficient, b = [ b 1 , b 2 ] W 1 , ( Ω ) d is the flow velocity, f L 2 ( Ω ) is the external source term, u is the unknown scalar term and u b H 1 / 2 ( Ω ) . First, the variational form of the equation is derived using the Galerkin technique. It is to find u such that for all v H 0 1 ( Ω )
a ( u , v ) = ( f , v )
where the bilinear form a ( · , · ) : H 1 ( Ω ) × H 0 1 ( Ω ) R is defined by
a ( u , v ) = Ω ϵ u · v d x + Ω b u v d x
( f , v ) = Ω f v d x
(·,·) is the L 2 ( Ω ) inner product. Let T h be a family of triangulations of Ω parametrized by positive parameters h whose only accumulation point is zero. The triangulations T h are assumed to consist of a finite number of open polygonal subsets K of Ω such that Ω ¯ = K T h K ¯ . The domain Ω is partitioned into Ω h P 2 and the weak formulation for this discretized domain is to choose a finite-dimensional space V h H 0 1 ( Ω ) comprising continuous piecewise polynomials and find u h V h such that for all v h V h we have
a h ( u h , v h ) = ( f , v h ) a u h , v h : = ε u h , v h + b · u h , v h = f , v h v h V h
The solution of this variational form contains spurious oscillations near the boundary layer, so a weighted residual term has been added to it as explained in the next subsection.

2.2. SUPG Stabilization

The residual of equation [2] is:
R ( u ) = ϵ Δ u + b · u f
The term R ( u ) has been added to the discretized weak formulation given in Equation (6). Now, modified weak form for domain Ω h , is to find u h V h such that v h V h :
a h ( u h , v h ) = f h ( v h ) v h V h where , a h ( u h , v h ) = ϵ ( u h , v h ) + ( b · u h , v h ) + i Ω h τ i ( ϵ Δ u h + b · u h f h , b · v h ) Ω h
τ i L 2 ( Ω ) is a user-chosen non-negative stabilization parameter. Its value plays an important role in the quality of the approximated solution. A very large value can show unexpected smearing, whereas a low value will not remove the spurious oscillations. So, an optimal value of stabilization parameter is required, which can control oscillations and smearing both. Many studies have established the bounds for τ for the SUPG technique in order to achieve quasi-optimal numerical convergence. One such widely used expression is given as follows.

2.3. Standard Stabilization Parameter

There is a standard mathematical expression to calculate the value of the stabilization parameter and it gives a nodally precise solution for a 1-dimensional problem and simple 2-dimensional problems. However, it is not generalizable to complex 2-dimensional problems such as problems with variable coefficients. To make it generalized we are proposing a NN-based method for predicting the stabilization parameter.
τ s t d T = h T 2 | b | ξ 0 P e T , P e T = | b | h T 2 ε where ξ 0 ( α ) = coth α 1 α ,
We will call it Std. τ for the sake of convenience. We are using a piecewise constant element to define τ on the mesh. τ K will be locally defined for each cell in the mesh.

3. Network Architecture

We predict a single stabilization parameter for the entire mesh and then locally divide it with the norm of gradients of the Galerkin solution in that cell. This way, we have taken care of cell-wise stabilization. By dividing with the gradient of the stabilization parameter, we obtain an adaptive τ . In this research, we use the QNN proposed by [15] and find the output for each input feature, namely | ψ of size M. The predicted stabilization parameter can be represented as: τ ^ = i = 1 M | ψ i M .

3.1. SPDE-Q-Net

We use QNN for approximating the stabilization parameter τ . In a QNN, we use initial states and then map them to quantum bits (qubits) using a quantum feature map. Generally, the feature map consists of a unitary circuit Q ^ ϕ ( x ) , which is a concatenation of x-dependent gates and performs single qubit rotation. When applied to an initial state | (a quantum initial state), the feature map transforms x P Ω data points into quantum states | ψ ( x ) : = Q ^ ϕ ( x ) | . These quantum states are processed by variational quantum circuit V ^ θ parametrized by parameters θ . Mathematically a QNN can be expressed as follows:
I n p u t : I = { ϵ , b 1 , b 2 , h T } , P e T = | b | h T 2 ϵ τ s t d T = h T 2 | b | ( coth ( P e T ) 1 P e T ) u s t d = S o l v e s u p g ( τ s t d ) Q ^ θ ( x ) = r = 1 R ( V ^ θ ( r ) Q ^ θ ( r ) ( x ) ) V ^ θ ( 0 ) τ ( θ ) = | Q ^ θ ( x ) C ^ Q ^ θ ( x ) | / ( | u s t d | ) m ( s ) = s s > 1 2.5 s 2 1.5 s 3 otherwise b = ( b 2 , b 1 ) | b | when | b | 0 0 when | b | = 0 L o s s ( θ ) = ϵ ( u h , v h ) + ( b · u h , v h ) + i Ω h τ i ( ϵ Δ u h + b · u h f h , b · v h ) Ω h f h ( v h ) + m ( | b · u h | ) θ * = a r g m i n L o s s ( I )
u h is the SUPG solution of Equation (7), h T is the mesh size, τ is stabilization parameter. We must learn Q ^ by finding optimal θ * through iterative back propagation of loss. SPDE-Q-Net has R number of layers, and initially, V ^ θ 0 operates the initial state. τ is the expected value of the cost operator. Here the cost operator is the weighted sum of Pauli operators, C ^ = j c j P ^ j or fix c j = constant j . SPDE-Q-Net has five input qubits and one output qubit. It has an Adam optimizer and Sigmoid activation function for six hidden layers with twenty quorons in each hidden layer.

3.2. Back Propagation in Quantum Neural Network

We discuss the back-propagation pipeline of SPDE-Q-Net here. We have numerically approximated the gradients of the expected value of cost by finite differences. It is affected by the quantum measurement noise, so there will be a bias involved in learning the network. The other way could be to use analytic derivatives of these expressions. Here, we use the parameter shift rule as described in [6]. Let us take an example case of unitary operator U ^ ( x ) = e x p ( ι ϕ ( x ) G ^ / 2 ) generated by a Hermitian operator G ^ . ψ ( x ) denotes an extra function applied to the feature variable, such as a feature non-linearity. Next, let us consider a simplified quantum model
q ( x ) = ψ ˜ | U ^ ( x ) C ˜ U ^ ( x ) | ψ ˜
where | ψ ˜ and C ˜ are some generic input states and dressed cost operators that do not depend on x. The circuit derivative reads
d U ˜ ( x ) d x = ( ι ψ ( x ) / 2 ) G ^ U ^ ( x )
and we can differentiate the quantum model as
d q ( x ) d x = ( ι ψ ( x ) / 2 ) ψ ˜ | U ^ ( x ) [ G ^ , C ˜ ] U ^ ( x ) | ψ ˜
where [ · , · ] denotes a commutator. In the general form, the expression above can be measured as an overlap between quantum states. However, to simplify the readout, derivatives of unitaries with a known spectrum of G ^ can be calculated exactly (without bias) by evaluating the model at shifted arguments of q ( x ) . Specifically, for unitaries that Pauli operators generate G ^ : = P ^ j (i.e., single qubit rotations at site j), where operators satisfy involutory property P ^ j 2 = 1 j and have just two unique eigenvalues of ± 1 . g ( x ) = ψ ( x ) [ g ( x + π / 2 ) g ( x π / 2 ) ] / 2 .

3.3. SPDE-Net

We compare the performance of SPDE-Q-Net with a classical neural network SPDE-Net trained for the same task and detailed as below. For a given input data sample,
I n p u t : I = { ϵ , b 1 , b 2 , h T } , P e T = | b | h T 2 ϵ τ s t d T = h T 2 | b | ( coth ( P e T ) 1 P e T ) τ ( θ ) = G θ ( P e T ) | | u s t d | | 2 L o s s ( θ k ) = ( Ω ( u ^ ( τ ^ ( θ k ) ) u ) d x ) 1 2 θ * = a r g m i n L o s s ( I , u ( τ ( θ ) ) )
where G θ is θ parameterized SPDE(Net), u ( τ ( θ ) ) is the SUPG solution of Equation (7), and τ ( θ ) is the predicted stabilization parameter. We have to learn G θ by finding optimal θ * by back propagating the loss iteratively.

3.4. Error Metrics

Here we discuss the metrics used to quantify the numerical errors in the solution. Four different error metrics are presented. The first is the L 2 -error, denoted as | e h | 0 , which measures the difference between the computed solution u h and the known analytical solution u in the L 2 norm over the domain Ω . The relative l 2 -error, denoted as | e h | 0 , , is a sum of the pointwise differences between u h and u normalized by the l 2 norm of u at specific points x i in the domain Ω h . The H 1 -error, denoted as | e h | 1 , measures the difference between the gradients of u h and u in the L 2 norm over Ω . Finally, the L -error, denoted as | e | L ( Ω ) , represents the essential supremum of the absolute difference between u h and u over the domain Ω . We use the following metrics to calculate numerical errors in the solution.
L 2 - error : e h 0 = u h u L 2 ( Ω ) = Ω ( u h u ) 2 d x 1 2 Relative l 2 - error : e h 0 , = i = 1 N u h ( x i ) u ( x i ) 0 , u 0 , , x i Ω h H 1 - error : | e h | 1 = u h u ) L 2 ( Ω ) = Ω ( u h u ) 2 d x 1 2 L - error : e L ( Ω ) = e s s s u p { | u h u | : x Ω } .
Here, τ ^ is the predicted stabilization parameter, u is the known analytical solution, u h ( τ ^ ) is the SUPG solution calculated with τ ^ , and D α is the weak derivative.

3.5. Order of Convergence

Given a sequence of approximations { x n } converging to a true solution x as n approaches infinity, the order of convergence of a numerical method is defined as follows:
Let { ϵ n } denote the sequence of errors, where ϵ n = | x n x | . If there exist positive constants C and p such that
lim n ϵ n + 1 ϵ n p = C ,
where p is known as the order of convergence, then the numerical method is said to converge with order p or is p-th order convergent. A higher value of p implies faster convergence. The value of C can provide additional information about the convergence behavior. If C = 0 , the method is said to have superlinear convergence, whereas C > 0 indicates sublinear convergence. If C = 1 , the method exhibits linear convergence, and if C > 1 , it is superlinearly convergent. The order of convergence provides a measure of how rapidly a numerical method approaches the true solution as the number of iterations increases.

4. Numerical Experiments

In this section, we consider four different examples of SPDEs based on the location of interior and boundary layers. For each example, we predict the τ using SPDE-Net, and SPDE-Q-Net. We compare the performance of different techniques in terms of errors metrices defined in Section 3.4. All these errors are computed with respect to the analytical solution. SPDE-Q-Net is optimized by minimizing the weak residual of the equation as mentioned in [20]. The network training was carried out for different mesh sizes, but here we only present the mesh results with 40 cells in the x and y-direction. The computation time taken by different techniques is given in Table 1.
Example 1. 
We have taken a two-dimensional convection–diffusion problem as given by Equation (2) with the following data for cheking the performance of SPDE-Q-Net:
ϵ = 10 8 , b = ( 1 , 0 ) , f = 1 , Ω = ( 0 , 1 ) 2 u = 0 on Ω
It has an exponential layer at x = 1 and two parabolic layers at y = 0 and y = 1 , respectively. In the interior domain, the analytical solution u ( x , y ) is quite close to x. The SPDE-Q-Net captures the pattern of τ in accordance with the location of the interior/boundary layer region. The L 2 error produced by all the considered techniques is compared in Table 2, and we can see that the L 2 error produced by SPDE-Q-Net is at par with SPDE-Net.
Example 2. 
Next, we consider Equation (2) with following data:
ϵ = 10 8 , b = ( 2 , 3 ) , Ω = ( 0 , 1 ) 2 , u = 0 on Ω u ( x , y ) = x y 2 x e x p ( 3 ( y 1 ) ϵ ) y 2 e x p ( 2 ( x 1 ) ϵ ) + e x p ( 2 ( x 1 ) + 3 ( y 1 ) ϵ )
It has two outflow boundary layers and is hence a suitable test case for checking the performance of the SPDE-Q-Net. We observed the clear pattern of the boundary layers in the heat map of the predicted τ. In the non-boundary region, the predicted τ is close to 1.0, whereas, in the boundary region, the value is close to the std. tau value and hence this localized τ ensures enough stabilization in the areas where it is required the most.
Example 3. 
ϵ = 10 8 , θ = π / 3 , b = ( c o s ( θ ) , s i n ( θ ) ) , f = 0.0 , Ω = ( 0 , 1 ) 2 , u = u b on Ω , Ω N = ϕ u b = 0 , for x = 1 or y 0.7 1 , otherwise
This example contains both the exponential and boundary layer. The value of the standard stabilization parameter τ for this example is 0.0049, whereas all the predicted τ values are more than 0.01. Both the interior and exponential layers are very well captured by the SPDE-Q-Net and it is quantitatively shown in terms of different error metrics given in Table 2, Table 3 and Table 4.
Example 4. 
ϵ = 10 8 , b = ( 1 , 0 ) T , u = 0 on Ω D , Ω D = Ω , Ω N = ϕ f = 0 , if | x 0.5 | 0.25 | y 0.5 | 0.25 32 ( x 0.5 ) , otherwise u = 0 , if | x 0.5 | 0.25 | y 0.5 | 0.25 16 ( x 0.25 ) ( y 0.75 ) , otherwise
This example has two interior characteristic layers in the convection direction. Its location is between (0.25, 0.25) and (0.25, 0.75). For this example, the solution of both the std. τ and SPDE-Q-Net has minimal oscillations. For all the 4 testing example we have computed the computational time taken by both SPDE-Net and SPDE-Q-Net as shown in the Table 1, wherein we can see that SPDE-Net takes longer time to converge for all the mesh-sizes.

5. Results

Although it is not recommended to compare a classical network with the quantum neural network, as the network design and training are quite different, we compare the two predictions in terms of different errors, such as L 2 , H 1 , relative l 2 , and L , as shown in Table 2, Table 3, Table 4 and Table 5. We demonstrate the prediction for different examples from both simulated SPDE-Q-Net and SPDE-Net. SPDE-Net is implemented as mentioned in [5]. We compare the computational time taken by SPDE-Net and SPDE-Q-Net in Table 1 and find the SPDE-Q-Net to be always faster than SPDE-Net.

6. Conclusions

In this study, we have explored the application of Quantum Neural Networks (QNNs) in predicting stabilization parameters for solving Singularly Perturbed Partial Differential Equation (SPDEs) using the Streamline-Upwind Petrov-Galerkin (SUPG) technique. The SUPG technique is commonly used in fluid flow simulations to stabilize numerical solutions and improve accuracy. By leveraging the capabilities of QNNs, we aim to enhance the efficiency and accuracy of the SUPG method.
To effectively employ QNNs, we utilize the parameter shift rule, a technique that enables the calculation of derivatives up to the second order. This rule plays a crucial role in determining the appropriate stabilization parameters for the SUPG method. By leveraging quantum computing resources, specifically a simulated quantum circuit provided by Qiskit, we carry out all the necessary computations and simulations.
Our experiments are designed as a proof of concept, focusing on a basic set of scenarios. These initial experiments demonstrate the feasibility of using QNNs for predicting stabilization parameters in fluid flow simulations. By successfully calculating derivatives and optimizing the stabilization parameters using QNNs, we have laid the foundation for addressing more complex problems in fluid dynamics.
The application of QNNs in fluid flow simulations has significant potential for tackling larger and more intricate problems. By harnessing the power of quantum computing and the capabilities of QNNs, we can potentially achieve improved accuracy and computational efficiency in solving SPDEs with the SUPG technique. These preliminary experiments provide valuable insights and open up avenues for future research in the field of quantum-assisted fluid dynamics simulations.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Long, Z.; Lu, Y.; Ma, X.; Dong, B. PDE-Net: Learning PDEs from Data. In Proceedings of the Machine Learning Research, Stockholm, Sweden, 10–15 July 2018; pp. 3208–3216. [Google Scholar]
  2. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics Informed Deep Learning (Part I): Data-driven Solutions of Nonlinear Partial Differential Equations. arXiv 2017, arXiv:1711.10561. [Google Scholar]
  3. Li, Z.; Kovachki, N.; Azizzadenesheli, K.; Liu, B.; Bhattacharya, K.; Stuart, A.; Anandkumar, A. Fourier Neural Operator for Parametric Partial Differential Equations. arXiv 2020, arXiv:2010.08895. [Google Scholar]
  4. Li, Z.; Kovachki, N.; Azizzadenesheli, K.; Liu, B.; Bhattacharya, K.; Stuart, A.; Anandkumar, A. Neural Operator: Graph Kernel Network for Partial Differential Equations. arXiv 2020, arXiv:2003.03485. [Google Scholar]
  5. Yadav, S.; Ganesan, S. SPDE-Net: Neural Network-based prediction of the stabilization parameter for SUPG technique. In Proceedings of the 13th Asian Conference on Machine Learning, Virtual, 17–19 November 2021; pp. 268–283. Available online: https://proceedings.mlr.press/v157/yadav21a.html (accessed on 10 June 2023).
  6. Schuld, M.; Sweke, R.; Meyer, J.J. Effect of data encoding on the expressive power of variational quantum-machine-learning models. Phys. Rev. A 2021, 103, 032430. [Google Scholar] [CrossRef]
  7. Biamonte, J.; Wittek, P.; Pancotti, N.; Rebentrost, P.; Wiebe, N.; Lloyd, S. Quantum machine learning. Nature 2017, 549, 195–202. [Google Scholar] [CrossRef] [PubMed]
  8. Aïmeur, E.; Brassard, G.; Gambs, S. Machine Learning in a Quantum World. In Proceedings of the Advances in Artificial Intelligence; Quebec, QC, Canada, 7–9 June 2006; Lamontagne, L., Marchand, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2006; pp. 431–442. [Google Scholar]
  9. Carleo, G.; Troyer, M. Solving the quantum many-body problem with artificial neural networks. Science 2017, 355, 602–606. [Google Scholar] [CrossRef]
  10. Tiersch, M.; Ganahl, E.; Briegel, H. Adaptive quantum computation in changing environments using projective simulation. Sci. Rep. 2014, 5, 12874. [Google Scholar] [CrossRef] [PubMed]
  11. Lovett, N.B.; Crosnier, C.; Perarnau-Llobet, M.; Sanders, B.C. Differential Evolution for Many-Particle Adaptive Quantum Metrology. Phys. Rev. Lett. 2013, 110, 220501. [Google Scholar] [CrossRef] [PubMed]
  12. Dunjko, V.; Briegel, H.J. Machine learning & artificial intelligence in the quantum domain: A review of recent progress. Rep. Prog. Phys. 2018, 81, 074001. [Google Scholar] [CrossRef]
  13. Alvarez-Rodriguez, U.; Lamata, L.; Escandell-Montero, P.; Martín-Guerrero, J. Supervised Quantum Learning without Measurements. Sci. Rep. 2017, 7, 13645. [Google Scholar] [CrossRef] [PubMed]
  14. Wan, K.H.; Dahlsten, O.C.O.; Kristjánsson, H.; Gardner, R.; Kim, M.S. Quantum generalisation of feedforward neural networks. npj Quantum Inf. 2016, 3, 36. [Google Scholar] [CrossRef]
  15. Ricks, B.; Ventura, D. Training a Quantum Neural Network. In Proceedings of the Advances in Neural Information Processing Systems, Whistler, BC, Canada, 9–11 December 2003; Thrun, S., Saul, L., Schölkopf, B., Eds.; MIT Press: Cambridge, MA, USA, 2003; Volume 16. [Google Scholar]
  16. Hughes, T.J.R.; Franca, L.P.; Hulbert, G.M. A new finite element formulation for computational fluid dynamics: VIII. The galerkin/least-squares method for advective-diffusive equations. Comput. Methods Appl. Mech. Eng. 1989, 73, 173–189. [Google Scholar] [CrossRef]
  17. Yadav, S.; Ganesan, S. How Deep Learning performs with Singularly Perturbed Problems? In Proceedings of the 2019 IEEE Second International Conference on Artificial Intelligence and Knowledge Engineering (AIKE), Sardinia, Italy, 3–5 June 2019; pp. 293–297. [Google Scholar] [CrossRef]
  18. Yadav, S.; Ganesan, S. AI-augmented stabilized finite element method. arXiv 2022, arXiv:2211.13418. [Google Scholar]
  19. Yadav, S.; Ganesan, S. SPDE-ConvNet: Predict Stabilization Parameter for Singularly Perturbed Partial Differential Equation. ECCOMAS Congress 2022. Available online: https://www.scipedia.com/public/Yadav_Ganesan_2022a (accessed on 10 June 2023).
  20. Khodayi-Mehr, R.; Zavlanos, M. VarNet: Variational Neural Networks for the Solution of Partial Differential Equations. In Proceedings of the 2nd Conference on Learning for Dynamics and Control, Berkeley, CA, USA, 11–12 June 2020; Proceedings of Machine Learning Research: PMLR; 2020; Volume 120, pp. 298–307. Available online: http://proceedings.mlr.press/v120/khodayi-mehr20a/khodayi-mehr20a.pdf (accessed on 10 June 2023).
Table 1. Computational Time: Average of the training time taken by all the four examples in seconds.
N c is the number of elements used to discretize the domain Ω in x-direction.
Table 1. Computational Time: Average of the training time taken by all the four examples in seconds.
N c is the number of elements used to discretize the domain Ω in x-direction.
N c 10203040
SPDE-Net200316473598
SPDE-Q-Net4096154189
Table 2. L 2   E r r o r .
Table 2. L 2   E r r o r .
Examples
1234
Std. τ 1.32 × 10 5 6.77 × 10 6 1.41 × 10 5 3.63 × 10 6
Std. τ with | | u | | 4.04 × 10 5 1.63 × 10 5 1.76 × 10 5 3.85 × 10 6
SPDE-Net 7.03 × 10 6 5.08 × 10 6 1.33 × 10 5 3.63 × 10 6
SPDE-Q-Net 6.05 × 10 6 5.04 × 10 6 1.20 × 10 5 3.63 × 10 6
Table 3. Relative l 2 Error.
Table 3. Relative l 2 Error.
Examples
1234
PINN 4.85 × 10 1 4.85 × 10 + 1 8.69 × 10 1
Std. 1.17 × 10 1 1.36 × 10 1 8.02 × 10 2 4.63 × 10 2
Std. with | | u | | 3.35 × 10 1 3.06 × 10 1 9.68 × 10 2 4.90 × 10 2
SPDE-Net 6.1 × 10 2 9.9 × 10 2 6.99 × 10 2 4.80 × 10 2
SPDE-Q-Net 6.17 × 10 2 9.73 × 10 2 6.94 × 10 2 4.60 × 10 2
Table 4. L Error.
Table 4. L Error.
Examples
1234
Std. τ 9.07 × 10 5 7.29 × 10 5 1.13 × 10 4 7.60 × 10 6
Std. τ with | | u | | 1.32 × 10 4 8.70 × 10 5 1.47 × 10 4 1.06 × 10 5
SPDE-Net 3.99 × 10 5 4.45 × 10 5 9.08 × 10 5 7.96 × 10 5
SPDE-Q-Net 3.95 × 10 5 4.05 × 10 5 9.08 × 10 5 7.35 × 10 6
Table 5. H 1 Error.
Table 5. H 1 Error.
Examples
1234
Std. τ 1.30 × 10 3 6.74 × 10 4 1.43 × 10 3 3.28 × 10 4
Std. τ with | | u | | 3.65 × 10 3 1.51 × 10 3 1.72 × 10 3 3.69 × 10 4
SPDE-Net 6.01 × 10 4 4.90 × 10 4 1.53 × 10 3 4.1 × 10 4
SPDE-Q-Net 5.98 × 10 4 4.80 × 10 4 1.23 × 10 3 3.29 × 10 4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yadav, S. QPDE: Quantum Neural Network Based Stabilization Parameter Prediction for Numerical Solvers for Partial Differential Equations. AppliedMath 2023, 3, 552-562. https://doi.org/10.3390/appliedmath3030029

AMA Style

Yadav S. QPDE: Quantum Neural Network Based Stabilization Parameter Prediction for Numerical Solvers for Partial Differential Equations. AppliedMath. 2023; 3(3):552-562. https://doi.org/10.3390/appliedmath3030029

Chicago/Turabian Style

Yadav, Sangeeta. 2023. "QPDE: Quantum Neural Network Based Stabilization Parameter Prediction for Numerical Solvers for Partial Differential Equations" AppliedMath 3, no. 3: 552-562. https://doi.org/10.3390/appliedmath3030029

Article Metrics

Back to TopTop