Next Article in Journal
Quality of Life Prediction on Walking Scenes Using Deep Neural Networks and Performance Improvement Using Knowledge Distillation
Next Article in Special Issue
Research Progress of Neural Synapses Based on Memristors
Previous Article in Journal
Intelligent Embedded Systems Platform for Vehicular Cyber-Physical Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Physics-Informed Recurrent Neural Network for RRAM Modeling

School of Microelectronics, Southern University of Science and Technology, Shenzhen 518055, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(13), 2906; https://doi.org/10.3390/electronics12132906
Submission received: 26 May 2023 / Revised: 14 June 2023 / Accepted: 30 June 2023 / Published: 2 July 2023

Abstract

:
Extracting behavioral models of RRAM devices is challenging due to their unique “memory” behaviors and rapid developments, for which well-established modeling frameworks and systematic parameter extraction processes are not available. In this work, we propose a physics-informed recurrent neural network (PiRNN) methodology to generate behavioral models of RRAM devices from practical measurement/simulation data. The proposed framework can faithfully capture the evolution of internal state and its impacts on the output. A series of modifications informed by the RRAM device physics are proposed to enhance the modeling capabilities. The integration strategy of Verilog-A equivalent circuits, is also developed for compatibility with existing general-purpose circuit simulators. The Verilog-A model can be easily adopted into the SPICE-type simulator for the circuit design with a variable step that differs from the training process. Numerical experiments with real RRAM devices data demonstrate the feasibility and advantages of the proposed methodology.

1. Introduction

Resistive random access memory (RRAM) has attracted much attention due to its advantages, including high density, low power consumption, long endurance, simple metal–insulator–metal construction, and good compatibility with CMOS technology since it was predicted by Leon O. Chua in 1971 [1] and first manufactured by HP Labs in 2008 [2]. Recent work has been dedicated to verifying the application of RRAM in neuromorphic computing, next-generation non-volatile memory technology and in-memory acceleration [3,4,5,6,7].
High-quality compact models of RRAM are a key for RRAM-based circuit design and optimization [8,9,10,11]. Existing modeling methods can be roughly divided into physics-based and empirical categories. Physical modeling provides excellent accuracy, but often at the cost of high-consumption computing power for its use in circuit simulation. Empirical modeling is more efficient but usually suffers from a poor generalization capacity.
To balance accuracy and efficiency, aphysics-informed compact modeling strategy is often applied, which incorporates device physics into the development of empirical-like compact models. However, developing high-quality physical-informed models is a time-consuming process, especially for emerging RRAM devices due to its much more complex conduction mechanism. Modern RRAM technologies use a wide variety of materials as dielectric layers to enable memristic properties. The switching of high- and low-resistance states in RRAM also depends heavily on the electrode materials. Therefore, it is increasingly challenging for physics-based model development to catch up with the rapid innovations in materials and structures in RRAM and other mem-device areas.
An effective strategy to combat these modeling challenges is to adopt neural network (NN) techniques since for emerging devices, high-quality measurement data are often more readily available than comprehensive physical understanding. In several application domains, machine learning has already demonstrated some promise [10,11,12,13,14]. Neural networks using back propagation can model the relationships between different variables with certain generalization capabilities. A recent methodological investigation indicates that using artificial neural networks (ANNs) for advanced transistors effectively reduces the demand of expertise and turn-around-time, and facilitates automatic parameter extraction [15]. Ref. [16] developed the NN-assisted compact model to capture cycle-to-cycle variations to evaluate the effectiveness of RRAM using advanced materials. To model the I-V curves of RRAM, Ref. [17] employs convolutional neural networks (CNNs) using I-V loop photos to extract the fitting parameters of the dynamic memdiode model. Although a purely data-driven NN model usually can fit the measurement data well, it still lacks a physical foundation, leading to poor generalization or prediction in practical use. Meanwhile, a purely data-driven neural network model would need a large amount of data, negating its benefit of rapid modeling. Therefore, neural networks with prior knowledge derived from physical or empirical understanding are proposed to enhance network performance. Physics-informed neural networks shows advantages of good generalizability on limited training datasets. Since physics is incorporated, the network model is effectively limited to a low-dimensional manifold and can be trained with a small amount of data [18]. Hence, the design of NNs should be informed or guided by an appropriate extent of device physics to avoid the low generalizability that a purely data-driven model might suffer from. A preliminary effort of using NNs to model thin-TFETs was reported [19], in which knowledge of different V g s and V d s characteristics guides the choices of excitation functions of the NN. The proposed model was proven to be accurate and predictive.
Different from conventional semiconductor devices, such as transistors, the output of RRAM, e.g., currents, are determined not only by the current inputs, but also by an internal hidden state that reflects the influence of previous inputs. In other words, the currents can be different even for the same voltage at different points of the hysteresis loop. As a consequence, conventional ANNs, in which the output depends solely on the current inputs, are not suitable for RRAM modeling. NNs with certain “memory” effects are needed to capture the unique physics of RRAM as well as other types of mem-devices.
In this paper, we investigate the possibility of using a recurrent neural network (RNN) with the embedded devices physics to model the steady-state I-V behavior of the metal-oxide-based RRAM devices. The RNN consists of a feedback loop to update the internal state to account for the input history, and the output depends on both the current input and the hidden state. The entire network structure is carefully crafted according to the RRAM device physics, rendering a physics-informed RNN (PiRNN) model that can mimic the physical behavior of a RRAM device. The training and testing procedures are then detailed, followed by integration with SPICE-type simulators. The Python model is converted into a Verilog-A equivalent circuit to employ a different step size in the simulation from that used in the training process or a variable step size. The feasibility and accuracy of the proposed PiRNN model are finally verified by experimental measurement data from realistic RRAM devices.

2. Fundamentals and Existing Models for RRAM

The conductive filament (CF) model is the mainstream theory for metal-oxide-based RRAM devices. The low-resistance state (LRS) and high-resistance state (HRS) of RRAM are associated with the formation and rupture of the CF controlled by the applied voltage waveforms [20,21,22,23,24]. As shown in Figure 1, the generation and recombination of oxygen vacancies and oxygen ions lead to the growth and rupture of CF, i.e., changes of the tunneling gap distance g. The internal state variable g of this theory can be viewed as the average distance from the tip of the CF to the top electrode (TE). And g controls the resistance of the RRAM through the electron tunneling conduction mechanism.
The conductance and current of RRAM are strongly correlated with the CF geometry evolution, which depends on the history of the applied voltages. During the SET process, oxygen vacancies and ions are generated with the applied bias. The continuous drift of the oxygen ions towards the top electrode produces a conductive filament linking the top electrode to the bottom electrode, eventually leading to the conversion of the HRS to the LRS. During the RESET process, the recombination of oxygen ions and oxygen vacancies leads to the continuous dissolution of CF and ultimately causes the transformation of LRS into HRS.
The relationship of the I-V can be expressed as [22]
I = I 0 e x p g g 0 s i n h V V 0
where I 0 , g 0 and V 0 are fitting parameters and the current is strongly correlated with the gap distance and the applied voltage. The evolution of g can be estimated as the following functions [22]:
d g d t = v 0 e x p q E a g k T e x p γ a 0 L q V k T e x p q E a r k T e x p γ a 0 L q V k T
γ = γ 0 β g g 1 3
where d g / d t represents the growth/dissolution velocity of the CF and the RHS is the difference between oxygen vacancy generation and recombination rates. Equation (2) describes the temporal evolution of g, where E a g and E a r are the activation energies for the generation and recombination of oxygen ions and oxygen vacancies, respectively. g contains g m a x and g m i n . During the SET process, the generation of the vacancies contributes to the growth of CF, and g gradually decreases to g m i n . During the RESET process, the recombination of the vacancies leads to the rupture of CF, and g gradually increases to g m a x . γ is a g-dependent local field acceleration parameter which considers the strong polarizability in high-k dielectrics [23], and L is the thickness of the oxide. v 0 , a 0 , γ 0 and g 1 are fitting parameters.
Existing RRAM models, e.g., [25,26], consist of two modules: one to describe the evolution of the internal state due to the applied bias and device parameters, often in the form of a differential equation or a dynamic sub-circuit, and the other to link the output current to the applied bias and the internal state of the device as shown in Figure 2. These internal state modules often demand substantial developmental efforts and complicate the parameter extraction process.
The existence of internal states complicates the development and deployment of RRAM compact models. First, the physical mechanisms governing internal states evolution remains less understood compared to other full-fledged devices, rendering the development of physics-based models a challenging and time-consuming task. Rapid innovations in RRAM technologies dictate frequent reworking of the models. Second, fully automatic parameter extraction is cumbersome for physics-based models. Third, interfacing with existing circuit simulators is not convenient. Some models even require modifications at the simulator level to handle the internal state variable. All these challenges hamper the quick turnaround in RRAM model development, undesirable for advanced design-technology co-optimization (DTCO) or path-finding activities.
Alternatively, one may choose to apply data-driven approaches to model RRAM devices, motivated by the fact that high-quality measurement data are often available earlier than comprehensive physical insights. However, purely data-driven models, such as look-up tables or simple NNs, usually do not provide adequate accuracy and generalizability for compact modeling [20]. In addition, the unique flux-dependent feature of RRAM renders the conventional ANNs without memory effect not applicable. New types of NNs are needed to account for the RRAM physics.

3. Physics-Informed RNN for RRAM Modeling

3.1. Basic Structure of PiRNN

We propose a physics-informed recurrent neural network (PiRNN) to model the steady-state behaviors of RRAM devices. Figure 3 depicts the structure of the proposed PiRNN model. The entire PiRNN takes in two consecutive voltages and the device width, then outputs the current flowing through the device.
Inspired by the two-module framework of existing RRAM models as shown in Figure 2, the PiRNN consists of an RNN module to model the internal state evolution associated to a given input voltage sequence, and an MLP module to generate the output current based on the internal state and the inputs. The RNN module, corresponding to the state module, takes the current voltage V, the previous voltage V p r e [27], and the device width W as the primary inputs and outputs the hidden state h, which is an m-bit vector. The hidden state is then fed back to form another input to the RNN, which is a critical operation that handles time-series data or the memory effect of RRAM devices in our application scenario. Given a sequence of input voltages, the goal of the RNN model is to determine a unique trajectory of the internal states of the RRAM devices.
The MLP module, on the other hand, is for determining the final output current, with a series of specialized designs guided by RRAM device physics. The MLP module is equivalent to the conduction module in Figure 2. By obtaining the internal state of the conductive filament, the current voltage, and the geometric parameters, it determines the current flowing through the device. In this work, we mainly follow the physical model reported in [28], which relates the output current to the input voltages and the internal state via Equation (1). Clearly, Equation (1) indicates that the current I depends on g and V in different (nonlinear) manners, representing different physical mechanisms. Informed by these physical insights, we choose to apply additionally an exponential activation function to the internal state h and a s i n h function to the input voltage V before sending them to the MLP. The width W is also squared to produce the conduction area input to the MLP based on the physical analysis in [29].
We adopt the following specific network structures for the PiRNN in this work. The RNN module has one layer with m  t a n h neurons whose outputs form the m-bit hidden state h in Figure 3. The MLP module consists of three hidden layers, with n 1 , n 2 , and n 3 neurons, respectively. The above structure results in m 2 + 9 m parameters for the RNN and m n 1 + n 1 n 2 + n 2 n 3 + 3 n 1 + n 2 + 2 n 3 + 1 parameters for the MLP.

3.2. RNN Cells

Figure 4 describes the structure of an RNN module in the proposed model and unrolled hidden layers through time. Each hidden layer in an RNN network can have an RNN cell with an n-bit hidden state. The input sequence updates the hidden state h t continually. Similar to an ANN, the RNN generates the hidden state for the current time step with the current input and the previous hidden state at each time step. The length of the input sequence determines the size of the unfolded RNN.

3.3. Training and Testing Procedures

Figure 5 illustrates the training process of the proposed model. The proposed PiRNN is trained with multiple sequences of steady-state I-V data for different device widths W acquired from experimental measurements or physics-based simulation. At a particular index of the sequence (called time t hereafter), the current voltage V t , the previous voltage V t 1 and the width W are fed to the model. These inputs, together with the previous state h t 1 and the previous output current I t 1 , determine the current hidden state h t . Then h t , V t and W, after the corresponding physics-informed transforms, are input into the MLP to generate the output current I t . The network equations read as
h t = f R N N ( V t , V t 1 , W , h t 1 , I t 1 )
I t = g M L P ( e x p ( h t ) , s i n h ( V t ) , W 2 )
where f R N N and g M L P refer to the manipulation of the inputs by the RNN module and the MLP module, respectively. In Equation (4), V t 1 stands for the previous input V p r e , which is the input of the RNN module.
Specifically, before each run, the hidden state and the output current ( h 0 , I 0 ) are initialized to zeros. All the weights and biases are randomly initialized within the range ( 1 / m , 1 / m ) , where m is the hidden state size. The first two voltages V 0 and V 1 , and W, are input to the RNN to generate the new state h 1 . Then, h 1 , V 1 and W are used to generate I 1 by the MLP. At the next time point, the pair ( h 1 , I 1 ) is fed back to the input of the RNN, together with V 1 and V 2 , to update the internal state to h 2 . The process repeats for all the data points in the training sequence. If there are multiple training data sequences for, e.g., different widths, they will be applied sequentially to train the model.
For validation and testing, one can send the corresponding voltage sequences into the trained model and evaluate the error between the generated currents and the true currents. The step size, i.e., the interval between two adjacent voltage points in the sequence, should be similar to the one used in the training process. The entire training and testing process can be fully automated utilizing existing machine learning software and hardware infrastructures.
In applications, the trained PiRNN model can be implemented in Verilog-A for integration with SPICE-type simulators [15,30], or used directly as a behavioral model in high-level memristor simulation platforms such as MNSim [31].

3.4. Verilog-A Implement

The RNN finite difference equations are transformed into differential equations in order to utilize a different step size in the simulation from that used in the training procedure or a variable step size. The equivalent model for these differential equations is ultimately implemented by Verilog-A [32].
At the given moment T = T t 1 + α h , a first-order expansion of continuous-time input voltages and hidden states is performed. 0 < α < 1 is the model parameter, and s is the time step of the training data. The input and hidden states of a discrete-time RNN can thus be expanded as follows:
V t = V + ( 1 α ) s V ˙
V t 1 = V α s V ˙
h t = h + ( 1 α ) s h ˙
h t 1 = h α s h ˙
During the implementation, the physical dimension width and the output current are treated as constants and thus involved in the calculation. Substituting Equations (6)–(9) into the hidden state Equation (4) yields the following. The physical dimension width and the output current are considered constants throughout implementation and are incorporated in the calculation. Equations (6)–(9) are substituted into the hidden state Equation (4) and the results are as follows:
h + ( 1 α ) s h ˙ = f R N N V + ( 1 α ) s V ˙ , V α s V ˙ , h α s h ˙ , W , I
The output is given by
I = g M L P e x p ( h + ( 1 α ) s h ˙ ) , s i n h ( V + ( 1 α ) s V ˙ ) , W 2
Figure 6 shows the Verilog-A implementation of Equations (10) and (11). In the Verilog-A model, the hidden states are represented by node voltages of the controlled current source, and n equals 20. Each current source is connected to the ground and implements Equation (12), which is the difference of the LHS and RHS of Equation (11). The current flowing through each current source equals zero. The result of the KCL equation for each node is the solution of Equation (11). The d d t function of Verilog-A is applied to generate V ˙ and h ˙ . Thus, the entire Verilog-A model takes the voltage and width as input since the previous voltage V i 1 is implemented as Equation (7), and generates the current flowing through the device:
I [ 1 : n ] = h + ( 1 α ) s h ˙ f R N N V + ( 1 α ) s V ˙ , V α s V ˙ , h α s h ˙ , W , I

4. Numerical Results

4.1. Model Configuration and Data Acquisition

The PiRNN model used in our numerical experiments consists of an RNN module with 1 hidden layer and m = 20 hidden state neurons, and a 3-layer MLP module with 256, 64, and 8 neurons in each layer, resulting in 28,565 parameters in total. The model is implemented by Python with the open-source deep learning framework PyTorch [33], with some network and training parameters reported in Table 1.
To collect data, we fabricated several HfO 2 RRAM devices with different widths, and measured their steady-state I-V data with the probe station as illustrated in Figure 7. The width, or side lengths, of the RRAM devices are respectively 3 μ m , 5 μ m , 10 μ m , and 15 μ m . A 501-point sequence is produced by the voltage sweeping from 2 V to 3 V with a uniform step size of 0.02 V. The sequence is generated by sweeping the voltage along 0 positive limit 0 negative limit 0 . The compliance current is set to 1 mA in the measurement of the I-V data. All the I-V curves are plotted in Figure 8.
To improve the numerical stability, the values of I in the sequences are converted to the logarithmic scale in the training process. The mean squared error (MSE) loss function with respect to the output current is used
M S E = 1 N i = 1 N I i I ^ i 2
where I ^ i is the prediction of PiRNN, I i is the measured data, and N is the length of the current sequence. The adaptive moment estimation (Adam) [34] is applied.

4.2. Result Analysis

We first confirm the accuracy of the proposed PiRNN for devices with a single width. We used two I-V data sequences collected with W = 5 μ m and W = 10 μ m to train two PiRNN models. For each data set, we pick the 251 odd data points for training and the 252 even data points for testing (with overlapping start and end points), respectively. As shown in Figure 9, the test loss drops rapidly in the first 200 epochs and converges at about 1500 epochs. The final training and test losses are 0.000226 and 0.000435 , respectively. The whole training process takes about 6 min. Figure 10 demonstrates good agreement between the predicted and measured currents in both linear and logarithmic scales for the two cases. The unique hysteresis loops of RRAM devices are well captured by the PiRNN model, owing to its modeling capability of internal state evolution.
Next, we show that the physics-informed modifications introduced in Section 3 indeed improve the model quality. Figure 10e,f show the IV curves from the PiRNN with the e x p ( h ) and the s i n h ( V ) functions removed, respectively, for the W = 10 μ m case. In comparison to the full PiRNN in Figure 10d at the same training epochs, one can see that removing these physics-informed modifications results in accuracy degradation, which in turn demonstrates the importance of including the relevant physical insights.
Furthermore, we prove that PiRNN can capture the dependence of the output current on device geometric parameters such as the width. We use the data sets of 3 μ m , 10 μ m , 15 μ m for training, and 5 μ m ) for testing. After applying the three training data sets sequentially, we achieve the final training and test losses of 0.187 and 0.196 , respectively. For the proposed model, the R-squared value is 0.98 . The whole training process takes about 20 min. Figure 11 compares the predicted and the measured currents at the testing data ( W = 5 μ m), which demonstrates reasonable accuracy.

4.3. Verilog-A Model

Finally, we implement the proposed PiRNN model in Verilog-A, build a simple test circuit with the Verilog-A model, and simulate it by Cadence Spectre. To validate the model, a piecewise-linear voltage source is applied in Figure 12 with α and s set to 0.1 and 0.005 , respectively. Figure 13 shows that the Verilog-A model correctly reproduces the results of the Python model.

5. Discussion

This paper presents a novel approach for modeling the steady-state behaviors of RRAM devices, utilizing a physics-informed recurrent neural network. The proposed model consists of two sub-modules, namely the MLP module and the RNN module. The RNN module, acting as the state module, captures the RRAM history-dependent features by leveraging its “memory” capacity to represent the internal state quantity g. On the other hand, the MLP module, analogous to the conduction module, generates the current output by incorporating the current internal state, voltage, and geometric parameters with physics-informed modifications. The accuracy of the proposed model is initially showcased for a device of single width in Figure 10a–d, and its capacity to apprehend the correlation between the output current and the geometrical parameters of the devices with a reasonable degree of precision is depicted in Figure 11. Figure 10e,f serve to demonstrate that the model’s quality is indeed enhanced through the incorporation of physics-informed adjustments. The model’s development and implementation are facile, and its accuracy is validated against measurement data obtained from practical RRAM devices. The PiRNN is implemented in Verilog-A for integration with SPICE-type simulators and thus allows a different step size in the simulation from that used in the training procedure or a variable step size.
Previous works report several physics-based compact models for RRAM, which commonly involve a dynamic module or sub-circuit to model the temporal evolution of the internal state (such as the conductive filament length) and the device responses (such as the current). These internal state modules often demand substantial developmental efforts and complicate the parameter extraction process. The proposed model distinguishes itself from prior models by integrating physics-informed neural networks (NNs) into the modeling process, thereby addressing the conventional challenge of developing high-quality physics-informed compact models in a time-efficient manner. Specifically, the model employs a specific neural network with fundamental physics embedded and reliable measurement data to rapidly model RRAM, obviating the need for crafting differential equations or dynamic circuits based on physical principles. Additionally, despite the limited training data, the model exhibits strong extrapolation capabilities, owing to the incorporation of recurrent neural networks (RNNs) and fundamental physics. Hence, the proposed PiRNN presents an effective method to cope with the swift progressions in RRAM technology, as well as the prompt development of RRAM models and the co-optimization of advanced design-technology co-optimization (DTCO) or path finding activities. Furthermore, the proposed approach is versatile and can be applied to other emerging devices that demonstrate history-dependent behavior.
This paper presents a proposed I-V model that incorporates the device geometry W. Nonetheless, the temperature during the switching process exerts a noteworthy influence on the growth and dissolution of CF. Consequently, temperature is an indispensable factor that warrants consideration in forthcoming research. Additionally, to meet the demands of circuit design, a charge-voltage, Q-V, model is formulated, utilizing a neural network and the C-V data obtained from the RRAM devices.

6. Conclusions

Motivated by the demand of rapid yet quality model development of emerging devices, we proposed a PiRNN model combining an RNN module and an MLP module for the behavioral modeling of RRAM devices. The model exploits the ”memory” effects of RNN to mimic the internal state evaluation of RRAMs and apply an array of physical insights to guide the model design for better modeling accuracy and capability. The model is validated against measurement data from practical RRAM devices and is also implemented in Verilog-A for SPICE integration. All the results confirm the feasibility and practicability of PiRNN. More thorough studies of the temperature and charge-voltage model will be conducted in the future.

Author Contributions

Conceptualization, Y.S. and Q.C.; methodology, Y.S.; software, Y.S.; validation, Y.S. formal analysis, J.L. and Y.L.; investigation, Y.S., J.L. and Y.L.; resources, J.L. and Y.L.; data curation, J.L. and Y.L.; writing—original draft preparation, Y.S.; writing—review and editing, Y.S., J.L., Y.L. and Q.C.; visualization, Y.S.; supervision, Q.C.; project administration, Q.C.; funding acquisition, Q.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported in part by the National Natural Science Foundation of China (NSFC) under the grants No. 62034007, No. 62141410, and in part by the Department of Science and Technology of Guangdong Province under the grant 2021B1101270003.

Data Availability Statement

Data is available on request due to restrictions.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chua, L. Memristor-the missing circuit element. IEEE Trans. Circuit Theory 1971, 18, 507–519. [Google Scholar] [CrossRef]
  2. Strukov, D.B.; Snider, G.S.; Stewart, D.R.; Williams, R.S. The missing memristor found. Nature 2008, 453, 80–83. [Google Scholar] [CrossRef] [PubMed]
  3. Veluri, H.; Chand, U.; Li, Y.; Tang, B.; Thean, A.V.Y. A Low-Power DNN Accelerator Enabled by a Novel Staircase RRAM Array. IEEE Trans. Neural Netw. Learn. Syst. 2021, 1–12. [Google Scholar] [CrossRef]
  4. Veluri, H.; Li, Y.; Niu, J.X.; Zamburg, E.; Thean, A.V.Y. High-Throughput, Area-Efficient, and Variation-Tolerant 3-D In-Memory Compute System for Deep Convolutional Neural Networks. IEEE Internet Things J. 2021, 8, 9219–9232. [Google Scholar] [CrossRef]
  5. Zhang, W.; Wang, S.; Li, Y.; Xu, X.; Dong, D.; Jiang, N.; Wang, F.; Guo, Z.; Fang, R.; Dou, C.; et al. Few-shot graph learning with robust and energy-efficient memory-augmented graph neural network (MAGNN) based on homogeneous computing-in-memory. In Proceedings of the 2022 IEEE Symposium on VLSI Technology and Circuits (VLSI Technology and Circuits), Honolulu, HI, USA, 12–17 June 2022; pp. 224–225. [Google Scholar] [CrossRef]
  6. Lan, J.; Li, Z.; Chen, Z.; Zhu, Q.; Wang, W.; Zaheer, M.; Lu, J.; Liang, J.; Shen, M.; Chen, P.; et al. Improved Performance of HfxZnyO-Based RRAM and its Switching Characteristics down to 4 K Temperature. Adv. Electron. Mater. 2023, 9, 2201250. [Google Scholar] [CrossRef]
  7. Lan, J.; Zhu, Q.; Zhang, Y.; Wang, W.; Zaheer, M.; Liang, J.; Shen, M.; Li, Z.; Chen, Z.; Wei, H.; et al. Zinc-Alloyed HFO2 Synaptic RRAM with Operating Voltage and Switching Energy Enhancement. In Proceedings of the 2022 China Semiconductor Technology International Conference (CSTIC), Shanghai, China, 20–21 June 2022; pp. 1–3. [Google Scholar]
  8. Xu, M.; Gao, B.; Xu, F.; Wu, W.; Tang, J.; Chen, J.; Qian, H. A Compact Model of Analog RRAM Considering Temperature Coefficient for Neural Network Evaluation. In Proceedings of the 2021 5th IEEE Electron Devices Technology & Manufacturing Conference (EDTM), Chengdu, China, 8–11 April 2021; pp. 1–3. [Google Scholar] [CrossRef]
  9. Krishnan, G.; Yang, L.; Sun, J.; Hazra, J.; Du, X.; Liehr, M.; Li, Z.; Beckmann, K.; Joshi, R.V.; Cady, N.C.; et al. Exploring Model Stability of Deep Neural Networks for Reliable RRAM-Based In-Memory Acceleration. IEEE Trans. Comput. 2022, 71, 2740–2752. [Google Scholar] [CrossRef]
  10. Prabhu, N.; Raghavan, N. Exploring RRAM Variability as Synapses on Inception Simulation Framework to Characterize the Prediction Accuracy and Power Estimation per Bit for Convolution Neural Network. In Proceedings of the 2020 IEEE International Symposium on the Physical and Failure Analysis of Integrated Circuits (IPFA), Singapore, 20–23 July 2020; pp. 1–5. [Google Scholar] [CrossRef]
  11. Huang, P.; Xiang, Y.; Zhao, Y.; Liu, C.; Gao, B.; Wu, H.; Qian, H.; Liu, X.; Kang, J. Analytic model for statistical state instability and retention behaviors of filamentary analog RRAM array and its applications in design of neural network. In Proceedings of the 2018 IEEE International Electron Devices Meeting (IEDM), San Francisco, CA, USA, 1–5 December 2018; pp. 40–44. [Google Scholar]
  12. Nassani, A.A.; Javed, A.; Rosak-Szyrocka, J.; Pilar, L.; Yousaf, Z.; Haffar, M. Major Determinants of Innovation Performance in the Context of Healthcare Sector. Int. J. Environ. Res. Public Health 2023, 20, 5007. [Google Scholar] [CrossRef]
  13. Prabhu, N.L.; Raghavan, N. Generalized convolution simulation stack for RRAM device based deep learning neural network. In Proceedings of the 2020 IEEE International Symposium on the Physical and Failure Analysis of Integrated Circuits (IPFA), Singapore, 20–23 July 2020; pp. 1–6. [Google Scholar]
  14. Sen, P.; Woods, W.H.; Sarkar, S.; Pratap, R.J.; Dufrene, B.M.; Mukhopadhyay, R.; Lee, C.H.; Mina, E.F.; Laskar, J. Neural-network-based parasitic modeling and extraction verification for RF/millimeter-wave integrated circuit design. IEEE Transa. Microw. Theory Technol. 2006, 54, 2604–2614. [Google Scholar] [CrossRef]
  15. Wang, J.; Kim, Y.H.; Ryu, J.; Jeong, C.; Choi, W.; Kim, D. Artificial Neural Network-Based Compact Modeling Methodology for Advanced Transistors. IEEE Trans. Electron Devices 2021, 68, 1318–1325. [Google Scholar] [CrossRef]
  16. Rohan, J.N.; Zhuang, P.; Nibhanupudi, S.T.; Banerjee, S.K.; Kulkarni, J.P. Neural Network Assisted Compact Model for Accurate Characterization of Cycle-to-cycle Variations in 2-D h-BN based RRAM devices. In Proceedings of the 2019 Device Research Conference (DRC), Ann Arbor, MI, USA, 23–26 June 2019; pp. 103–104. [Google Scholar]
  17. Aguirre, F.L.; Piros, E.; Kaiser, N.; Vogel, T.; Petzold, S.; Gehrunger, J.; Oster, T.; Hochberger, C.; Suñé, J.; Alff, L.; et al. Fast Fitting of the Dynamic Memdiode Model to the Conduction Characteristics of RRAM Devices Using Convolutional Neural Networks. Micromachines 2022, 13, 2002. [Google Scholar] [CrossRef]
  18. Karniadakis, G.E.; Kevrekidis, I.G.; Lu, L.; Perdikaris, P.; Wang, S.; Yang, L. Physics-informed machine learning. Nat. Rev. Phys. 2021, 3, 422–440. [Google Scholar] [CrossRef]
  19. Li, M.; İrsoy, O.; Cardie, C.; Xing, H.G. Physics-inspired neural networks for efficient device compact modeling. IEEE J. Explor. Solid-State Comput. Devices Circuits 2016, 2, 44–49. [Google Scholar] [CrossRef]
  20. Huang, P.; Liu, X.Y.; Chen, B.; Li, H.T.; Wang, Y.J.; Deng, Y.X.; Wei, K.L.; Zeng, L.; Gao, B.; Du, G.; et al. A physics-based compact model of metal-oxide-based RRAM DC and AC operations. IEEE Trans. Electron Devices 2013, 60, 4090–4097. [Google Scholar] [CrossRef]
  21. Yao, J.; Zhong, L.; Natelson, D.; Tour, J.M. Intrinsic resistive switching and memory effects in silicon oxide. Appl. Phys. A 2011, 102, 835–839. [Google Scholar] [CrossRef]
  22. Chen, P.Y.; Yu, S. Compact modeling of RRAM devices and its applications in 1T1R and 1S1R array design. IEEE Trans. Electron Devices 2015, 62, 4022–4028. [Google Scholar] [CrossRef]
  23. Yu, S.; Wong, H.S.P. Characterization and modeling of the conduction and switching mechanisms of HfOx based RRAM. MRS Online Proc. Library (OPL) 2014, 1631, 403. [Google Scholar] [CrossRef]
  24. Lim, E.W.; Ismail, R. Conduction Mechanism of Valence Change Resistive Switching Memory: A Survey. Electronics 2015, 4, 586–613. [Google Scholar] [CrossRef] [Green Version]
  25. Garcia-Redondo, F.; Gowers, R.P.; Crespo-Yepes, A.; López-Vallejo, M.; Jiang, L. SPICE compact modeling of bipolar/unipolar memristor switching governed by electrical thresholds. IEEE Trans. Circuits Syst. I Regul. Pap. 2016, 63, 1255–1264. [Google Scholar] [CrossRef] [Green Version]
  26. Messaris, I.; Serb, A.; Stathopoulos, S.; Khiat, A.; Nikolaidis, S.; Prodromakis, T. A data-driven verilog-a reram model. IEEE Trans. Comput.-Aided Design Integr. Circuits Syst. 2018, 37, 3151–3162. [Google Scholar] [CrossRef] [Green Version]
  27. Lin, A.S.; Liu, P.N.; Pratik, S.; Yang, Z.K.; Rawat, T.; Tseng, T.Y. RRAM Compact Modeling Using Physics and Machine Learning Hybridization. IEEE Trans. Electron Devices 2022, 69, 1835–1841. [Google Scholar] [CrossRef]
  28. Yu, S.; Gao, B.; Fang, Z.; Yu, H.; Kang, J.; Wong, H.S.P. A neuromorphic visual system using RRAM synaptic devices with Sub-pJ energy and tolerance to variability: Experimental characterization and large-scale modeling. In Proceedings of the 2012 International Electron Devices Meeting, San Francisco, CA, USA, 10–13 December 2012; pp. 10.4.1–10.4.4. [Google Scholar] [CrossRef]
  29. Zhong, L.; Jiang, L.; Huang, R.; De Groot, C. Nonpolar resistive switching in Cu/SiC/Au non-volatile resistive memory devices. Appl. Physics Lett. 2014, 104, 093507. [Google Scholar] [CrossRef] [Green Version]
  30. Deng, W.; Zhang, W.; Peng, Y.; Wu, W.; Huang, J.; Luo, Z. Compact modeling of metal-oxide TFTs based on artificial neural network and improved particle swarm optimization. J. Comput. Electron. 2021, 20, 1043–1049. [Google Scholar] [CrossRef]
  31. Xia, L.; Li, B.; Tang, T.; Gu, P.; Yin, X.; Huangfu, W.; Chen, P.Y.; Yu, S.; Cao, Y.; Wang, Y.; et al. MNSIM: Simulation platform for memristor-based neuromorphic computing system. In Proceedings of the 2016 Design, Automation Test in Europe Conference Exhibition (DATE), Dresden, Germany, 14–18 March 2016; pp. 469–474. [Google Scholar]
  32. Chen, Z.; Raginsky, M.; Rosenbaum, E. Verilog-A compatible recurrent neural network model for transient circuit simulation. In Proceedings of the 2017 IEEE 26th Conference on Electrical Performance of Electronic Packaging and Systems (EPEPS), San Jose, CA, USA, 15–18 October 2017; pp. 1–3. [Google Scholar] [CrossRef]
  33. Jain, A.; Awan, A.A.; Subramoni, H.; Panda, D.K. Scaling TensorFlow, PyTorch, and MXNet using MVAPICH2 for High-Performance Deep Learning on Frontera. In Proceedings of the 2019 IEEE/ACM Third Workshop on Deep Learning on Supercomputers (DLS), Denver, CO, USA, 17 November 2019; pp. 76–83. [Google Scholar]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. The formation (abc) and rupture (cba) of the CF between the top and the bottom electrode, which corresponds to the internal state g. g is the average tunneling gap distance between the tip of CF and TE, which determines the RRAM resistance.
Figure 1. The formation (abc) and rupture (cba) of the CF between the top and the bottom electrode, which corresponds to the internal state g. g is the average tunneling gap distance between the tip of CF and TE, which determines the RRAM resistance.
Electronics 12 02906 g001
Figure 2. Two-module framework for RRAM modeling.
Figure 2. Two-module framework for RRAM modeling.
Electronics 12 02906 g002
Figure 3. Structure of the proposed PiRNN, which includes an RNN module responsible for internal state evolution and an MLP module for generating output currents. Some RRAM physics-informed treatments are applied between the RNN and the MLP modules.
Figure 3. Structure of the proposed PiRNN, which includes an RNN module responsible for internal state evolution and an MLP module for generating output currents. Some RRAM physics-informed treatments are applied between the RNN and the MLP modules.
Electronics 12 02906 g003
Figure 4. Structure of the basic RNN cell and unrolled hidden layers through time.
Figure 4. Structure of the basic RNN cell and unrolled hidden layers through time.
Electronics 12 02906 g004
Figure 5. Training procedures for the proposed PiRNN model.
Figure 5. Training procedures for the proposed PiRNN model.
Electronics 12 02906 g005
Figure 6. The equivalent model implemented by Verilog-A.
Figure 6. The equivalent model implemented by Verilog-A.
Electronics 12 02906 g006
Figure 7. (a,b) Vertical and profile view of a single RRAM device. (c) TEM result From top to bottom, the material and thickness of each layer is Pt (20 nm)/Ti (20 nm)/ HfO 2 (5 nm)/Pt (20 nm)/Ti (5 nm). Devices of four different widths are fabricated, namely 3 μ m , 5 μ m , 10 μ m , and 15 μ m .
Figure 7. (a,b) Vertical and profile view of a single RRAM device. (c) TEM result From top to bottom, the material and thickness of each layer is Pt (20 nm)/Ti (20 nm)/ HfO 2 (5 nm)/Pt (20 nm)/Ti (5 nm). Devices of four different widths are fabricated, namely 3 μ m , 5 μ m , 10 μ m , and 15 μ m .
Electronics 12 02906 g007
Figure 8. Measurement data of RRAM devices with different widths.
Figure 8. Measurement data of RRAM devices with different widths.
Electronics 12 02906 g008
Figure 9. Training loss and testing loss for the single-width test ( W = 10 μ m).
Figure 9. Training loss and testing loss for the single-width test ( W = 10 μ m).
Electronics 12 02906 g009
Figure 10. Accuracy of PiRNN for single-width devices (501 points, 251 points for training and 252 points for testing). (a,b) Linear and log scale comparisons for W = 5 μ m . (c,d) Linear and log scale comparisons for W = 10 μ m . (e,f) PiRNN without e x p ( h ) and without s i n h ( V ) after 1500 training epochs.
Figure 10. Accuracy of PiRNN for single-width devices (501 points, 251 points for training and 252 points for testing). (a,b) Linear and log scale comparisons for W = 5 μ m . (c,d) Linear and log scale comparisons for W = 10 μ m . (e,f) PiRNN without e x p ( h ) and without s i n h ( V ) after 1500 training epochs.
Electronics 12 02906 g010
Figure 11. Comparison of I-V curves for the multi-width test. The model is trained with 3 μ m , 10 μ m , 15 μ m widths and tested with 5 μ m width.
Figure 11. Comparison of I-V curves for the multi-width test. The model is trained with 3 μ m , 10 μ m , 15 μ m widths and tested with 5 μ m width.
Electronics 12 02906 g011
Figure 12. Simulation result with the Verilog-A model.
Figure 12. Simulation result with the Verilog-A model.
Electronics 12 02906 g012
Figure 13. SPICE simulation result with the Verilog-A model comparing with the Python model. The circuit schematic is shown in the inset.
Figure 13. SPICE simulation result with the Verilog-A model comparing with the Python model. The circuit schematic is shown in the inset.
Electronics 12 02906 g013
Table 1. Details of the proposed model.
Table 1. Details of the proposed model.
ParametersQuantity
Model configuration(20, 256, 64, 8)
RNN parameters580
MLP parameters 27 , 985
Learning rate 0.001
Training Epochsless than 5000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sha, Y.; Lan, J.; Li, Y.; Chen, Q. A Physics-Informed Recurrent Neural Network for RRAM Modeling. Electronics 2023, 12, 2906. https://doi.org/10.3390/electronics12132906

AMA Style

Sha Y, Lan J, Li Y, Chen Q. A Physics-Informed Recurrent Neural Network for RRAM Modeling. Electronics. 2023; 12(13):2906. https://doi.org/10.3390/electronics12132906

Chicago/Turabian Style

Sha, Yanliang, Jun Lan, Yida Li, and Quan Chen. 2023. "A Physics-Informed Recurrent Neural Network for RRAM Modeling" Electronics 12, no. 13: 2906. https://doi.org/10.3390/electronics12132906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop