Next Article in Journal
Influence of Metal Gear Tooth Geometry on Load and Wear within Metal-Polymer Gear Pairs
Previous Article in Journal
25-hydroxyvitamin D3 Levels and Their Clinical Associations in a Polish Cohort of Systemic Sclerosis Patients: A Cross-Sectional Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RALR: Random Amplify Learning Rates for Training Neural Networks

1
School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
2
Yangtze Delta Region Institution (Quzhou), University of Electronic Science and Technology of China, Quzhou 324000, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2022, 12(1), 268; https://doi.org/10.3390/app12010268
Submission received: 13 December 2021 / Revised: 24 December 2021 / Accepted: 26 December 2021 / Published: 28 December 2021
(This article belongs to the Topic Machine and Deep Learning)

Abstract

:
It has been shown that the learning rate is one of the most critical hyper-parameters for the overall performance of deep neural networks. In this paper, we propose a new method for setting the global learning rate, named random amplify learning rates (RALR), to improve the performance of any optimizer in training deep neural networks. Instead of monotonically decreasing the learning rate, we expect to escape saddle points or local minima by amplifying the learning rate between reasonable boundary values based on a given probability. Training with RALR rather than conventionally decreasing the learning rate achieves further improvement on networks’ performance without extra consumption. Remarkably, the RALR is complementary with state-of-the-art data augmentation and regularization methods. Besides, we empirically study its performance on image classification tasks, fine-grained classification tasks, object detection tasks, and machine translation tasks. Experiments demonstrate that RALR can bring a notable improvement while preventing overfitting when training deep neural networks. For example, the classification accuracy of ResNet-110 trained on the CIFAR-100 dataset using RALR achieves a 1.34% gain compared with ResNet-110 trained traditionally.

1. Introduction

After decades of lower activity, studies in artificial neural networks were revitalized after a 2006 breakthrough [1] in the research on deep learning. Deep neural networks have been a tremendous practical success in many fields, such as image recognition [2], object detection [3], face recognition [4], object tracking [5], and speech recognition [6]. However, a deep neural network is a complex mathematical function, consisting of millions of parameters that need to be computed to solve a problem. In addition, the parameters of deep neural networks include weight parameters and hyper-parameters. Each hyper-parameter has its impact on training deep neural networks.
Deep neural networks are usually learned to perform a specific task, and the goal during training is to minimize the objective function f = R n R , typically updated by stochastic gradient descent (SGD). The basic procedure to optimize f is:
θ t + 1 = θ t η t f t ( θ t )
Here, θ t R n is the weight parameter vector at the t-th iteration, f t ( θ t ) is the loss gradient information obtained on a relatively small batch of training datasets at the t-th iteration, and η t is the learning rate at the t-th iteration.
According to Equation (1), we perceive that the learning rate is a crucial hyper-parameter that controls how much we are adjusting the weight parameter vector of deep neural networks concerning the loss gradient information obtained by training datasets. It is well known that a small learning rate will make the training procedure converge slowly and more prone to get stuck in saddle points or local minima, and a large learning rate will cause the training procedure to diverge.
Traditionally, the learning rate should be a single value that monotonically decreases during training at specific epochs. This paper describes the remarkable phenomenon that is randomly amplifying the learning rate during training is beneficial overall. Hence, we propose randomly magnifying the global learning rate between reasonable boundary values based on the traditional monotonically decaying way, named random amplify learning rates (RALR). The RALR effectively makes the networks’ performance near the optimal classification accuracy compared to the conventional monotonically decaying way. Furthermore, unlike the adaptive learning rates methods, the RALR procedure requires virtually no additional computation.
To better understand the potential profits of RALR, we visually plot the training and validation loss, validation accuracy, and the global learning rate changing of both our RALR method and traditional method while training the CIFAR-100 dataset using network ResNet-110 in Figure 1. The blue curve in Figure 1b shows the baseline of ResNet-110p; the baseline of ResNet-110 reaches the best classification accuracy of 73.94% after 300 epochs. In contrast, it is possible to achieve 2.11% gain in the same epochs using the RALR technique, and the orange curve in Figure 1 shows the validation accuracy while training using RALR, the network ResNet-110 reaches the best classification accuracy of 76.05% at the 291st epoch. In Figure 1a, we found that training with traditional strictly decreasing methods (blue line) is more prone to overfitting compared with training using RALR (orange line). As shown in Figure 1, enlarging the global learning rate may temporarily reduce the classification accuracy and increase the training and verification loss. Still, the classification accuracy perhaps reaches a higher level after the global learning rate is restored.
The contributions of this paper are:
  • A methodology for setting the global learning rates for training deep neural networks makes the performance near-optimal classification accuracy compared with traditional monotonically decaying methods, with essentially no additional computation. Numerical and graphical results show that our method outperforms the traditional learning rate decreasing way.
  • RALR is applicable to any optimizers, such as SGD with Nesterov momentum [7], adaptive gradient (Adagrad [8]), and adaptive moment estimation (Adam [9]).
  • RALR is complementary to existing data augmentation and regularization techniques, such as Mixup [10], Cutout [11], CutMix [12], and Label Smoothing [13].
  • RALR is demonstrated with various architectures on image classification tasks, fine-grained classification tasks, object detection tasks, and machine translation tasks.

2. Motivation

Training deep neural networks involves using an optimization algorithm to find the weight parameter vector to best map inputs and outputs. Many researchers discover how to accelerate the learning speed, reduce overfitting, and obtain optimal networks’ performance with deep learning models. The training procedure is challenging, not least because the optimization function is non-convex and can be falling into the local minima or saddle points problems, as shown in Figure 1b. Training with conventionally decaying way cannot receive better performance after 159 epochs (blue line). Yet, training with RALR is less prone to overfitting, which improves the networks’ performance continuously almost in the whole training process (orange line), and the network gains the best classification accuracy at the 291st epoch. The training procedure of RALR is more effective and can speed up convergence ability and mitigate against overfitting (see Figure 1b). The validation accuracy of the baseline reaches the highest 65.25% at the 139th epoch in the first stage. In comparison, the validation accuracy of RALR achieves 66.02% at the 31st epoch and reaches the highest 73.30% at the 142nd epoch in the first stage.
The global learning rate is a hyper-parameter that controls how much to change the model. Traditionally, the global learning rate decreases at specific epochs, i.e., the learning rate may become smaller and smaller, and the learning rate might be too small after some epochs, resulting in the optimizer sticking in saddle points or local minima that may slow down the training speed and restrict the network performance in a long time [14]. The x-axis on each subfigure of Figure 2 indicates the value of the activation units, and the y-axis indicates the number of activation units under this specific value. As shown in Figure 2a, the activation values of test images or the network parameters are substantially unchanged for the last 100 epochs, which indicates, the network cannot update effectively in the remaining training time. Increasing the learning rate between reasonable boundary values at the specific probability can potentially help the training escape saddle points and local minima and find flatter minima with good generalization performance by the leap forward of the network’s weights updates. Figure 2b shows that the network’s activation values and weight parameters are updated effectively with the RALR for the last 100 epochs. That means the network may restructure the parameter weight vector to gain near-optimal performance. Each column of Figure 2 indicates the activation values of different convolutional layers for one given image. The first, second, and last column indicate the activation values of the layer1, layer2, and layer3 of ResNet-110.
From the above discussion, if the training procedure does not improve the network performance anymore, we can amplify the learning rate between reasonable boundary values to accelerate the network’s training and achieve better performance. Increasing the learning rate allows the optimizer of deep neural networks to escape saddle points and local minima effectively. In this way, it is possible to further explore the optimization algorithm and leads the network to better generalization, as illustrated in Figure 2.
We propose the RALR method to alter the learning rate around the global one. Amplifying the learning rate at some epochs can help the optimization algorithm escape local minima or saddle points and quickly find the correct descent direction. Particularly, the larger learning rate will make the optimizer explore the larger hidden space, which indicates the network could be updated more effectively. As a result, we speed the training up and train the network more effectively in comparison with the previous one.

3. Related Works

Training deep neural networks is a challenging optimization task. Many researchers try to achieve better network performance and accelerate the training speed. Some works are described briefly in the section below.
Adaptive Learning Rates: In order to improve the convergence speed of gradient descent learning algorithms, momentum [7] is proposed to accelerate SGD in the relevant direction. Instead of utilizing the gradient of the current step individually to guide the direction, momentum also accumulates the gradient of the previous step. As one of the early adaptive methods, AdaGrad [8] is a dynamic learning rate decreasing algorithm. AdaGrad adapts different learning rates associated with different features, i.e., frequently occurring features updated using small learning rates and infrequently occurring features updated using large learning rates. RMSprop is introduced by Geoffrey Hinton in Lecture 6e of his Coursera Class (http://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf, accessed on 25 December 2021). RMSprop divides the learning rate by an exponentially decaying average of squared gradients. AMSGrad [15] pinpoints the exponential moving average of past squared gradients as a reason for the poor generalization behavior of earlier adaptive learning rate methods and uses the maximum of past squared gradients rather than the exponential average to update the parameters.
In recent years, several optimizers have been proposed. These include AdamW [16], which fixes weight decay in Adam; QHAdam [17], which averages a standard SGD step with a momentum SGD step; and AggMo [18], which combines multiple momentum terms; and others.
Cycling learning rates: CLR and SGDR: The idea of cycling increasing the global learning rate has been explored by algorithms such as SGDR [19] and CLR [20]. SGDR reset the global learning rate to a large value in a periodic manner. After this, the global learning rate declines to a small value following a cosine annealing schedule. In addition, CLR lets the global learning rate cyclically vary between reasonable boundary values. Leslie N. Smith [20] argued that we could estimate a reasonable learning rate by initially training the model with a very small learning rate and increasing the global learning rate linearly or exponentially at each iteration in CLR.
Unlike the above techniques, RALR amplifies the global learning rate based on the given probability. It prevents the model from overfitting by escaping saddle points or local minima. Further experiments demonstrate the effectiveness of the RALR in achieving better performance.
Others: Biler et al. [21] stressed that the learning rate of the gradient descent strongly affects performance in neural networks and propose the Alrao algorithm. Blier et al. believed that each unit or feature in the network gets its learning rate sampled from a random distribution spanning several orders of magnitude, hoping that enough units will get a close-to-optimal learning rate.
Unlike Alrao (the experimental results of Alrao show that Alrao cannot effectively improve the network performance on CIFAR and ImageNet datasets. Besides, the code of Alrao is not publicly available. Therefore, we are not able to provide the comparison results with ALRAO here), RALR is easy to implement and takes the same learning rate for all units or features of the network.

4. Method

4.1. Observation

From Equation (1), we have known that the learning rate is a crucial hyper-parameter for SGD. To characterize the surprising phenomenon of RALR, we use the class activation mapping (CAM) [22] to visualize the activation units of network ResNet-110 trained on the Tiny-ImageNet dataset (http://tiny-imagenet.herokuapp.com/, accessed on 25 December 2021), and a class activation map for a particular category indicates the discriminative image regions used by the convolutional neural network (CNN) to identify that category. Figure 3 shows the comparison of class activation mapping, where objects of RALR are more precise and more accurate than the original ones.

4.2. Schedule of RALR

RALR is conducted with a certain probability of p r in the training procedure, and we name this probability factor p r as the amplification probability factor. For a training epoch e, the probability of it undergoing learning rate amplification is p r , and the probability of it being kept unchanged is 1 p r . In this process, training epochs with RALR are generated.
The procedure of RALR for training deep neural networks is shown in Algorithm 1. RALR usually sets a random boundary interval [ l o w , h i g h ] , the amplification factor is set to p r . Assume our training procedure takes a new training epoch e, e is a natural number, we randomly initialize the probability p e for this epoch, if p e < p r , we amplify the learning rate with a random numerical value r a n d o m _ r a t e [ l o w , h i g h ] , i.e., the learning rate η e of this epoch e is η e = η o × r a n d o m _ r a t e , η o is the global learning rate according to conventionally decaying way. Otherwise, we use the original global learning rate η o in this epoch for training, i.e., η e = η o .
Algorithm 1 Schedule of RALR based on epoch for training deep neural networks.
Require: 
Model θ , N Training Samples ( x i , y i ) i = 1 N , Optimizer SGD, Momentum, Weight decay, Batch Size, Number of epochs e, Decrease learning rate epochs e d , Loss Function f ( θ ) . Initial learning rate η 0 , Gamma γ .
Require: 
Random interval factor [ l o w , h i g h ] , Amplification probability p r
Ensure: 
Initialize θ , η 0 , best Loss L * f ( θ )
1:
for  i [ 1 , e ] do
2:
   if  i e d  then
3:
      η 0 * = γ
4:
   end if
5:
   Initialize random probability for each epoch p i
6:
   if  p i p r  then
7:
     Amplification rate r a n d o m _ r a t e [ l o w , h i g h ]
8:
     Amplify learning rate η = η 0 × r a n d o m _ r a t e
9:
     Train for 1 epoch.
10:
     Evaluate new loss L f ( θ )
11:
   else
12:
     learning rate η = η 0
13:
     Train for 1 epoch.
14:
     Evaluate new loss L f ( θ )
15:
   end if
16:
   if  L L *  then
17:
     Update L * = L . Save checkpoint θ and L *
18:
   end if
19:
end for

5. Experiment

This section will demonstrate the RALR schedule’s effectiveness on various convolutional networks for some computer vision tasks. If not specified, the learning rate is randomly amplified at the probability of p r = 0.5 , the random interval factor is set to l o w = 1 , h i g h = 3 All experimental results are averaged over four runs and we report the standard deviation. Moreover, this section analyzes the amplification probability p r and random interval boundary values l o w and h i g h . All of the experiments are performed with PyTorch [23] on Tesla M40 or GeForce GTX 1080 GPUs.

5.1. RALR on Different Tasks

CIFAR-10 and CIFAR-100: Both of the CIFAR datasets [24] consist of 60,000, 32 × 32 color images. CIFAR-10 has ten classes, such as dog, cat, car, and boat, with 6000 images per class, including 5000 training images per class and 1000 validation images per class. There are 50,000 training images and 10,000 validation images in total. The CIFAR-100 dataset is similar the CIFAR-10, except it has 100 classes containing 600 images each. There are 500 training images and 100 validation images per class. CIFAR-100 requires much more fine-grained recognition compared to CIFAR-10, as some classes are very visually similar. For example, it contains five different species of flowers: orchids, poppies, roses, sunflowers, and tulips.
In the experiment on the CIFAR dataset, the images are first zero-padded with 4 pixels on each side to get the 40 × 40 pixel images. Then, 32 × 32 pixel images are cropped randomly. Besides, images are randomly mirrored horizontally at a probability of p = 0.5 .
We compare the classification accuracy of some different architectures on the CIFAR-10 dataset, including deep residual networks (ResNet) [25], dense network(DenseNet) [26], and wide residual network (WRN) [27]. The networks includs (1) ResNet-56 and ResNet-110: a ResNet with a depth of 56 and 110; (2) DenseNet-BC-100-12: a DenseNet with a depth of 100 and a growth rate of 12; and (3) WRN-28-10, a WRN with a depth of 28 and a widening factor of 10.
We first show the feature embedding in Figure 4, which is the results of network ResNet-110 updated by SGD and trained on the CIFAR-10 dataset with the cross-entropy loss. Compare with the baseline one, the model trained using the RALR technique brings stronger intra-class compactness and larger inter-class separability. This means our RALR strategy can efficiently improve the network performance.
We trained for 300 epochs on the CIFAR-10 dataset for the architectures listed in Table 1. For all architectures, we trained with batches of 128 images using SGD. The hyper-parameter of Nesterov momentum is set to 0.9 , weight decay is 1 × 10 4 , and the global learning rate is initially set to 0.1 and decreased by 10× at the 150th and 225th epoch. In Table 1, we observe that the RALR technique consistently outperforms the baseline scheme.
We trained for 300 epochs on the CIFAR-100 dataset for architectures listed in Table 2. For ResNets and WRNs, we trained with batches of 128 images using SGD, and we trained with batches of 64 images using SGD for DenseNet. The hyper-parameter of Nesterov momentum is set to 0.9, weight decay is 1 × 10 4 , and the global learning rate is initially set to 0.1 and decreased by 10× at the 150th and 225th epoch for all architectures. In Table 2, we observe that the RALR technique consistently outperforms the baseline scheme. For example, our approach achieves 75.87% accuracy when applying network ResNet-110, which is a 1.34% improvement over baseline.
Fine-grained classification: The fine-grained image classification task focuses on differentiating between hard to distinguish object classes, such as species of birds, flowers, or animals, and identifying the makes or models of vehicles. Since collecting fine-grained samples often requires expert-level domain knowledge, most fine-grained datasets are hard to extend to a large scale, i.e., the fine-grained datasets are usually insufficient to gain reliable classification accuracy of deep neural networks when adopting the random initialization method. In the fine-grained classification tasks, the deep learning models are always fine-tuned on the parameters trained on ImageNet.
We test the effectiveness of RALR on several fine-grained benchmark datasets: CUB-200-2011 [28], StandfordDogs [29], and StandfordCars [30]. In the experiments, we evaluate the performance of the fine-grained classification tasks above using ResNet-50 pre-trained on ImageNet, and the model is fine-tuned by the optimizer SGD. ’The model is trained for 95 epochs, and the learning rate is initially set to 0.001 and decayed at 30th, 60th, and 90th epoch. The comparison result is shown in Table 3. For all the cases in Table 3, models trained using RALR achieve significant boosts over baseline.
Object Detection: Object detection is a task to predict the bounding boxes of each object of an image, with associated real-value confidence. SSD [31] is one object detection architecture. In this section, we use SSD (https://github.com/sgrvinod/a-PyTorch-Tutorial-to-Object-Detection, accessed on 25 December 2021) to evaluate the effectiveness of RALR on object detection. For comparison, the backbone of our experiments is VGG-16 and using VOC 2007 and VOC 2012 as the training dataset simultaneously. The results are shown in Table 4. RALR can achieve further detection performance compared with the baseline one even trained with the network VGG-16 pre-trained on ImageNet-2012.
Machine Translation - WMT 2014: We evaluate the capability of RALR technique on the standard WMT 2014 English–German dataset consisting of about 4.5 million sentence pairs. Each training batch contained a set of sentence pairs containing approximately 25,000 source tokens and 25,000 target tokens. The backbone is from the paper [32]. Table 5 shows the comparison results (https://github.com/sgrvinod/a-PyTorch-Tutorial-to-Machine-Translation/, accessed on 25 December 2021). The results in Table 5 show that the RALR technique outperforms the baseline and even uses the factor l o w = 1 , h i g h = 3 , and p r = 0.5 which gained in computer vision tasks.

5.2. RALR on Different Optimizers

To verify the generalization of our proposed method, we do some experiments with ResNet-110 on the CIFAR-100 dataset. We provide the comparison results on four different optimizers: SGD, SGD with momentum, Adam, and Adagrad in Table 6.
Specifically, we initially set the learning rate to 0.001 and decayed at the 150th epoch for optimizer Adam (besides, we also test the Adam with other hyper-parameters, but the network performance cannot be improved), and we initially set the learning rate to 0.1 and decayed at the 150th epoch for optimizer Adagrad. For optimizer SGD, the hyper-parameters are the sam as the CIFAR-100 in Section 5.1.

5.3. RALR’s Effectiveness on Other Techniques

To verify that RALR is complementary to existing popular data augmentation and regularization techniques, we do some experiments with ResNet-56 on the CIFAR100 dataset. We compare our proposed method with four state-of-the-art methods: Cutout, Mixup, CutMix, and Label Smoothing.
As seen in Table 7, we observe that if we combine the techniques above with RALR methods, the classification performances are outperformed than when we use these methods alone.
Specifically, we select the hole size of 8 × 8 pixels for the CIFAR-100 dataset when training on the full datasets for Cutout, which is the best parameters adapted in [11]. For Mixup, we tested the hyper-parameter α = 1.0 . p r o b = 0.5 and α = 1.0 are adapted in CutMix. We use s m o o t h i n g = 0.1 for Label Smoothing technique.

5.4. Ablation Study

Effect of the random interval factor: We explore the effect of the random interval factor l o w and h i g h according to the classification performance. We change the random interval factor setting l o w from 1 to 5 and h i g h from 2 to 9; the probability to take RALR method is fixed to p r = 0.5 . In addition, we evaluate the classification performance of network ResNet-110 on the CIFAR-100 dataset using the above factors. Partial validation accuracy is demonstrated in Table 8. As shown in Table 8, with the increase of random factor l o w , the classification accuracy decreases after l o w 2 . With the increase of random factor h i g h when setting the random factor to l o w = 1 , the accuracy raises and reaches the highest when h i g h = 3 , and then the accuracy decreases.
Effect of the probability: Another important hyper-parameter of RALR is amplification probability p r . We do some experiments to explore the influence of probability p r with a different threshold. We set amplification factor p r from 0.1 to 0.9 , and evaluate the accuracy of network ResNet-110 on the CIFAR-100 dataset. As illustrated in Table 9, the accuracy reaches the highest at amplification probability p r = 0.5 , and variance reaches the lowest at the same time. The variance reaches the lowest and accuracy reaches the highest when p r = 0.5 , l o w = 1 , and h i g h = 3 .

6. Conclusions

In this work, we propose a new learning rate changing policy, where the global learning rate randomly amplifies at a certain probability between reasonable boundaries, obtaining a near-optimal network performance compared to the traditional monotonically decreasing way. The experimental results demonstrate that our RALR method achieves consistent improvements in various tasks and datasets. The RALR policy is easy to implement and, unlike adaptive learning rate methods, incurs essentially no additional computational expense. This work is limited by not entirely searching all pf the applications for the RALR methods yet. We expect to determine if similar policies work for training networks over different research domains in future work.

Author Contributions

Conceptualization, M.L. (Ming Liu) and J.D.; methodology, J.D.; validation, M.L. (Minghui Liu), T.X., and X.C.; formal analysis, J.D. and H.G.; writing—original draft preparation, J.D.; writing—review and editing, X.C. and X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RALRRandom amplify learning rate
LRLearning rate
SGDStochastic gradient descent
CAMClass activation mapping

References

  1. Hinton, G.E.; Osindero, S.; Teh, Y.W. A Fast Learning Algorithm for Deep Belief Nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef] [PubMed]
  2. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  3. Zhang, S.; Wen, L.; Bian, X.; Lei, Z.; Li, S.Z. Single-Shot Refinement Neural Network for Object Detection. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  4. Wang, Y.; Ji, X.; Zhou, Z.; Wang, H.; Li, Z. Detecting Faces Using Region-based Fully Convolutional Networks. arXiv 2017, arXiv:1709.05256. [Google Scholar]
  5. Li, B.; Yan, J.; Wu, W.; Zhu, Z.; Hu, X. High Performance Visual Tracking With Siamese Region Proposal Network. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  6. Shaham, T.R.; Dekel, T.; Michaeli, T. SinGAN: Learning a Generative Model From a Single Natural Image. In Proceedings of the 17th IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019. [Google Scholar]
  7. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the importance of initialization and momentum in deep learning. In Proceedings of the 30th International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  8. Duchi, J.C.; Hazan, E.; Singer, Y. Adaptive Subgradient Methods for Online Learning and Stochastic Optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  9. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  10. Zhang, H.; Cissé, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond Empirical Risk Minimization. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  11. Devries, T.; Taylor, G.W. Improved Regularization of Convolutional Neural Networks with Cutout. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  12. Yun, S.; Han, D.; Chun, S.; Choe, J.; Yoo, Y. CutMix: Regularization Strategy to Train Strong Classifiers with Localizable Features. In Proceedings of the 17th IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27–28 October 2019. [Google Scholar]
  13. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  14. Dauphin, Y.N.; Pascanu, R.; Gülçehre, Ç.; Cho, K.; Ganguli, S.; Bengio, Y. Identifying and attacking the saddle point problem in high-dimensional non-convex optimization. In Proceedings of the 28th Annual Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  15. Reddi, S.J.; Kale, S.; Kumar, S. On the Convergence of Adam and Beyond. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  16. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  17. Ma, J.; Yarats, D. Quasi-hyperbolic momentum and Adam for deep learning. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  18. Lucas, J.; Sun, S.; Zemel, R.; Grosse, R. Aggregated Momentum: Stability Through Passive Damping. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  19. Loshchilov, I.; Hutter, F. SGDR: Stochastic Gradient Descent with Warm Restarts. In Proceedings of the 5th International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  20. Smith, L.N. Cyclical Learning Rates for Training Neural Networks. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision, Santa Rosa, CA, USA, 24–31 March 2017. [Google Scholar]
  21. Blier, L.; Wolinski, P.; Ollivier, Y. Learning with Random Learning Rates. In Proceedings of the European Conference on Machine Learning and Knowledge Discovery in Databases, Würzburg, Germany, 16–20 September 2019. [Google Scholar]
  22. Zhou, B.; Khosla, A.; Lapedriza, À.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  23. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037. [Google Scholar]
  24. Krizhevsky, A.; Hinton, G. Learning Multiple Layers of Features from Tiny Images; Technical Report; Computer Science Department, University of Toronto: Toronto, ON, USA, 2009. [Google Scholar]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 29th IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  26. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 June 2017. [Google Scholar]
  27. Zagoruyko, S.; Komodakis, N. Wide Residual Networks. In Proceedings of the 27th British Machine Vision Conference, York, UK, 19–22 September 2016. [Google Scholar]
  28. Wah, C.; Branson, S.; Welinder, P.; Perona, P.; Belongie, S. The Caltech-UCSD Birds-200-2011 Dataset; Technical Report; California Institute of Technology: Pasadena, CA, USA, 2011. [Google Scholar]
  29. Khosla, A.; Jayadevaprakash, N.; Yao, B.; Fei-Fei, L. Novel Dataset for Fine-Grained Image Categorization. In Proceedings of the First Workshop on Fine-Grained Visual Categorization, IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 1–2 June 2011. [Google Scholar]
  30. Krause, J.; Stark, M.; Deng, J.; Fei-Fei, L. 3D Object Representations for Fine-Grained Categorization. In Proceedings of the 4th International IEEE Workshop on 3D Representation and Recognition (3dRR-13), Sydney, Australia, 1–8 December 2013. [Google Scholar]
  31. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the 14th European Conference on Computer Vision; Amsterdam, The Netherlands, 11–14 October 2016, 2016. [Google Scholar]
  32. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
Figure 1. Training and validation loss, validation accuracy, and changing of learning rate while training the CIFAR-100 dataset using network ResNet-110. (b) The orange curve shows the validation accuracy of the new learning rate policy. (a) Loss. (b) Validation accuracy. (c) Learning rate.
Figure 1. Training and validation loss, validation accuracy, and changing of learning rate while training the CIFAR-100 dataset using network ResNet-110. (b) The orange curve shows the validation accuracy of the new learning rate policy. (a) Loss. (b) Validation accuracy. (c) Learning rate.
Applsci 12 00268 g001
Figure 2. The activation values changing of a network ResNet-110 trained for 300 epochs. The blue column indicates activation values of the network ResNet-110 obtained at the 200th epoch, and the red column indicates activation values of the network ResNet-110 obtained at the 300th epoch. (a) Baseline. (b) RALR.
Figure 2. The activation values changing of a network ResNet-110 trained for 300 epochs. The blue column indicates activation values of the network ResNet-110 obtained at the 200th epoch, and the red column indicates activation values of the network ResNet-110 obtained at the 300th epoch. (a) Baseline. (b) RALR.
Applsci 12 00268 g002
Figure 3. The first line shows the part of the original images of the Tiny-ImageNet test dataset. The second line shows the localized class-discriminative regions with ResNet-110 training with the traditional way. The third line shows the localized class-discriminative regions with ResNet-110 training with RALR.
Figure 3. The first line shows the part of the original images of the Tiny-ImageNet test dataset. The second line shows the localized class-discriminative regions with ResNet-110 training with the traditional way. The third line shows the localized class-discriminative regions with ResNet-110 training with RALR.
Applsci 12 00268 g003
Figure 4. The feature embedding of ResNet-110 on the CIFAR-10 dataset. The network is trained with the cross-entropy loss and updated by SGD. Each color on the picture represents a category of the CIFAR-10 dataset. (a) Baseline. (b) Random learning rates.
Figure 4. The feature embedding of ResNet-110 on the CIFAR-10 dataset. The network is trained with the cross-entropy loss and updated by SGD. Each color on the picture represents a category of the CIFAR-10 dataset. (a) Baseline. (b) Random learning rates.
Applsci 12 00268 g004
Table 1. Validation accuracy (%) over different architectures on CIFAR-10.
Table 1. Validation accuracy (%) over different architectures on CIFAR-10.
ModelBaselineRALR
ResNet-2091.41 ± 0.0892.28 ± 0.05
ResNet-5693.77 ± 0.0694.35 ± 0.05
ResNet-11094.42 ± 0.0695.05 ± 0.04
WRN-28-1095.82 ± 0.0396.36 ± 0.03
Table 2. Validation accuracy (%) over different architectures on CIFAR-100.
Table 2. Validation accuracy (%) over different architectures on CIFAR-100.
ModelBaselineRALR
ResNet-2069.88 ± 0.0670.48 ± 0.03
ResNet-5673.69 ± 0.0774.54 ± 0.04
ResNet-11074.53 ± 0.4175.87 ± 0.07
WRN-28-1078.81 ± 0.2479.54 ± 0.11
DesNet-BC-100-1277.22 ± 0.1577.95 ± 0.04
Table 3. Validation accuracy (%) of ResNet-50 on different fine-grained tasks.
Table 3. Validation accuracy (%) of ResNet-50 on different fine-grained tasks.
CUBStanford CarsStanford Dogs
+RALRAcc.+RALRAcc.+RALRAcc.
RseNet-50×85.32 ± 0.23×90.18 ± 0.18×84.15 ± 0.25
86.36 ± 0.0890.83 ± 0.0584.72 ± 0.13
Table 4. Comparison results of SSD on PASCAL VOC 2012.
Table 4. Comparison results of SSD on PASCAL VOC 2012.
Backbone+RALRAcc.
SSD-VGG16×77.0
77.7
Table 5. Comparison results of WMT 2014.
Table 5. Comparison results of WMT 2014.
TokenizationCasedBLEU
Baseline+RALR
13aYes25.125.5
No25.625.9
InternationalYes25.926.1
No26.326.7
Table 6. Validation accuracy (%) over different optimizer based on ResNet-110 trained on CIFAR-100.
Table 6. Validation accuracy (%) over different optimizer based on ResNet-110 trained on CIFAR-100.
Optimizer+RALRAcc.
SGD-×67.48 ± 0.35
69.99 ± 0.15
+ momentum×74.53 ± 0.41
75.87 ± 0.07
Adam×72.55 ± 0.06
73.55 ± 0.09
Adagrad×64.56 ± 0.80
65.01 ± 0.86
Table 7. Validation accuracy (%) over different techniques based on ResNets trained on CIFAR-100.
Table 7. Validation accuracy (%) over different techniques based on ResNets trained on CIFAR-100.
Method+RALRResNet-56ResNet-110
+baseline×73.69 ± 0.0774.53 ± 0.41
74.54 ± 0.0475.87 ± 0.07
+Cutout×74.55 ± 0.0575.34 ± 0.12
75.02 ± 0.0276.11 ± 0.05
+Mixup×75.66 ± 0.0777.89 ± 0.05
76.02 ± 0.0478.26 ± 0.05
+CutMix×76.67 ± 0.1177.71 ± 0.13
77.13 ± 0.0378.52 ± 0.07
+Label Smoothing×74.05 ± 0.1175.01 ± 0.11
74.94 ± 0.0775.98 ± 0.04
Table 8. Impact on parameters of different l o w , h i g h with fixed probability of p r = 0.5 when taking RALR method.
Table 8. Impact on parameters of different l o w , h i g h with fixed probability of p r = 0.5 when taking RALR method.
ModelLowHighAcc.
ResNet-1101275.37 ± 0.82
1375.87 ± 0.07
1575.35 ± 0.22
1770.17 ± 1.58
2375.71 ± 0.57
2575.81 ± 0.11
2775.13 ± 0.45
3571.88 ± 2.11
3771.79 ± 3.01
Table 9. Impact on parameters of probability with fixed l o w and h i g h when taking RALR method.
Table 9. Impact on parameters of probability with fixed l o w and h i g h when taking RALR method.
ModelProbLowHighAcc.
ResNet-1100.31374.44 ± 0.87
2573.98 ± 1.36
0.41375.31 ± 0.11
2575.26 ± 0.23
0.51375.87 ± 0.07
2575.81 ± 0.11
0.61375.53 ± 0.31
2575.40 ± 0.62
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Deng, J.; Gong, H.; Liu, M.; Xie, T.; Cheng, X.; Wang, X.; Liu, M. RALR: Random Amplify Learning Rates for Training Neural Networks. Appl. Sci. 2022, 12, 268. https://doi.org/10.3390/app12010268

AMA Style

Deng J, Gong H, Liu M, Xie T, Cheng X, Wang X, Liu M. RALR: Random Amplify Learning Rates for Training Neural Networks. Applied Sciences. 2022; 12(1):268. https://doi.org/10.3390/app12010268

Chicago/Turabian Style

Deng, Jiali, Haigang Gong, Minghui Liu, Tianshu Xie, Xuan Cheng, Xiaomin Wang, and Ming Liu. 2022. "RALR: Random Amplify Learning Rates for Training Neural Networks" Applied Sciences 12, no. 1: 268. https://doi.org/10.3390/app12010268

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop