Augmentation on the fly - an important tool in the training of neural networks

The most important part of machine learning is data. No matter how good the model and method of training, if the training sample is small or does not describe most of the cases of the real world, it will be almost impossible to achieve high quality work. At the same time, the task of creating training datasets is by no means simple and does not suit everyone, since in addition to the long and exhausting annotation of data by people, additional funding for this process is usually required.


Augmentation, or the generation of new data on the basis of the available data, makes it possible to quite simply and cheaply solve some of the problems with the training set using the available methods. In the case of neural networks, a widespread phenomenon has become to embed augmentation directly into the learning process, modifying the data of each era. However, a very small number of articles focuses on the importance of such an approach and what properties it brings to the learning process. In this article, we will examine what is useful can be extracted from augmentation on the fly, and how critical the choice of transformations and their parameters is within the framework of this approach.



 


Augmentation: offline or online?


โ€œโ€, โ€œโ€ โ€œโ€, . , . , โ€“ , , . .


, โ€œโ€ โ€“ . : , . , . , โ€œโ€ , . , , AlexNet, ImageNet 2012 . , , : CPU, , GPU. - 1.



 
. 1. -


, , . , , .



: . : , . , , , โ€“ , , .


: . , , ( ). , , , ! , , . 2 , .



 
. 2.



, โ€œ โ€ . , , . , โ€“ . , , (, AutoAugment Google, ). โ€“ . , , , .


. , : //. โ€“ , , . , , (AxialWarp), (RandStructElem) [1] (BrightnessMod). 3.



 
. 3.


, , . , , :


  • ;
  • , , , (, );
  • , . , , 0.9.

( ), , ( ). , , 10% SVHN (Street View House Numbers, , Google Street View), 1. SimpleNet [2] โ€“ LeNet , , . , . 4.


1. SimpleNet SVHN


,Error rate,Error rate,
SimpleNet, 310 .2.37%2.17%


 
. 4. ( ) SVHN


,


, . , , , . , , โ€“ .


, ResNet- [3] , ( , ). 5. SVHN. , , , (. 2). 2 โ€“ , ! , , .



 
. 5. ResNet- . โ€”


2. ResNet SVHN



w:
mul:
Error rate,Error rate,
ResNet, w: 40 ., mul: 1.63 .5.68%5.06%
ResNet, w: 18 ., mul: 730 .6.71%5.73%


. , , โ€“ , . , , .


, . โ€œโ€ , . , โ€“ , , .


ICMV 2018 (, . ): Gayer, A., Chernyshova, Y., & Sheshkus, A. (2019). Effective real-time augmentation of training dataset for the neural networks learning. Proceedings of SPIE, 11041, 110411I-110411I-8



  1. .. / .., .., .., .. // . 2018. . 32. โ„– 3
  2. Hasanpour S. H., Rouhani M., Mohsen F., Sabokrou M. Letโ€™s keep it simple, Using simple architectures to outperform deeper and more complex architectures // ArXiv e-prints https://arxiv.org/abs/1608.06037
  3. K. He, X. Zhang, S. Ren, and J. Sun. Deep residual learning for image recognition. // ArXiv e-prints https://arxiv.org/abs/1512.03385

All Articles