Bipolar morphological networks: a neuron without multiplication

Nowadays it is difficult to find a problem that has not yet been proposed to be solved by neural networks. And in many problems other methods are no longer even considered. In such a situation, it is logical that in pursuit of the β€œsilver bullet”, researchers and technologists offer more and more new modifications of neural network architectures, which should bring the applicants β€œhappiness for all, for nothing, and let no one go offended!” However, in industrial tasks it often turns out that the accuracy of the model mainly depends on the cleanliness, size and structure of the training sample, and the interface needs to be reasonable on a neural network model (for example, it is unpleasant when the logic answer should be a variable-length list).


Another thing is productivity, speed. Here the dependence on architecture is direct and quite predictable. However, not all scientists are interested. It is much more pleasant to think for centuries, epochs, to mentally aim for a century when magically the computing power will be unimaginable, and energy is extracted from the air. However, there are also enough mundane people. And it is important for them that the neural networks are more compact, faster and more energy-efficient right now. For example, this is important when working on mobile devices and in embedded systems where there is no powerful video card or you need to save battery. A lot has been done in this direction: here are small-sized integer neural networks, and the removal of excess neurons, and tensor decomposition decompositions, and much more.


We managed to remove the multiplications from the calculations inside the neuron, replacing them with additions and taking the maximum, although we left the opportunity to use multiplications and nonlinear operations in the activation function. We called the proposed model a bipolar morphological model of a neuron.



 


, , β€œ ” - . , , . . , . , , , .


, , , , . , , . . . Labor omnia vΔ«cit improbus et dΕ«rΔ«s urgΔ“ns in rΔ“bus egestās.


, β€” . 90- [1, 2]. . , [3], [4]. , , . [5], [6]. , . .


, , , , .



:


y(x,w)=Οƒ(βˆ‘i=1Nwixi+wN+1)


, xw, Οƒ.


(-), . , . , 4 , :


βˆ‘i=1Nxiwi=βˆ‘i=1Npi00xiwiβˆ’βˆ‘i=1Npi01xi|wi|βˆ’βˆ‘i=1Npi10|xi|wi+βˆ‘i=1Npi11|xi||wi|,



pikj={1,  (βˆ’1)kxi>0 and (βˆ’1)jwi>00, 


. :


M=maxj(xjwj)k=βˆ‘i=1NxiwiMβˆ’1


:


βˆ‘i=1Nxiwi=exp⁑{lnβ‘βˆ‘i=1Nxiwi}=exp⁑{ln⁑M(1+k)}=(1+k)exp⁑ln⁑M==(1+k)exp⁑{ln⁑(maxj(xjwj))}=(1+k)exp⁑maxjln⁑(xjwj)==(1+k)exp⁑maxj(ln⁑xj+ln⁑wj)=(1+k)exp⁑maxj(yj+vj)β‰ˆexp⁑maxj(yj+vj),


yjβ€” , vj=ln⁑wjβ€” . , , kβ‰ͺ1. 0≀k≀Nβˆ’1, , (k=0), β€” (k=Nβˆ’1). N. , β€” , . , , β€” , . - .


- . 1. ReLU 4 : . . , .


, , . , , . (, -), β€” .



. 1. .


, - :


BM(x,w)=exp⁑maxj(ln⁑ReLU(xj)+vj0)βˆ’exp⁑maxj(ln⁑ReLU(xj)+vj1)βˆ’βˆ’exp⁑maxj(ln⁑ReLU(βˆ’xj)+vj0)+exp⁑maxj(ln⁑ReLU(βˆ’xj)+vj1),



vjk={ln⁑|wj|,  (βˆ’1)kwj>0βˆ’βˆž, 


, . , . , , .



, , , : - ! , . ( 1) ( 2). ? , . , .


, -: - , , -, , . - : , , , .


, , , incremental learning β€” , . . - , . β€œβ€ β€” ( 1), β€” ( 2). β€œβ€ , , . , -, , , -.



MNIST


MNIST β€” , 60000 28 28. 10000 . 10% , β€” . . 2.



. 2. MNIST.


:


conv(n, w_x, w_y) β€” n w_x w_y;
fc(n) β€” n ;
maxpool(w_x, w_y) β€” max-pooling w_x w_y;
dropout(p) β€” dropout p;
relu β€” ReLU(x)=max(x,0);
softmax β€” softmax.


MNIST :


CNN1: conv1(30, 5, 5) β€” relu1 β€” dropout1(0,2) β€” fc1(10) β€” softmax1.


CNN2: conv1(40, 5, 5) β€” relu1 β€” maxpool1(2, 2) β€” conv2(40, 5, 5) β€” relu2 β€” fc1(200) β€” relu3 β€” dropout1(0,3) β€” fc2(10) β€” softmax1.


. 1. β€œβ€ . () ().


1. MNIST. β€” , β€” .


1,1, +2,2, +
CNN1-98,72-98,72-
CNN1conv142,4798,5138,3898,76
CNN1conv1 β€” relu1 β€” dropout1 β€” fc126,89-19,8694,00
CNN2-99,45-99,45-
CNN2conv194,9099,4196,5799,42
CNN2conv1 β€” relu1 β€” maxpool1 β€” conv221,2598,6836,2399,37
CNN2conv1 β€” relu1 β€” maxpool1 β€” conv2 β€” relu2 β€” fc110,0174,9517,2599,04
CNN2conv1 β€” relu1 β€” maxpool1 β€” conv2 β€” relu2 β€” fc1 β€” dropout1 β€” relu3 β€” fc212,91-48,7397,86

-, , - . , - , . , .


: . , . : - .


MRZ


MRZ- , (. . 3). 280 000 21 17 37 MRZ, .



. 3. MRZ .


CNN3: conv1(8, 3, 3) β€” relu1 β€” conv2(30, 5, 5) β€” relu2 β€” conv3(30, 5, 5) β€” relu3 β€” dropout1(0,25) β€” fc1(37) β€” softmax1.


CNN4: conv1(8, 3, 3) β€” relu1 β€” conv2(8, 5, 5) β€” relu2 β€” conv3(8, 3, 3) β€” relu3 β€” dropout1(0,25) β€” conv4(12, 5, 5) β€” relu4 β€” conv5(12, 3, 3) β€” relu5 β€” conv6(12, 1, 1) β€” relu6 β€” fc1(37) β€” softmax1.


2. β€œβ€ . () ().


, MNIST: -, , . - , - .


2. MRZ. β€” , β€” .


1,1, +2,2, +
CNN3-99,63-99,63-
CNN3conv197,7699,6483,0799,62
CNN3conv1 β€” relu1 β€” conv28,5999,4721,1299,58
CNN3conv1 β€” relu1 β€” conv2 β€” relu2 β€” conv33,6798,7936,8999,57
CNN3conv1 β€” relu1 β€” conv2 β€” relu2 β€” conv3 β€” relu3 β€” dropout1 β€” fc112,58-27,8493,38
CNN4-99,67-99,67-
CNN4conv191,2099,6693,7199,67
CNN4conv1 β€” relu1 β€” conv26,1499,5273,7999,66
CNN4conv1 β€” relu1 β€” conv2 β€” relu2 β€” conv323,5899,4270,2599,66
CNN4conv1 β€” relu1 β€” conv2 β€” relu2 β€” conv3 β€” relu3 β€” dropout1 β€” conv429,5699,0477,9299,63
CNN4conv1 β€” relu1 β€” conv2 β€” relu2 β€” conv3 β€” relu3 β€” dropout1 β€” conv4 β€” relu4 β€” conv534,1898,4517,0899,64
CNN4conv1 β€” relu1 β€” conv2 β€” relu2 β€” conv3 β€” relu3 β€” dropout1 β€” conv4 β€” relu4 β€” conv5 β€” relu5 β€” conv65,8398,0090,4699,61
CNN4conv1 β€” relu1 β€” conv2 β€” relu2 β€” conv3 β€” relu3 β€” dropout1 β€” conv4 β€” relu4 β€” conv5 β€” relu5 β€” conv6 -relu6 β€” fc14,70-27,5795,46


, , . , - . MNIST MRZ.


? , - . , (, ) . , β€” TPU, .


, , : , .


PS. ICMV 2019:
E. Limonova, D. Matveev, D. Nikolaev and V. V. Arlazarov, β€œBipolar morphological neural networks: convolution without multiplication,” ICMV 2019, 11433 ed., Wolfgang Osten, Dmitry Nikolaev, Jianhong Zhou, Ed., SPIE, Jan. 2020, vol. 11433, ISSN 0277-786X, ISBN 978-15-10636-43-9, vol. 11433, 11433 3J, pp. 1-8, 2020, DOI: 10.1117/12.2559299.



  1. G. X. Ritter and P. Sussner, β€œAn introduction to morphological neural networks,” Proceedings of 13th International Conference on Pattern Recognition 4, 709–717 vol.4 (1996).
  2. P. Sussner and E. L. Esmi, Constructive Morphological Neural Networks: Some Theoretical Aspects and Experimental Results in Classification, 123–144, Springer Berlin Heidelberg, Berlin, Heidelberg (2009).
  3. G. X. Ritter, L. Iancu, and G. Urcid, β€œMorphological perceptrons with dendritic structure,” in The 12th IEEE International Conference on Fuzzy Systems, 2003. FUZZ ’03., 2, 1296–1301 vol.2 (May 2003).
  4. G. X. Ritter and G. Urcid, β€œLattice algebra approach to single-neuron computation,” IEEE Transactions on Neural Networks 14, 282–295 (March 2003).
  5. H. Sossa and E. Guevara, β€œEfficient training for dendrite morphological neural networks,” Neurocomputing 131, 132–142 (05 2014).
  6. E. Zamora and H. Sossa, β€œDendrite morphological neurons trained by stochastic gradient descent,” in 2016 IEEE Symposium Series on Computational Intelligence (SSCI), 1–8 (Dec 2016).

All Articles