Page 197 - The-5th-MCAIT2021-eProceeding
P. 197

alternative in selecting a suitable weight initialization strategy that can produce consistent results over the wide
        range of datasets.

        2. Related Works

           The rise of research papers discussing weight initialization technique have led to various introduction of
        image filters in the deep learning environment. Many initialization techniques have been proposed such as zeros
        or identity initialization and random initialization (Li et al., 2020). Due to problems faced when applying the
        standard weight initialization strategy, this motivates other researchers to investigate other image filters such as
        PCA filter and LDA filter. LDA differs from PCA where, LDA aims to look for the components that maximize
        the class separation which requires class labels while, PCA aims to extract components that maximize the
        variance in the dataset (Alberti et al., 2017).
           PCA initialization network has been thoroughly investigated by various works such as PCANet proposed by
        (Chan et al., 2015). Other works, such as 2-dimensional PCA weight initialization with parametric equalization
        normalization by (Wang et al., 2020), implementing PCA-initialized LeNet-5 model for image classification
        (Ren et al., 2016) and vehicular image classification with PCA CNN proposed by (Soon et al., 2020). With the
        increasing number of PCA initialization in deep learning, this shows promising future in using PCA filters in
        the convolutional layer. In this work, variant of PCA filters will be introduced and applied as an initialization
        strategy on LeNet-5 model and AlexNet model. The results showed that, PCA variant initialization is able to
        obtain consistent results and is highly robust towards any changes in the employed image dataset.

        3. PCA Variant Initialization
           The generated PCA variant filters will be inserted in the first convolutional layer of LeNet-5 and AlexNet
        model  which  will  then  be  evaluated  on  numerous  image  datasets.  Suppose  that  there  are    input  training
        images, {   }      of dimension    × ℎ and the patch or sliding window size is    ×   .
                      =1
        3.1. Principal Component Analysis (PCA) filters

           The process of generating PCA filter will utilize rectangular window,     and it will be moved over the test
                                                                      ,  
        image B (   × ℎ) with a pre-defined step horizontally and then vertically to obtain sub-images,    . The test
                                                                                          ,  
        image B (   × ℎ) at horizontal position    ∈ {0, … ,    − 1} and vertical position    ∈ {0, … , ℎ − 1} while, the
        sliding window     (   ×   ) at position    ∈ {0, … ,    − 1} and position    ∈ {0, … , ℎ − 1}.
                        ,  
           The weighting of the pixels for each sub-image,     is a product of multiplying the pixel values and the
                                                      ,  
        specific window weights,     at the associated window locations. The window weights of rectangular window
                                ,  
            and corresponding sub-image     can be obtained based on equation (1) and equation (2) respectively,
                                       ,  
           ,  
                                                                                               (1)

                                                                                               (2)

           The  collected  non-overlapping  sub-images  of  the   -th  image,    ,    , … ,      ,  ℎ ̂ ∈ ℝ   ×    where  each   
                                                                      ,2
                                                                                                 ,  
                                                                           ̂
                                                                  ,1
                                                                 
                                                       ̂
        denotes the   -th vectorized patch in     and    =    − ( ) , ℎ = ℎ − ( ). Each image vector is then normalized
                                           ̂
                                         
                                                    2          2
        by  subtracting  its  respective  mean  to  ensure  that  the  image  vector  is  centered  which  results  to,     ̅ =
        (   ,    , … ,    ̅   ,  ℎ ̂) where each   ̅  is a mean removed patch. By constructing the same matrix for all input
              ̅
          ̅
                ,2
            ,1
                                     ,  
                     ̂
        images and assemble them together will results to,
                                                                                               (3)
           The PCA filters are then expressed as follows, assuming that there are    principal eigenvectors,
                                                                                               (4)
        E- Proceedings of The 5th International Multi-Conference on Artificial Intelligence Technology (MCAIT 2021)   [184]
        Artificial Intelligence in the 4th Industrial Revolution
   192   193   194   195   196   197   198   199   200   201   202