Page 458 - Kaleidoscope Academic Conference Proceedings 2024
P. 458

2024 ITU Kaleidoscope Academic Conference




               the entire image. The window slides after each
               operation and the features are learnt by the
               feature maps.
                             (N X N)* (f X f )= N – F + 1(1)
                  Equation 1 indicates the size of the  output
               matrix with no padding also known as the
               feature map matrix, in this study image matrix
               28*28  map 3*3 filter (28-3+1)=26, i.e. 26*26 is
               the feature map. Equation 2 presents The size of
               the output matrix with padding.
                          (NXN) * (fXf) = (N+2P–f)/(s+1)(2)
               Here p is padding and  s is stride. The
               convolution operation is defined as  Conv(m,n)
               =  l(x,y)  ⊗  f(x,y), where  ⊗  is  convolution     Fig6: Summary of proposed CNN Model.
               operation,  I(x,y) is expressing input image     A specific linear process called a convolution
               matrix, F(x,y) is expressing filter or kernel    layer is used to extract important information.
               function. So Convolution is a mathematical       To reduce the covariance shift and boost neural
               technique that accepts two inputs, such as an    network stability, batch  normalization is
               image matrix and a filter or kernel. The image   performed. By taking the batch mean away and
               matrix is a digital representation of  picture   dividing it by the batch standard  deviation, it
               pixels, and the filter is another matrix used to   normalizes the  output of an earlier activation
               process it. It can process any aspect of the image   layer.  By offering an abstracted version of the
               because the kernel is significantly smaller than   representation,  max pooling aids in reducing
               the image. This paper uses 3-by-3 filters. To be   over-fitting. To avoid overfitting, the Dropout
               employed in a layered  architecture with         layer randomly sets input units to 0 with  a
               numerous convolutional layers using kernels (or   frequency of rate at each step during training.
               filters) and a Pooling operation, each model     The sum of all inputs is maintained by scaling
               must first be trained,  followed by testing.     up non-zero inputs by 1/(1 - rate).
               Rotation, Width, Height, Shear, and Zoom
               variables are taken into  account for  data
               augmentation.   The accuracy of the model is
               boosted when the proper  values for these
               parameters are filled in. As indicated in Table 2,
               the CNN model's remapping parameter was
               deemed false.












                          Fig 5: Architecture CNN                  Table2:data augmentation argument
               We normalize the dataset's features but not the
               dataset's labels.  Feathers values between 0 and 255
               normalize to 0.  Pixel values are simply divided to
               255 for this.  Thus, machines can comprehend with
               ease. When a statistic stops improving, lowers the
               learning rate.  Once learning reaches  a plateau,
               models frequently gain by decreasing the learning
               rate by a factor of 2–10. This callback keeps track
               of a quantity, and it slows down learning if there    Table3:Learning rate argument
               hasn't been any improvement for a predetermined   RESULT ANALYSIS:
               amount  of  epochs. The proposed CNN Model's     For analyzing the impact of model precision, recall,
               layers are depicted in Fig6.                     F-1 score, and accuracy are to be considered.
                                                                Precision is a measure of a model’s accuracy in
                                                                classifying a sample as positive. Recall measures
                                                                the ability to detect positive  samples. F1-Score is
                                                                used to balance precision and recall. Accuracy is




                                                          – 414 –
   453   454   455   456   457   458   459   460   461   462   463