Evaluating Model Performance with Hard-Swish Activation Function Adjustments
RECPAD 2024 In the field of pattern recognition, achieving high accuracy is essential. While training a model to recognize different complex images, it is vital to fine-tune the model to achieve the highest accuracy possible. One strategy for fine-tuning a model involves changing its activation func...
Saved in:
Main Authors: | , , , , , , |
---|---|
Format: | Journal Article |
Language: | English |
Published: |
09-10-2024
|
Subjects: | |
Online Access: | Get full text |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Summary: | RECPAD 2024 In the field of pattern recognition, achieving high accuracy is essential.
While training a model to recognize different complex images, it is vital to
fine-tune the model to achieve the highest accuracy possible. One strategy for
fine-tuning a model involves changing its activation function. Most pre-trained
models use ReLU as their default activation function, but switching to a
different activation function like Hard-Swish could be beneficial. This study
evaluates the performance of models using ReLU, Swish and Hard-Swish activation
functions across diverse image datasets. Our results show a 2.06% increase in
accuracy for models on the CIFAR-10 dataset and a 0.30% increase in accuracy
for models on the ATLAS dataset. Modifying the activation functions in
architecture of pre-trained models lead to improved overall accuracy. |
---|---|
DOI: | 10.48550/arxiv.2410.06879 |