Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
169 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Image edge enhancement for effective image classification (2401.07028v1)

Published 13 Jan 2024 in cs.CV

Abstract: Image classification has been a popular task due to its feasibility in real-world applications. Training neural networks by feeding them RGB images has demonstrated success over it. Nevertheless, improving the classification accuracy and computational efficiency of this process continues to present challenges that researchers are actively addressing. A widely popular embraced method to improve the classification performance of neural networks is to incorporate data augmentations during the training process. Data augmentations are simple transformations that create slightly modified versions of the training data and can be very effective in training neural networks to mitigate overfitting and improve their accuracy performance. In this study, we draw inspiration from high-boost image filtering and propose an edge enhancement-based method as means to enhance both accuracy and training speed of neural networks. Specifically, our approach involves extracting high frequency features, such as edges, from images within the available dataset and fusing them with the original images, to generate new, enriched images. Our comprehensive experiments, conducted on two distinct datasets CIFAR10 and CALTECH101, and three different network architectures ResNet-18, LeNet-5 and CNN-9 demonstrates the effectiveness of our proposed method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (18)
  1. (1998). Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324.
  2. Transfer learning for image classification using vgg19: Caltech-101 image data set. Journal of ambient intelligence and humanized computing, pages 1–12.
  3. Canny, J. (1986). A computational approach to edge detection. IEEE Transactions on pattern analysis and machine intelligence, (6):679–698.
  4. Few-shot learning by integrating spatial and frequency representation. In 2021 18th Conference on Robots and Vision (CRV), pages 49–56. IEEE.
  5. A review of medical image data augmentation techniques for deep learning applications. Journal of Medical Imaging and Radiation Oncology, 65(5):545–563.
  6. An adaptive gaussian filter for noise reduction and edge detection. In 1993 IEEE conference record nuclear science symposium and medical imaging conference, pages 1615–1619. IEEE.
  7. Deep residual learning for image recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 770–778.
  8. Puzzle mix: Exploiting saliency and local statistics for optimal mixup. In International Conference on Machine Learning, pages 5275–5285. PMLR.
  9. Automix: Unveiling the power of mixup for stronger classifiers. In European Conference on Computer Vision, pages 441–458. Springer.
  10. Improving robustness to out-of-distribution data by frequency-based augmentation. In 2022 IEEE International Conference on Image Processing (ICIP), pages 3116–3120. IEEE.
  11. Do cifar-10 classifiers generalize to cifar-10? arXiv preprint arXiv:1806.00451.
  12. Fads: Fourier-augmentation based data-shunting for few-shot classification. IEEE Transactions on Circuits and Systems for Video Technology.
  13. Pde based unsharp masking, crispening and high boost filtering of digital images. In Contemporary Computing: Second International Conference, IC3 2009, Noida, India, August 17-19, 2009. Proceedings 2, pages 8–13. Springer.
  14. Batchmixup: Improving training by interpolating hidden states of the entire mini-batch. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, pages 4908–4912.
  15. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF international conference on computer vision, pages 6023–6032.
  16. mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412.
  17. Overfitting and underfitting analysis for deep learning based end-to-end communication systems. In 2019 11th international conference on wireless communications and signal processing (WCSP), pages 1–6. IEEE.
  18. Mpce: a maximum probability based cross entropy loss function for neural network classification. IEEE Access, 7:146331–146341.
Citations (1)

Summary

We haven't generated a summary for this paper yet.