Hybrid Approach for Denoising and Segmentation: N2S with Swin Transformerenhanced U-Net

Ashwini G.*, Ramashri T.**
*_** Department of Electronics and Communication Engineering, Sri Venkateswara University College of Engineering, Tirupati, Andhra Pradesh, India.
Periodicity:January - March'2025
DOI : https://doi.org/10.26634/jip.12.1.21658

Abstract

Accurate segmentation in medical imaging, particularly for modalities such as Chest X-rays, CT scans, and microscopic images, is critical for diagnosis and treatment. However, noisy and low-quality data can significantly affect performance. This paper presents a novel framework that integrates Noise2Split denoising with a Hybrid Swin Transformer U-Net to enhance segmentation accuracy in these challenging medical imaging tasks. By combining Noise2Split's effective noise reduction with the Swin Transformer's advanced feature extraction and U-Net's robust segmentation architecture, the model efficiently addresses both noise and segmentation challenges. The Swin Transformer effectively captures both local and global context, while the skip connections in U-Net contribute to recovering detailed high- resolution features. Extensive experiments on Chest X-rays, CT scans, and microscopic images demonstrate that this integrated model performs better than traditional methods in terms of segmentation accuracy, making it a valuable tool for clinical applications where imaging quality is compromised.

Keywords

Image Segmentation, N2S, De-Noising, Swin Transformer, U-Net, Deep Learning.

How to Cite this Article?

Ashwini, G., and Ramashri, T. (2025). Hybrid Approach for Denoising and Segmentation: N2S with Swin Transformerenhanced U-Net. i-manager’s Journal on Image Processing, 12(1), 50-62. https://doi.org/10.26634/jip.12.1.21658

References

[3]. Bibaut, A., Kallus, N., Dimakopoulou, M., Chambaz, A., & van Der Laan, M. (2021). Risk minimization from adaptively collected data: Guarantees for supervised and policy learning. Advances in Neural Information Processing Systems, 34, 19261-19273.
[7]. Chen, L. C., Zhu, Y., Papandreou, G., Schroff, F., & Adam, H. (2018). Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV) (pp. 801-818).
[10]. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (pp. 770-778).
[14]. Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z & Guo, B. (2021). Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (pp. 10012-10022).
[17]. Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., & Jégou, H. (2021). Training data- efficient image transformers & distillation through attention. In International Conference on Machine Learning (pp. 10347-10357). PMLR.
If you have access to this article please login to view the article or kindly login to purchase the article

Purchase Instant Access

Single Article

North Americas,UK,
Middle East,Europe
India Rest of world
USD EUR INR USD-ROW
Pdf 35 35 200 20
Online 15 15 200 15
Pdf & Online 35 35 400 25

Options for accessing this content:
  • If you would like institutional access to this content, please recommend the title to your librarian.
    Library Recommendation Form
  • If you already have i-manager's user account: Login above and proceed to purchase the article.
  • New Users: Please register, then proceed to purchase the article.