Attention-based U-Net for image demoiréing

Main Article Content

Tomasz Lehmann


Keywords : image demoireing, computer vision, attention U-Net, cross-sampling
Abstract

Image demoiréing is a particular example of a picture restoration problem. Moiré is an interference pattern generated by overlaying similar but slightly offset templates.
In this paper, we present a deep learning based algorithm to reduce moiré disruptions. The proposed solution contains an explanation of the cross-sampling procedure – the training dataset management method which was optimized according to limited computing resources.
Suggested neural network architecture is based on Attention U-Net structure. It is an exceptionally effective model which was not proposed before in image demoiréing systems. The greatest improvement of this model in comparison to U-Net network is the implementation of attention gates. These additional computing operations make the algorithm more focused on target structures.
We also examined three MSE and SSIM based loss functions. The SSIM index is used to predict the perceived quality of digital images and videos. A similar approach was applied in various computer vision areas.
The author’s main contributions to the image demoiréing problem contain the use of the novel architecture for this task, innovative two-part loss function, and the untypical use of the cross-sampling training procedure.

Article Details

How to Cite
Lehmann, T. (2022). Attention-based U-Net for image demoiréing. Machine Graphics and Vision, 31(1/4), 3–17. https://doi.org/10.22630/MGV.2022.31.1.1
References

I. Alhashim and P. Wonka. High quality monocular depth estimation via transfer learning. arXiv, 2018. https://doi.org/10.48550/arXiv.1812.11941.

Y. Zhang an Y. Tian, Y. Kong, B. Zhong, and Y. Fu. Residual dense network for image restoration. IEEE Transactions on Pattern Analysis and Machine Intelligence, 43(7):2480--2495, 2020. https://doi.org/10.1109/tpami.2020.2968521. (Crossref)

P. Anderson, X. He, C. Buehler, et al. Bottom-up and top-down attention for image captioning and visual question answering. In Proc. 2018 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), pages 6077-6086, Salt Lake City, Utah, 18-23 Jun 2018. IEEE. https://doi.org/10.1109/CVPR.2018.00636. (Crossref)

X. Cheng, Z. Fu, and J. Yang. Multi-scale dynamic feature encoding network for image demoiréing. In Proc. 2019 IEEE/CVF Int. Conf. Computer Vision Workshops (ICCVW), pages 3486-3493, Seoul, Korea, 27 Oct - 2 Nov 2019. IEEE. https://doi.org/10.1109/ICCVW.2019.00432. (Crossref)

J. Gurrola-Ramos, O. Dalmau, and T. E. Alarcon. A residual dense U-Net neural network for image denoising. IEEE Access, 9:31742-31754, 2021. https://doi.org/10.1109/ACCESS.2021.3061062. (Crossref)

C. Wang K. Batmanghelich D. Tao H. Fu, M. Gong. Deep ordinal regression network for monocular depth estimation. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2018. https://doi.org/10.1109/cvpr.2018.00214. (Crossref)

B. He, C. Wang, B. Shi, and L. Y. Duan. Mop moiré patterns using MopNet. In Proc. 2019 IEEE/CVF Int. Conf. Computer Vision (ICCV), pages 2424-2432, Seoul, Korea, 27 Oct - 2 Nov 2019. IEEE. https://doi.org/10.1109/ICCV.2019.00251. (Crossref)

B. He, C. Wang, B. Shi, and L.-Y. Duan. FHDe2Net: Full high definition demoireing network. In A. Vedaldi, H. Bischof, T. Brox, and J.-M. Frahm, editors, Computer Vision - Proc. ECCV 2020, volume 12367 of Lecture Notes in Computer Science, pages 713-729, Glasgow, UK, 23–28 Aug 2020. Springer International Publishing. https://doi.org/10.1007/978-3-030-58542-6_43. (Crossref)

X. Hu, M. A. Naiel, A. Wong, et al. RUNet: A robust UNet architecture for image super-resolution. In Proc. 2019 IEEE/CVF Conf. Computer Vision and Pattern Recognition Workshops (CVPRW), pages 505-507, Long Beach, California, 16-20 Jun 2019. IEEE. https://doi.org/10.1109/CVPRW.2019.00073. (Crossref)

D. P. Kingma and J. Ba. Adam: A method for stochastic optimization. In Y. Bengio and Y. LeCun, editors, Proc. 3rd Int. Conf. Learning Representations, ICLR 2015, San Diego, CA, 7-9 May 2015. Accessible in arXiv. https://doi.org/10.48550/arXiv.1412.6980.

O. Kupyn, T. Martyniuk, J. Wu, and Z. Wang. DeblurGAN-v2: Deblurring (orders-of-magnitude) faster and better. In Proc. 2019 IEEE/CVF Int. Conf. Computer Vision (ICCV), pages 8877-8886, Seoul, Korea, 27 Oct - 2 Nov 2019. IEEE. https://doi.org/10.1109/ICCV.2019.00897. (Crossref)

Z. Lu and Y. Chen. Single image super resolution based on a modified U-net with mixed gradient loss. Signal, Image and Video Processing, 16(5):1143–1151, 2022. https://doi.org/10.1007/s11760-021-02063-5. (Crossref)

S. Nah, T. H. Kim, and K. M. Lee. Deep multi-scale convolutional neural network for dynamic scene deblurring. In Proc. 2017 IEEE Conf. Computer Vision and Pattern Recognition (CVPR), pages 257-265, Honolulu, Hawaii, 21-26 Jul 2017. IEEE. https://doi.org/10.1109/CVPR.2017.35. (Crossref)

O. Oktay, J. Schlemper, L. L. Folgoc, et al. Attention U-Net: Learning where to look for the pancreas. arXiv, 2018. https://doi.org/10.48550/arXiv.1804.03999.

G. Palubinskas. Image similarity/distance measures: what is really behind MSE and SSIM? International Journal of Image and Data Fusion, 8(1):32–53, 2016. https://doi.org/10.1080/19479832.2016.1273259. (Crossref)

A. Paszke, S. Gross, F. Massa, et al. PyTorch: An imperative style, high-performance deep learning library. In Advances in Neural Information Processing Systems 32 - Proc. 33rd Conf. Neural Information Processing Systems (NeurIPS 2019), volume 11, pages 8024-8035, Vancouver, Canada, 8-14 Dec 2019. Accessible in arXiv. https://doi.org/10.48550/arXiv.1912.01703.

B. D. Ripley. Pattern Recognition and Neural Networks. Cambridge University Press, 1996. https://doi.org/10.1017/CBO9780511812651. (Crossref)

O. Ronneberger, P. Fischer, and T. Brox. U-Net: Convolutional networks for biomedical image segmentation. In N. Navab, J. Hornegger, W. M. Wells, et al., editors, Medical Image Computing and Computer-Assisted Intervention - Proc. MICCAI 2015, volume 9351 of Lecture Notes in Computer Science, pages 234-241, Munich, Germany, 5-9 Oct 2015. Springer International Publishing. https://doi.org/10.1007/978-3-319-24574-4_28. (Crossref)

De R. I. M. Setiadi. PSNR vs SSIM: imperceptibility quality assessment for image steganography. Multimedia Tools and Applications, 80(6):8423–8444, 2021. https://doi.org/10.1007/s11042-020-10035-z. (Crossref)

Y. Sun, Y. Yu, and W. Wang. Moiré photo restoration using multiresolution convolutional neural networks, 8 May 2018. https://paperswithcode.com/paper/moire-photo-restoration-using-multiresolution. [Accessed: May, 2022]. (Crossref)

F.-J Tsai, Y.-T. Peng, Y.-Y. Lin, et al. Stripformer: Strip transformer for fast image deblurring. In S. Avidan, G. Brostow, M. Cissé, et al., editors, Computer Vision - Proc. ECCV 2022, volume 13679, Part XIX of Lecture Notes in Computer Science, pages 146-162, Tel Aviv, Israel, 23-27 Oct 2022. Springer Nature Switzerland. https://doi.org/10.1007/978-3-031-19800-7_9. (Crossref)

Z. Tu, H. Talebi, H. Zhang, et al. MAXIM: Multi-Axis MLP for image processing. In Proc. 2022 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), pages 5759-5770, New Orleans, Louisiana, 18-24 Jun 2022. IEEE. https://doi.org/10.1109/CVPR52688.2022.00568. (Crossref)

X. Wang, Y. Li, H. Zhang, and Y. Shan. Towards real-world blind face restoration with generative facial prior. In Proc. 2021 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), pages 9164-9174, Virtual conference, 20-25 Jun 2021. IEEE. https://doi.org/10.1109/CVPR46437.2021.00905. (Crossref)

Z. Wang, A. C. Bovik, H. R. Sheikh, and E. P. Simoncelli. Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Processing, 13(4):600-612, Apr 2004. https://doi.org/10.1109/TIP.2003.819861. (Crossref)

Z. Wang, X. Cun, J. Bao, et al. Uformer: A general U-shaped transformer for image restoration. In Proc. 2022 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), pages 17662-17672, New Orleans, Louisiana, 18-24 Jun 2022. IEEE. https://doi.org/10.1109/CVPR52688.2022.01716. (Crossref)

W. Wang Y. Sun, Y. Yu. Moiré photo restoration using multiresolution convolutional neural networks. IEEE Trans. Image Processing, 27(8):4160–4172, 2018. https://doi.org/10.1109/tip.2018.2834737. (Crossref)

X. Yu, P. Dai, W. Li, L. Ma, et al. Towards efficient and scale-robust ultra-high-definition image demoiréing. In S. Avidan, G. Brostow, M. Cissé, et al., editors, Computer Vision - Proc. ECCV 2022, volume 13678, Part XVIII of Lecture Notes in Computer Science, pages 646-662, Tel Aviv, Israel, 23-27 Oct 2022. Springer Nature Switzerland. https://doi.org/10.1007/978-3-031-19797-0_37. (Crossref)

S. W. Zamir, A. Arora, S. Khan, et al. Restormer: Efficient transformer for high-resolution image restoration. In Proc. 2022 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), pages 5718-5729, New Orleans, Louisiana, 18-24 Jun 2022. IEEE. https://doi.org/10.1109/CVPR52688.2022.00564. (Crossref)

B. Zheng, S. Yuan, G. Slabaugh, and A. Leonardis. Image demoireing with learnable bandpass filters. In Proc. 2020 IEEE/CVF Conf. Computer Vision and Pattern Recognition (CVPR), pages 3633-3642, Virtual conference, 14-19 Jun 2020. IEEE. https://doi.org/10.1109/CVPR42600.2020.00369. (Crossref)

Statistics

Downloads

Download data is not yet available.