Batik Sketch Coloring Using Generative Adversarial Network Pix2pix

Main Article Content

Fanky Abdilqoyyim
Muhammad Ali Syakur
Fitri Damayanti

Abstract

Batik, an Indonesian cultural heritage recognized by UNESCO, involves a complex and time-consuming coloring process. Digitalization offers a crucial solution for the preservation and development of batik art in the modern era. This research implements a Generative Adversarial Network (GAN), specifically the Pix2Pix model, to automate the transformation of batik sketches into colored images. The primary focus is a performance comparison between the U-Net generator architecture, which excels at preserving spatial details via skip-connections, and the ResNet architecture, which is capable of learning deeper and more complex features. This study utilized 1164 paired images, divided into 931 training, 117 validation, and 116 test data points. The models were trained with consistent hyperparameters, including an Adam optimizer and a combination of L1 and binary cross-entropy loss functions, with evaluations at 50 and 100 epochs. Quantitative evaluation was performed using Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index (SSIM), and Fréchet Inception Distance (FID) metrics. The results indicate that the model with the ResNet generator trained for 100 epochs achieved the most balanced and superior performance, with a PSNR of 8.11, SSIM of 0.39, and an FID of 120.72. Overall, the ResNet generator proved more capable of producing realistic and high-quality colored batik images, offering an innovative solution to enhance the efficiency of the coloring process while supporting cultural preservation.

Article Details

Section
Articles

References

G. F. Shidik, E. Noersasongko, A. Nugraha, P. N. Andono, J. Jumanto, and E. J. Kusuma, "A systematic review of batik classification using deep learning," Journal of King Saud University - Computer and Information Sciences, vol. 34, no. 10, pp. 9725–9736, Nov. 2022, doi: 10.1016/j.jksuci.2021.11.009.

Y. A. Sari, R. K. Dewi, and C. Fatichah, "Selective Active Learning for Batik Image Classification Using Convolutional Neural Network," International Journal of Intelligent Engineering and Systems, vol. 15, no. 5, pp. 468–480, 2022, doi: 10.22266/ijies2022.1031.43.

H. Sakdiyah and L. W. Hidayah, "Blue Ocean Strategy in a Creative Industry Environment: A Madura Batik Tulis Context," in IOP Conference Series: Earth and Environmental Science, Institute of Physics Publishing, Apr. 2020. doi: 10.1088/1755-1315/469/1/012101.

Z. Li, B. Guan, Y. Wei, Y. Zhou, J. Zhang, and J. Xu, "Mapping New Realities: Ground Truth Image Creation with Pix2Pix Image-to-Image Translation," Apr. 2024, [Online]. Available: http://arxiv.org/abs/2404.19265

R. Giri, B. R. Lamichhane, and B. Pokhrel, "Sketch to Image Translation using Generative Adversarial Network," Journal homepage: Journal of Engineering and Sciences (nepjol.info) Publisher: Research Management Cell, vol. 2.

A. Tsourtis, G. Papoutsoglou, and Y. Pantazis, "GAN-Based Training of Semi-Interpretable Generators for Biological Data Interpolation and Augmentation," Applied Sciences (Switzerland), vol. 12, no. 11, Jun. 2022, doi: 10.3390/app12115434.

N. T. Tran, V. H. Tran, N. B. Nguyen, T. K. Nguyen, and N. M. Cheung, "On Data Augmentation for GAN Training," IEEE Transactions on Image Processing, vol. 30, pp. 1882–1897, 2021, doi: 10.1109/TIP.2021.3049346.

N. A. Nezla, T. P. Mithun Haridas, and M. H. Supriya, "Semantic Segmentation of Underwater Images using UNet architecture based Deep Convolutional Encoder Decoder Model," in 2021 7th International Conference on Advanced Computing and Communication Systems, ICACCS 2021, Institute of Electrical and Electronics Engineers Inc., Mar. 2021, pp. 28–33. doi: 10.1109/ICACCS51430.2021.9441804.

C. S. Wickramasinghe, D. L. Marino, and M. Manic, "ResNet Autoencoders for Unsupervised Feature Learning from High-Dimensional Data: Deep Models Resistant to Performance Degradation," IEEE Access, vol. 9, pp. 40511–40520, 2021, doi: 10.1109/ACCESS.2021.3064819.

J. Henry, T. Natalie, and D. Madsen, "Pix2Pix GAN for Image-to-Image Translation", doi: 10.13140/RG.2.2.32286.66887.

L. Jiao, C. Hu, L. Huo, and P. Tang, "Guided-Pix2Pix: End-to-End Inference and Refinement Network for Image Dehazing," IEEE J Sel Top Appl Earth Obs Remote Sens, vol. 14, pp. 3052–3069, 2021, doi: 10.1109/JSTARS.2021.3061460.

L. Kong et al., "Breaking the Dilemma of Medical Image-to-image Translation." [Online]. Available: https://github.com/Kid-Liet/Reg-GAN.

H. Abdelmotaal, A. A. Abdou, A. F. Omar, D. M. El-Sebaity, and K. Abdelazeem, "Pix2pix conditional generative adversarial networks for Scheimpflug camera colour-coded corneal tomography image generation," Transl Vis Sci Technol, vol. 10, no. 7, Jun. 2021, doi: 10.1167/tvst. 10.7.21.

M. Heusel, H. Ramsauer, T. Unterthiner, B. Nessler, and S. Hochreiter, "GANs trained by a two time-scale update rule converge to a local Nash equilibrium," Adv Neural Inf Process Syst, vol. 2017-Decem, no. Nips, pp. 6627–6638, 2017, doi: 10.18034/ajase.v8i1.9.