Wang, T., & Lin, Y. (2024). CycleGAN with Better Cycles. Technical Report. arXiv:2408.15374v2 [cs.CV]
This research paper aims to improve the quality and realism of images generated by CycleGAN, a deep learning model used for unpaired image-to-image translation. The authors identify limitations with the model's cycle consistency loss, which can lead to unrealistic artifacts in the generated images.
The authors propose three modifications to the cycle consistency loss in CycleGAN:
The authors evaluate their proposed modifications on the horse2zebra dataset and compare their results to the original CycleGAN model.
The authors demonstrate that their proposed modifications lead to improved image quality and realism compared to the original CycleGAN model. The generated images exhibit fewer artifacts and more closely resemble real images from the target domain.
The authors conclude that their proposed modifications to the cycle consistency loss in CycleGAN effectively address limitations in the original model and result in more realistic image-to-image translation.
This research contributes to the field of image-to-image translation by improving the quality and realism of generated images. The proposed modifications to CycleGAN have the potential to enhance various applications, including domain adaptation, image editing, and data augmentation.
The authors acknowledge the need for further parameter tuning to optimize the performance of their proposed modifications. They also suggest exploring the use of pretrained discriminators and incorporating stochastic input into the generator network for improved diversity in generated images. Additionally, investigating alternative consistency constraints and exploring the latent space representation in CycleGAN are promising avenues for future research.
In un'altra lingua
dal contenuto originale
arxiv.org
Approfondimenti chiave tratti da
by Tongzhou Wan... alle arxiv.org 11-25-2024
https://arxiv.org/pdf/2408.15374.pdfDomande più approfondite