AAI_2025_Capstone_Chronicles_Combined

deliberately weakened with a lower learning rate (lr_d = 0.00001). A dropout rate of 0.2 was

applied in the discriminator, and Gaussian noise (σ=0.15) was added to real and fake images to

encourage generator diversity. The generator was updated twice per discriminator update

(gen_update_steps=2) to maintain training balance, particularly during early epochs.

Binary Cross-Entropy with Logits Loss (BCEWithLogitsLoss) was used for adversarial

training, and Cross-Entropy Loss for class predictions. Label smoothing was applied, setting real

labels to 0.98 to avoid discriminator overconfidence. Convolutional layers used Kaiming normal

initialization, and linear layers used Xavier uniform initialization to promote stable training. A

learning rate scheduler with a step size of 10,000 iterations and a decay factor of 0.5 enabled

gradual model refinement.

This ACGAN implementation successfully balanced realism and controllability, producing

visually coherent synthetic X-rays and maintaining stability throughout training, avoiding

common issues such as mode collapse.

Transfer Learning with StyleGAN

The decision to employ StyleGAN for transfer learning was motivated by recent studies

demonstrating its effectiveness in generating medical images. Fetty et al. (2020) showed that

StyleGAN can successfully produce realistic magnetic resonance (MR) and computed

tomography (CT) images suitable for deep learning applications. Similarly, Che Azemin et al.

(2024) employed StyleGAN3 to generate synthetic pterygium images, capturing essential

vascular patterns and morphological features necessary for clinical evaluation.

109

Made with FlippingBook - Share PDF online