AAI_2025_Capstone_Chronicles_Combined
rates. The search was conducted over 25 trials using Kermany’s pneumonia dataset. The best
performing model used 64 filters in both convolutional layers, 64 neurons in the dense layer,
and a dropout rate of 0.3. L2 regularization (λ = 0.001) was applied to reduce overfitting. Adam
was chosen as the optimizer, with a learning rate of 0.0001 yielding the most stable and
accurate results when minimizing binary cross-entropy loss.
Training followed a structured methodology to ensure robustness. Image preprocessing
was performed using Keras’s ImageDataGenerator, applying real-time data augmentations such
as normalization, shear transformations, horizontal flipping, and brightness adjustments. All
images were resized to 128×128 pixels and converted to grayscale to standardize inputs and
minimize variations due to lighting or color differences. Data augmentation techniques helped
prevent overfitting and encouraged the model to learn actual features rather than memorize
feature locations.
The model was trained three times, corresponding to each data source: Kaggle training
data, synthetics generated via custom ACGAN, and synthetics produced through StyleGAN. For
each synthetic dataset, 1,000 images representing pneumonia and normal states were
aggregated, then divided into training and validation sets with an 80/20 split. Actual Kaggle test
images served as the test dataset for all models.
Leveraging Pre-Trained VGG16 Model
To improve validation performance metrics, transfer learning was tested using the
VGG-16 model due to its versatility and proven success in similar applications. The model
106
Made with FlippingBook - Share PDF online