AAI_2025_Capstone_Chronicles_Combined

Cinema Analytics and Prediction System

27

An LSTM model with pre-trained GloVe embeddings captured sequential patterns in

descriptions, performing well on dominant genres but showing low precision and recall for

minority classes. Training curves indicated no overfitting,

but performance plateaued due to class imbalance and

limited data for rare genres.From the best training results,

we can see (Figure 26) how the validation loss curves

demonstrate the success of the final LSTM model

architecture, which combined GloVe embeddings,

Figure 26: LSTM Training vs Validation Loss

metadata features, and oversampling to address class

imbalance. Both losses decline steadily in parallel,

converging to approximately 0.40 (training) and 0.45 (validation), indicating effective

optimization with minimal overfitting. This smooth convergence validates the model’s design

choices: the frozen-then-fine-tuned embedding strategy maintained stable feature extraction

during early training, while dropout regularization and carefully tuned LSTM units (64/128)

prevented overfitting despite the model’s complexity.

From the following prediction table (Figure 27), it is clearly shown high accuracy for

animation/family films with multiple genres. On the other hand, systematic underprediction of

horror/thriller genres and over-simplification of crime-drama hybrids. The model also has

occasional severe misclassifications (e.g., comedy -> animation). Predictions are most reliable

for genres with strong visual/thematic stereotypes and least reliable for mood-dependent or

adult-oriented genres.

194

Internal

Made with FlippingBook - Share PDF online