AAI_2025_Capstone_Chronicles_Combined

Cinema Analytics and Prediction System

11

To address these limitations, transformer-based models like BERT (Bidirectional Encoder

Representations from Transformers) have been developed. Introduced by Devlin et al. (2019),

BERT leverages a deep bidirectional transformer architecture to generate context-aware

embeddings. Unlike TF-IDF, BERT understands the semantics of entire sentences, allowing it to

detect deeper thematic similarities. This makes it especially powerful for tasks involving plot

understanding, genre prediction, and nuanced recommendation, and it has been widely

adopted in academic and industry NLP pipelines.

Recurrent neural networks, such as Long Short-Term Memory (LSTM) models, have also

been used in text classification and recommendation tasks. LSTMs are well-suited for handling

sequential data and have been shown to capture long-range dependencies in natural language

(Hochreiter & Schmidhuber, 1997). When combined with pre-trained GloVe word embeddings

(Pennington et al., 2014), LSTMs can provide more meaningful input representations than raw

token sequences, though they may still underperform compared to transformer-based models

when it comes to modeling context at scale.

Hybrid architectures that integrate contextual embeddings with structured data have also

emerged as an effective method for recommendation tasks. For example, Liu et al. 2024,

proposed a multimodal recommender that fuses text embeddings from language models with

numerical metadata (e.g., ratings, runtime) to improve predictive accuracy. These approaches

are particularly beneficial for domains like movies, where metadata such as popularity and

runtime often correlate with genre or audience preferences. Models that combine BERT with

normalized metadata have been shown to outperform text-only models (Yang et al., 2024), this

178

Internal

Made with FlippingBook - Share PDF online