AAI_2025_Capstone_Chronicles_Combined
ResolveAI
Model Optimization
Optimization of the model is achieved through an iterative process that involves several strategies. Initially, the architecture includes bidirectional LSTM layers to capture contextual information from both directions in the text, which is crucial for understanding the nuances in ticket descriptions. However, early experiments revealed challenges such as overfitting and high validation loss. To address these issues, the training procedure was refined by increasing dropout rates and incorporating L2 regularization to impose weight decay on the network layers. Additionally, the model complexity was reduced by decreasing the number of LSTM units and layers. Learning rate adjustments were also explored using scheduling strategies that reduce the learning rate when the validation loss stagnates, allowing the model to converge more gradually. Despite these efforts, the performance of the multi-class model, tasked with distinguishing among low, medium, and high priorities remained suboptimal. Consequently, the problem was re-framed as a binary classification task by grouping medium and high priority tickets together. This simplification helped address issues of data imbalance and overfitting, resulting in a more robust and generalizable model.
LLM with Retrieval Augmented Generation
To implement the chatbot portion of our support agent we utilized a combination of Langchain, ChromaDB and OpenAI’s ChatGPT 3.5 Turbo model. We used both Langchain and ChromaDB to process the question/answer pairs in the training dataset into a vector database. To serve the most relevant examples based on a user query a Retrieval Augmented Generation
18
66
Made with FlippingBook - Share PDF online