AAI_2025_Capstone_Chronicles_Combined

‭ResolveAI‬

‭(RAG) system was built with LangChain to feed into our LLM for the chatbot. To keep low cost‬ ‭and quick inference we used ChatGPT version 3.5 turbo for the LLM to interact with the user.‬ ‭The full flow of the chatbot included collecting the query from the user, pulling the most relevant‬ ‭examples from our database using RAG, then prompting the LLM with both the user query and‬ ‭relevant examples. Finally the response from the LLM was provided back to the user to complete‬ ‭the chatbot interface.‬

‭Results‬

‭Model Effectiveness‬

‭During multi-class training (low, medium, high priority),‬‭the model’s training accuracy‬ ‭steadily increased while training loss sharply declined, suggesting that it was successfully‬ ‭learning the nuances of the data. However, the validation accuracy plateaued and remained‬ ‭noticeably lower, and the validation loss trended upward. These signs indicate the model‬ ‭struggled to generalize beyond the training set. Initially we attempted to improve regularization‬ ‭of the model by using both aggressive dropout and reducing model size to limit variance and‬ ‭overfitting. However the model did not improve significantly with these measures. The poor‬ ‭performance was likely due to data imbalance or insufficient representation of medium or high‬ ‭tickets, prompting a shift toward binary classification to improve performance. This could also‬ ‭be due to the fact that we’re working with the upper limit of the available data and perhaps the‬ ‭“priority” classes aren’t as linearly separable in this dataset as we had hoped. In a production‬ ‭environment, there would be a more refined or expanded dataset to address the potential‬ ‭separability issue and improve overall performance.‬

‭19‬

67

Made with FlippingBook - Share PDF online