AAI_2025_Capstone_Chronicles_Combined

2

Draw, Detect, Navigate ​

Abstract

This capstone project extends traditional sketch classification by incorporating real-time object

detection and augmented reality (AR) integration. While convolutional neural networks (CNNs)

have demonstrated high accuracy in identifying hand-drawn pictograms, most approaches omit

spatial localization and real-time interaction. Our system addresses these gaps by using

bounding box predictions and edge-optimized models capable of identifying multiple doodles

in real time within an AR environment. To overcome limitations in existing datasets, we

developed a synthetic data generation pipeline using Unity and Python, producing randomized,

annotated images that mirror real-world drawing variability. We trained and evaluated both

Faster R-CNN and YOLOv8 variants, ultimately selecting the YOLOv8 nano model for deployment

due to its speed, size, and high accuracy. The model achieved an F1 score of 0.96 and

processed images at 28 frames per second (FPS), enabling seamless AR integration. The final

application uses a webcam feed and ArUco marker tracking to detect hand-drawn symbols,

anchor 3D models to the drawings, and compute navigation routes using the A* algorithm. The

system combines symbolic vision with spatial reasoning to support interactive use cases such

as route planning, adversarial simulations, and strategic modeling. The model’s real-time

performance on new inputs, along with its successful deployment through Unity, supports its

use in live scenarios. These results show that sketch-based simulations can support quick

decision making in settings where fast, visual input is needed.

29

Made with FlippingBook - Share PDF online