Quantization in Deep Learning

Research Question:

How can Retrieval-Augmented Generation (RAG) technology be leveraged to enhance developers' understanding of code repositories as a developer tool?

Project Summary:

I'm exploring model quantization techniques for deep neural networks, focusing on how they impact performance and efficiency in convolutional neural networks and large language models. I'll present a literature review of multiple SOTA methods like AWQ (based on past research papers), followed by a practical demonstration comparing the same model quantized and non-quantized.

Student: Beau Hayes-Pollard (Computer Science, '26)

Collaborators: Austin Bunn, Yijia Dai (Graduate student mentor) 

Top