Skip to content

jwalith/Quantized-LLM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Quantized-LLM

A mobile application that runs large language models (LLMs) directly on Android devices. Designed to showcase expertise in mobile machine learning optimization and deployment, the app loads quantized models (e.g., Qwen2.5-1.5B) and executes them entirely on-device—no internet connection required. It also features built-in performance monitoring to measure inference speed, memory usage, and response quality.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors