[ICLR-2025-SLLM Spotlight 🔥]MobiLlama : Small Language Model tailored for edge devices
-
Updated
May 10, 2025 - Python
[ICLR-2025-SLLM Spotlight 🔥]MobiLlama : Small Language Model tailored for edge devices
on-mobile-llm is a research-style project that evaluates how well small language models (0.5B–2B parameters) run fully offline on an Android smartphone using GGUF + llama.cpp + Termux. The goal is to measure speed, memory usage, thermals, stability, and output quality across a variety of SLM architectures.
Add a description, image, and links to the mobile-llm topic page so that developers can more easily learn about it.
To associate your repository with the mobile-llm topic, visit your repo's landing page and select "manage topics."