llamafu pushes the boundaries of on-device LLM inference. Built with Flutter and llama.cpp, it runs full language models on mobile hardware with zero cloud dependency — studying the practical limits of memory, latency, and model quality on consumer devices.