| Hi everyone, I’m the maintainer of Box — a fork of Google’s AI Edge Gallery that I’ve been extending into a fully offline AI assistant for Android. Full disclosure: I built this project. It runs entirely on-device (no cloud, no accounts, no external inference), and combines multiple local inference backends in a single app. What I’ve been experimenting withThe goal was to see how far a fully offline mobile AI stack could be pushed using:
All running on Android with hardware acceleration where available (GPU / NPU / TPU). Current capabilities
Architecture focusWhat I’ve found interesting while building this:
Repo (for reference)Why I’m posting this hereI’m mainly sharing this for feedback from people also working on local inference systems, especially around:
Not trying to push adoption — more interested in technical critique than anything else. Happy to answer questions or go deeper into any part of the stack if useful. [link] [comments] |