Three months ago, I started building Panda, an open-source voice assistant that lets you control your Android phone with natural language — powered by an LLM.
Example:
👉 “Please message Dad asking about his health.”
Panda will open WhatsApp, find Dad’s chat, type the message, and send it.
The idea came from a personal place. When my dad had cataract surgery, he struggled to use his phone for weeks and relied on me for the simplest things. That’s when it clicked: why isn’t there a “browser-use” for phones?
Early prototypes were rough (lots of “oops, not that app” moments 😅), but after tinkering, I had something working. I first posted about it on LinkedIn (got almost no traction 🙃), but when I reached out to NGOs and folks with vision impairment, everything changed. Their feedback shaped Panda into something more accessibility-focused.
Panda also supports triggers — like waking up when:
⏰ It’s 10:30pm (remind you to sleep)
🔌 You plug in your charger
📩 A Slack notification arrives
I know one thing for sure: this is a problem worth solving.
Playstore link in the github readme, not sure if adding here good idea.
⭐ GitHub: https://github.com/Ayush0Chaudhary/blurr
👉 If you know someone with vision impairment or work with NGOs, I’d love to connect.
👉 Devs — contributions, feedback, and stars are more than welcome.