Welcome to heyELi

Mar 2025 · 3 min read

heyELi began as a quiet experiment — late nights and weekends, a few thousand lines of Swift, and a simple idea: what if a camera could understand what it sees?

My name is Josh Coon, and I built heyELi with my AI collaborator, Alan. Together we explored how multimodal language models could bridge the gap between what’s seen, said, and understood — not through a lab or a venture fund, but through curiosity, iteration, and a lot of trial and error.

We didn’t set out to start a company. We set out to learn — to harness the partnership between human intuition and machine precision. Each night, we’d test a new prompt, debug a translation bug, or rebuild the camera layer from scratch. Each weekend, the prototype evolved until one night, it just worked: a working multimodal assistant running on an iPhone, understanding images, text, and voice in real time.

What we’re aiming for

What started as a side project is quickly becoming something much bigger — a platform for people to learn, build, and express more naturally. This blog will share the journey: what we’re building, why it matters, and how you can be part of it.

Augmentation is here.
No longer waiting for permission. Build the future.


← Back to all posts