tokyo // nyc — reustle.org
64 Followers
I built a prototype showcasing how I think AI assistants will interact with existing software during an OpenAI Realtime Hackathon this weekend. Won 2nd place and $15k in credits. (skip to 1:00) https://hirebellhop.com
Important clarification: everything is happening on the client side except for the the audio being proxied to OpenAI, then receiving tool calls back.
Thanks! 3 in parallel, 4o realtime audio model (basic listening) + whisper model (transcripts) + o3-mini (double-checking regularly)
There are a few routes to go down w/ that. What I've been brainstorming for a few weeks, heavily related to MCP, is how to essentially let these agents go hunt down the API docs for the service & your API key when logged in, then it can do most (?) things w/out fumbling w/ the browser.