You can now control a computer with just your voice. Here’s how we did it. We built on Replit’s template to combine Hume’s empathic voice interface with Anthropic’s computer use API. EVI processes speech in real time, sends instructions to the agentic computer control loop, explains its actions with voice, and can even be interrupted to change course. Try it out with our open-source demo: https://lnkd.in/e_weNThz This works because Hume’s frontier speech-LLM, EVI 2, can generate its own language, but can also read out lines from other language models like an actor reading a script. EVI is the only voice-to-voice model that’s both interoperable with any LLM and available today as an API. Learn more here: https://lnkd.in/e9Kqu-YM
Excellent flow, congratulations.
Will Jackson check this out.
Jake Marrello Supreet Deshpande
Wow 🤯
All about Intelligence - Natural & Artificial Math, Brains & ML |Prev Data Science @DELL | Generative Modeling Co-Founder @CADSA | Google CSRMP Fellow 🏆 TreeHacks'22 | HackMIT'22 🏆| CalHacks '24 🥇| HackUMass ‘21 🏆
1moThat’s the way it’s meant to be🔥