The idea is I’d go on stage singing and playing guitar with a looper and some samples, then bring a robot toy and introduce the robot “controlling” the looping and sampling as the bandmate.
It’s a gimmick that’s been done before, but with LLMs driving verbal interaction and now I could use this to animate a robot…it becomes pretty compelling. I’d plug the LLMs into the audio feed so I could banter with it and get responses then have the robot avatar animate accordingly.
If only my full time job saw value in this project.
I’d prefer much less wealth inequality, but it’s not like the only benefit of automation is profit.
Also for the author in case they read this, the central/decentral hub architecture writing in privacy section is overly complex. Simplify it as "locally hosted, not cloud connected"
One example on unmanned boats: a human could radio to the boat over VHF and say “move 100 meters south”… that speech-to-text would feed to an LLM which extracts the meaning and calls the MCP.
I’ll have to install this and play around.
I tried the MCP server with the demo (https://merliot.io/demo) using Cursor and asked:
What is the location of the "GPS nano" device?
The location of the "GPS nano" device is: Latitude: 30.448336 Longitude: -91.12896
also great - improving the ide interface: the last time i worked with esp32's using (micro) python the best solution i found was the Thonny ide - not the most productive setup.
Is there a magic button to shut down the AI and go to 100% manual mode?