Well first of all, welcome! this is the first devlog I am going to write completely myself without the use of any AI, so please excuse any grammatical or spelling mistakes as I am going to make quite a few of them. anyway that’s besides the point, the thing is I have been working on a new project for quite some time [as you can see, around 12 hours logged]. Now, why were there no devlogs in that time? The thing was that 2 of my friends had committed to building this project together and I was waiting for them to join, and as it was not my own project I decided against putting it on flavortown, but guess what… they both bailed on me (hence the project then being posted here, rest assured all the commits were by me)!Â
Anyway, eidolon is ment to function as a bridge to allow you deploy a live AI model [full duplex to be exact] as your personal assistant which can send and receive phone calls on your mobile device on your behalf. It functions by using the host to set up a Bluetooth HFP that a mobile considers as a headset and sends audio to, as it can send and receive audio, it can interact natively with the call.
Anyway curuntly its in very early PoC stage where I have just made the linux version working correctly [macOS version is also in its final stage of the PoC but will not work as intended because of kernel level constraints]. I am curruntly working on the esp32 version [ohh did I mention I will use the esp32 I bought from the shop to expand it to work without any desktop devices, its going to be sooo flipping cool!!]
also also, curuntly the linux and any other working version just have a basic TTS and STT implementation ment to function as a PoC, in the next (first ship) ill push all the working versions of it [esp32, linux, windows n mac] with the PoC TTS and STT. after it is done I’ll begin to work on actually implementing a live AI API [prolly openAI as I can get their credits from the shop and they are the only ones which provide a live API]
Also did I mention Eidolon will also have a mobile app operating over wifi to share live data about calls like transcription, what the agent is saying and providing the ability to guide it? again, it’ll be sooo fun, can’t wait!! [the attached screenshot is from the CursesTUI which I am using in the PoC working on linux (and me connecting to my pixel 9 and testing it with a call, also debug flag was set to true hence so many logs)!]