dnhkng / GlaDOS
- пятница, 3 мая 2024 г. в 00:00:01
Personality Core
This is a project dedicated to building a real-life version of GLaDOS.
This is a hardware and software project that will create an aware, interactive, and embodied GLaDOS.
This will entail:
The initial goals are to develop a low-latency platform, where GLaDOS can respond to voice interactions within 600ms.
To do this, the system constantly records data to a circular buffer, waiting for voice to be detected. When it's determined that the voice has stopped (including detection of normal pauses), it will be transcribed quickly. This is then passed to streaming local Large Language Model, where the streamed text is broken by sentence, and passed to a text-to-speech system. This means further sentences can be generated while the current is playing, reducing latency substantially.
This will be based on servo- and stepper-motors. 3D printable STL will be provided to create GlaDOS's body, and she will be given a set of animations to express herself. The vision system will allow her to track and turn toward people and things of interest.
If you want to install the TTS Engine on your machine, please follow the steps below. This has only been tested on Linux, but I think it will work on Windows with small tweaks.
Install the espeak
synthesizer
according to the installation
instructions
for your operating system.
Install the required Python packages, e.g., by running pip install -r requirements.txt
For the LLM, install Llama.cpp, and compile it for your CPU or GPU. Edit the LLAMA_SERVER_PATH parameter in glados.py to match your installation path.
For voice recognition, install Whisper.cpp, and after compiling, run make libwhisper.so
and then move the "libwhisper.so" file to the "glados" folder or add it to your path. For Windows, check out the discussion in my whisper pull request.
Download the models:
and put them in the "models" directory.
You can test the systems by exploring the 'demo.ipynb'.