A.L.I.C.E (Artificial Labile Intelligence Cybernated Existence)
Is a REST API of an A.I companions.
- Nvidia RTX 3060 12 GB
- Intel i5 12400
- 24 GB DDR4 RAM
- OS: Ubuntu 24.04 LTS
This work only for Llama GPTQ
and GPTJ splitted
model for now.
I will add type GGML model soon.
- Python 3.10.9
- CUDA 12.4
- Clone this repository
git clone https://github.com/Aqirito/A.L.I.C.E.git
and go to the project folder. - Install CUDA 12
sudo apt install nvidia-cuda-toolkit
will install the latest version of CUDA. for Ubuntu 24.04 LTS it's version 12 - Inside
/configs
folder, Make a copy of all*-sample.json
file and rename by remove the-sample
. - Download your LLM models and put your local model inside
models/LLM
folder. - MoeTTS model Download your VITS models and put your local VITS TTS model inside
models/TTS/vits
folder. - Create and activate python virtual environment
python -m venv env && source env/bin/activate
- Install wheel package:
pip install wheel
- Install the requirements
pip install -r requirements.txt
- Install Exllama
python -m pip install git+https://github.com/jllllll/exllama
- Add .env follow the .env-sample and change the value
cp .env-sample .env
- Run
uvicorn fast:app
and go tolocalhost:8000/docs
to access the Swagger UI.
MODEL_NAME_OR_PATH
- You must download the model manually and save it to the
model/LLMS/
folder. Model name from hub is not working for a moment. - PLEASE don't include any slashes like
\
or/
otherwise the system will try to find and download the models from huggingface hub or maybe an error accured. - Example for local model:
Pygmalion-6b
orPygmalion-13B-SuperHOT-8K-GPTQ
- You must download the model manually and save it to the
-
TEMPLATE_TYPE
For changing the characteristic of the models.pygmalion
Roleplay, specials for pygmalion models.prompt
For all models basically just a normal AI chat bot.
-
MODEL_LOADER
A model loader for the LLM model.AutoGPTQ
AutoGPTQ model loader (GPTQ type only).HuggingFaceBig
HuggingFace Big model loader (more than one .safetensors file).ExLlama
Exllama model loader. use this for faster response. (GPTQ type only)
-
Base model config
-
Exllama model config
- See examples of Exllama
-
tts_type
Type of the TTS model used. -
language
The language of TTS (moe-tts)[EN]
English[JA]
Japanese[ZH]
Chinese[KO]
Korean
-
speed
Speed of the TTS speaker (moe-tts) -
speaker_id
Go to speakers.json to see all the speaker_id (moe-tts) -
voice
Can search speaker from edge-tts -
pitch
The pitch voice of the speaker (edge-tts) -
rate
Speed of the speaker (edge-tts) -
volume
Volume of the speaker (edge-tts)
- CjangCjengh Author of moe-tts.
- ZoltanAI Author of AI character editor.
- turboderp Author of Exllama.
- jllllll Makes Python module for Exllama.