🐳 With Docker
Libre Chat is available as a docker image, it is recommended to use docker for deploying in production as it uses gunicorn to run multiple workers.
Shared memory for multiple users
Memory of the chatbot is shared betweem the users that are on the same worker.
⚡ Quickstart
If you just want deploy it using the pre-trained Mixtral model, you can use docker:
⚙️ Configure with docker compose
- Create a
chat.yml
file with your chat web service configuration. -
Create the
docker-compose.yml
in the same folder:docker-compose.ymlversion: "3" services: libre-chat: image: ghcr.io/vemonet/libre-chat:main volumes: # ⚠️ Share files from the current directory to the /data dir in the container - ./chat.yml:/data/chat.yml - ./models:/data/models - ./documents:/data/documents - ./embeddings:/data/embeddings - ./vectorstore:/data/vectorstore ports: - 8000:8000 environment: - LIBRECHAT_WORKERS=1
-
Start your chat web service with:
Using multiple workers
Using multiple worker is still experimental. When using a documents-based QA chatbot you will need to restart the API after adding new documents to make sure all workers reload the newly built vectorstore.