项目地址:llama-cpp-python
镜像地址:hub-mirror
模型地址:ggml-vic13b-q5_1.bin
前端UI地址(目前仍不兼容):BetterChatGPT
部署Docker镜像
version: '3.3'
services:
llama:
ports:
- '1234:8000'
restart: always
environment:
MODEL: /models/ggml-vic13b-q5_1.bin
volumes:
- './models:/models'
image: togettoyou/ghcr.io.abetlen.llama-cpp-python:latest
command: ["python3", "-m", "llama_cpp.server", "--model", "/models/ggml-vic13b-q5_1.bin"]
- mkdir -p ~/app/llama && cd ~/app/llama && nano docker-compose.yml
- mkdir models && cd models
- wget https://huggingface.co/eachadea/ggml-vicuna-13b-1.1/resolve/main/ggml-vic13b-q5_1.bin
- cd ~/app/llama
- sudo docker-compose up -d
- sudo docker-compose logs
测试是否成功
- 查看文档:http://localhost:1234/docs
curl http://localhost:1234/v1/chat/completions \
-H 'Content-Type: application/json' \
-d '{
"model": "gpt-3.5-turbo",
"messages": [{"role": "user", "content": "Hello!"}]
}'
Comments NOTHING