v1.7.1 - 2024-06-13
- Copy over new binary names
v1.7.0 - 2024-06-12
- Update nvidia/cuda to 12.5.0
- Add libomp to clang build
- Add libgomp to cuda build
- LLAMA_CUBLAS is deprecated
v1.6.1 - 2024-05-21
- Include new models as makefile targets (again)
v1.6.0 - 2024-05-20
- Add llama 3 8B and phi 3 mini
v1.5.1 - 2024-04-15
- Include new models as makefile targets
v1.5.0 - 2024-04-15
- Add command-r 35b model
- Add starling 7b beta model
- Sort by lmsys leaderboard elo score
- Update nvidia/cuda to 12.4.1
v1.4.0 - 2024-03-06
- Use clang 16 instead of gcc in cpu version
v1.3.1 - 2024-01-06
- Move entrypoint to bottom of Dockerfile
- Suppress missing nvidia-smi command output
v1.3.0 - 2024-01-03
- Improve docker sudo detection in Makefile
- Move download to docker-entrypoint.sh
- Rename target to llama 2 to match download
v1.2.1 - 2023-12-27
- Ensure build stages are named
v1.2.0 - 2023-12-25
- Automatically build and run gpu or cpu version
v1.1.0 - 2023-12-22
- Convert env vars to command line args
v1.0.0 - 2023-12-20
- Do not add downloaded models to git
- Run llama.cpp with GPU enabled Docker Compose