Releases: mudler/LocalAI
v2.6.1
This is a patch release containing bug-fixes around parallel request support with llama.cpp models.
What's Changed
Bug fixes 🐛
- fix(llama.cpp): Enable parallel requests by @tauven in #1616
- fix(llama.cpp): enable cont batching when parallel is set by @mudler in #1622
Exciting New Features 🎉
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1623
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1619
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1620
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1626
New Contributors
Full Changelog: v2.6.0...v2.6.1
v2.6.0
What's Changed
Bug fixes 🐛
- move BUILD_GRPC_FOR_BACKEND_LLAMA logic to makefile: errors in this section now immediately fail the build by @dionysius in #1576
- prepend built binaries in PATH for BUILD_GRPC_FOR_BACKEND_LLAMA by @dionysius in #1593
Exciting New Features 🎉
- minor: replace shell pwd in Makefile with CURDIR for better windows compatibility by @dionysius in #1571
- Makefile: allow to build without GRPC_BACKENDS by @mudler in #1607
- feat: 🐍 add mamba support by @mudler in #1589
- feat(extra-backends): Improvements, adding mamba example by @mudler in #1618
👒 Dependencies
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1567
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1568
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1573
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1578
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1583
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1587
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1590
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1594
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1599
Other Changes
- Moving the how tos to self hosted by @lunamidori5 in #1574
- docs: missing golang requirement for local build for debian by @dionysius in #1596
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1597
- docs/examples: enhancements by @mudler in #1572
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1604
- Update README.md by @lunamidori5 in #1601
- docs: re-use original permalinks by @mudler in #1610
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1612
- Expanded and interlinked Docker documentation by @jamesbraza in #1614
- Modernized LlamaIndex integration by @jamesbraza in #1613
New Contributors
- @dionysius made their first contribution in #1571
Full Changelog: v2.5.1...v2.6.0
v2.5.1
Patch release to create /build/models
in the container images.
What's Changed
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1562
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1563
Full Changelog: v2.5.0...v2.5.1
v2.5.0
What's Changed
This release adds more embedded models, and shrink image sizes.
You can run now phi-2
( see here for the full list ) locally by starting localai with:
docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core phi-2
LocalAI accepts now as argument a list of short-hands models and/or URLs pointing to valid yaml file. A popular way to host those files are Github gists.
For instance, you can run llava
, by starting local-ai
with:
docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/llava.yaml
Exciting New Features 🎉
👒 Dependencies
- deps(conda): use transformers-env with vllm,exllama(2) by @mudler in #1554
- deps(conda): use transformers environment with autogptq by @mudler in #1555
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1558
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1557
Full Changelog: v2.4.1...v2.5.0
v2.4.1
What's Changed
Exciting New Features 🎉
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1546
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1547
- docs: improve getting started by @mudler in #1553
Full Changelog: v2.4.0...v2.4.1
v2.4.0
What's Changed
Bug fixes 🐛
- fix(entrypoint): cd to backend dir before start by @mudler in #1530
- fix(diffusers): add omegaconf dependency by @mudler in #1540
Exciting New Features 🎉
- feat(preload): prepare models in galleries by @mudler in #1515
- feat: share models by url by @mudler in #1522
- feat(prepare): allow to specify additional files to download by @mudler in #1526
- feat: add trimsuffix by @mudler in #1528
👒 Dependencies
- deps(llama.cpp): update and sync grpc server by @mudler in #1527
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1531
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1517
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1518
- docs(phi-2): add example by @mudler in #1525
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1533
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1544
- ci(dockerhub): push images also to dockerhub by @mudler in #1542
Full Changelog: v2.3.1...v2.4.0
v2.3.1
What's Changed
Bug fixes 🐛
- Fix: Set proper Homebrew install location for x86 Macs by @cnatale in #1510
- fix(download): correctly check for not found error by @mudler in #1514
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1509
Other Changes
- Update version.json (2.3.0) by @lunamidori5 in #1511
New Contributors
Full Changelog: v2.3.0...v2.3.1
v2.3.0
What's Changed
Bug fixes 🐛
- fix: exllama2 backend by @sozercan in #1484
- fix: guidance_scale not work in sd by @hibobmaster in #1488
- fix(image-gen): remove default to stablediffusion by @JZacharie in #1500
Exciting New Features 🎉
- env(conda): use transformers for vall-e-x by @mudler in #1481
- feat: add 🐸 coqui by @mudler in #1489
- feat: partial download by @muka in #1486
- feat: add tiny dream stable diffusion support by @M0Rf30 in #1283
👒 Dependencies
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1479
- ⬆️ Update donomii/go-rwkv.cpp by @localai-bot in #1478
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1469
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1482
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1483
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1492
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1495
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1496
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1501
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1503
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1477
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1480
- docs: add langchain4j integration by @mudler in #1476
- ci(apple): speedups by @mudler in #1471
New Contributors
- @hibobmaster made their first contribution in #1488
- @muka made their first contribution in #1486
- @JZacharie made their first contribution in #1500
Full Changelog: v2.2.0...v2.3.0
v2.2.0
What's Changed
This release brings updates to the backends and includes a fix for recompilation of LocalAI with go-rwkv (#1473). To note, it also tries to reduce the image size by allowing some backends (transformers-based) to share the same environment.
With this release inline templates and models as URLs are supported, for example:
name: mixtral
parameters:
model: https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GGUF/resolve/main/mixtral-8x7b-v0.1.Q2_K.gguf
# or huggingface://TheBloke/Mixtral-8x7B-v0.1-GGUF/mixtral-8x7b-v0.1.Q2_K.gguf@main
template:
completion: |
Complete the following: {{.Input}}
Exciting New Features 🎉
- docs: automatically track latest versions by @mudler in #1451
- feat: inline templates and accept URLs in models by @mudler in #1452
- feat(conda): share envs with transformer-based backends by @mudler in #1465
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1450
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1455
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1457
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1462
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1461
- ⬆️ Update donomii/go-rwkv.cpp by @localai-bot in #1470
Other Changes
- How To (Updates and Fixes) by @lunamidori5 in #1456
- Revert ":arrow_up: Update donomii/go-rwkv.cpp" by @mudler in #1474
Full Changelog: v2.1.0...v2.2.0
v2.1.0
What's Changed
Breaking Changes 🛠
Bug fixes 🐛
- fix(piper): pin petals, phonemize and espeak by @mudler in #1393
- update(llama.cpp): update server, correctly propagate LLAMA_VERSION by @mudler in #1440
Exciting New Features 🎉
- Added Check API KEYs file to API.go by @lunamidori5 in #1381
- exllama(v2): fix exllamav1, add exllamav2 by @mudler in #1384
- Fix: API Key / JSON Fast Follow #1 by @dave-gray101 in #1388
- feat: add transformers-musicgen backend by @dave-gray101 in #1387
- feat(diffusers): update, add autopipeline, controlnet by @mudler in #1432
- feat(img2vid,txt2vid): Initial support for img2vid,txt2vid by @mudler in #1442
👒 Dependencies
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1378
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1379
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1430
- ⬆️ Update mudler/go-piper by @localai-bot in #1441
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1434
Other Changes
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1385
- docs: site update fixing old image text / How To update updating GPU and CPU docker pages by @lunamidori5 in #1399
- feat: cuda transformers by @mudler in #1401
- feat(entrypoint): optionally prepare extra endpoints by @mudler in #1405
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1390
- ⬆️ Update mudler/go-piper by @localai-bot in #1400
- tests: add diffusers tests by @mudler in #1419
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1418
- How To Updates / Model Used Switched / Removed "docker-compose" (RIP) by @lunamidori5 in #1417
- fix(transformers*): add sentence-transformers and transformers-musicgen tests, fix musicgen wrapper by @mudler in #1420
- extras: add vllm,bark,vall-e-x tests, bump diffusers by @mudler in #1422
- Documentation for Hipblas by @sfxworks in #1425
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1391
- docs: add aikit to integrations by @sozercan in #1412
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1429
- docs(mixtral): add mixtral example by @mudler in #1449
New Contributors
Full Changelog: v2.0.0...v2.1.0