Skip to content

Releases: mudler/LocalAI

v2.6.1

23 Jan 17:22
d5d82ba
Compare
Choose a tag to compare

This is a patch release containing bug-fixes around parallel request support with llama.cpp models.

What's Changed

Bug fixes 🐛

  • fix(llama.cpp): Enable parallel requests by @tauven in #1616
  • fix(llama.cpp): enable cont batching when parallel is set by @mudler in #1622

Exciting New Features 🎉

  • feat(grpc): backend SPI pluggable in embedding mode by @coyzeng in #1621

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.6.0...v2.6.1

v2.6.0

20 Jan 17:34
06cd9ef
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • move BUILD_GRPC_FOR_BACKEND_LLAMA logic to makefile: errors in this section now immediately fail the build by @dionysius in #1576
  • prepend built binaries in PATH for BUILD_GRPC_FOR_BACKEND_LLAMA by @dionysius in #1593

Exciting New Features 🎉

  • minor: replace shell pwd in Makefile with CURDIR for better windows compatibility by @dionysius in #1571
  • Makefile: allow to build without GRPC_BACKENDS by @mudler in #1607
  • feat: 🐍 add mamba support by @mudler in #1589
  • feat(extra-backends): Improvements, adding mamba example by @mudler in #1618

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.5.1...v2.6.0

v2.5.1

09 Jan 08:00
5309da4
Compare
Choose a tag to compare

Patch release to create /build/models in the container images.

What's Changed

Other Changes

Full Changelog: v2.5.0...v2.5.1

v2.5.0

08 Jan 13:55
574fa67
Compare
Choose a tag to compare

What's Changed

This release adds more embedded models, and shrink image sizes.

You can run now phi-2 ( see here for the full list ) locally by starting localai with:

docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core phi-2

LocalAI accepts now as argument a list of short-hands models and/or URLs pointing to valid yaml file. A popular way to host those files are Github gists.

For instance, you can run llava, by starting local-ai with:

docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/llava.yaml

Exciting New Features 🎉

  • feat: more embedded models, coqui fixes, add model usage and description by @mudler in #1556

👒 Dependencies

  • deps(conda): use transformers-env with vllm,exllama(2) by @mudler in #1554
  • deps(conda): use transformers environment with autogptq by @mudler in #1555
  • ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1558

Other Changes

Full Changelog: v2.4.1...v2.5.0

v2.4.1

06 Jan 00:05
ce724a7
Compare
Choose a tag to compare

What's Changed

Exciting New Features 🎉

  • feat: embedded model configurations, add popular model examples, refactoring by @mudler in #1532

Other Changes

Full Changelog: v2.4.0...v2.4.1

v2.4.0

04 Jan 16:30
bcf0244
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • fix(entrypoint): cd to backend dir before start by @mudler in #1530
  • fix(diffusers): add omegaconf dependency by @mudler in #1540

Exciting New Features 🎉

👒 Dependencies

Other Changes

Full Changelog: v2.3.1...v2.4.0

v2.3.1

30 Dec 14:38
a95bb05
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

  • Fix: Set proper Homebrew install location for x86 Macs by @cnatale in #1510
  • fix(download): correctly check for not found error by @mudler in #1514

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.3.0...v2.3.1

v2.3.0

29 Dec 19:01
9723c3c
Compare
Choose a tag to compare

What's Changed

Bug fixes 🐛

Exciting New Features 🎉

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.2.0...v2.3.0

v2.2.0

21 Dec 16:48
9ae47d3
Compare
Choose a tag to compare

What's Changed

This release brings updates to the backends and includes a fix for recompilation of LocalAI with go-rwkv (#1473). To note, it also tries to reduce the image size by allowing some backends (transformers-based) to share the same environment.

With this release inline templates and models as URLs are supported, for example:

name: mixtral

parameters:
  model: https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GGUF/resolve/main/mixtral-8x7b-v0.1.Q2_K.gguf
# or huggingface://TheBloke/Mixtral-8x7B-v0.1-GGUF/mixtral-8x7b-v0.1.Q2_K.gguf@main
template:
  completion: |
            Complete the following: {{.Input}}

Exciting New Features 🎉

  • docs: automatically track latest versions by @mudler in #1451
  • feat: inline templates and accept URLs in models by @mudler in #1452
  • feat(conda): share envs with transformer-based backends by @mudler in #1465

👒 Dependencies

Other Changes

Full Changelog: v2.1.0...v2.2.0

v2.1.0

16 Dec 21:04
3d83128
Compare
Choose a tag to compare

What's Changed

Breaking Changes 🛠

  • feat(alias): alias llama to llama-cpp, update docs by @mudler in #1448

Bug fixes 🐛

  • fix(piper): pin petals, phonemize and espeak by @mudler in #1393
  • update(llama.cpp): update server, correctly propagate LLAMA_VERSION by @mudler in #1440

Exciting New Features 🎉

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.0.0...v2.1.0