Skip to content

Releases: BodhiSearch/llama.cpp

Llama Server Windows build 8cd21747

17 Mar 04:08
Compare
Choose a tag to compare

Llama Server Windows Builds

Automated build from commit 8cd2174

Available Builds

  • llama-server-macos-cpu
  • llama-server-macos-metal
  • llama-server-windows-generic.exe
  • llama-server-windows-haswell.exe

Llama Server Windows build 8c63e5b9

17 Mar 11:58
Compare
Choose a tag to compare

Llama Server Windows Builds

Automated build from commit 8c63e5b

Available Builds

  • llama-server-macos-cpu
  • llama-server-macos-metal
  • llama-server-windows-alderlake.exe
  • llama-server-windows-generic.exe
  • llama-server-windows-haswell.exe
  • llama-server-windows-icelake.exe
  • llama-server-windows-sandybridge.exe
  • llama-server-windows-skylakex.exe
  • llama-server-windows-sse42.exe
  • llama-server-windows-zen4.exe

Llama Server Windows build 6d9a3328

15 Mar 16:45
Compare
Choose a tag to compare

Llama Server Windows Builds

Automated build from commit 6d9a332

Available Builds

  • llama-server-generic.exe
  • llama-server-haswell.exe

Llama Server Windows build 451fda4a

16 Mar 07:52
Compare
Choose a tag to compare

Llama Server Windows Builds

Automated build from commit 451fda4

Available Builds

  • llama-server-generic.exe
  • llama-server-haswell.exe
  • llama-server-macos-cpu
  • llama-server-macos-metal

Llama Server Build 396c05da

19 Mar 08:55
Compare
Choose a tag to compare

Llama Server Windows Builds

Automated build from commit 396c05d

Available Builds

  • llama-server-linux-generic
  • llama-server-linux-haswell
  • llama-server-linux-icelake
  • llama-server-linux-sandybridge
  • llama-server-linux-sapphirerapids
  • llama-server-linux-skylakex
  • llama-server-linux-sse42
  • llama-server-linux-zen4
  • llama-server-macos-cpu
  • llama-server-macos-metal
  • llama-server-windows-alderlake.exe
  • llama-server-windows-generic.exe
  • llama-server-windows-haswell.exe
  • llama-server-windows-icelake.exe
  • llama-server-windows-sandybridge.exe
  • llama-server-windows-sapphirerapids.exe
  • llama-server-windows-skylakex.exe
  • llama-server-windows-sse42.exe
  • llama-server-windows-zen4.exe

server-bodhiapp_20250126-b4559-d95a6fb

27 Jan 03:17
Compare
Choose a tag to compare
[BodhiApp] changes for bodhiapp.

[21-dec-24] using prompt if passed in chat completions, and not using messages
- added python integration tests for changes in server
- having add_special as request param to pass to upstream, allows pre-formatted chat messages to not be formatted again
- modified workflow to download and cache the llama2-7b model used for integration testing
[17-jan-25] updated to latest llama.cpp
server.cpp had changes where ctx_server.vocab was used instead of the earlier ctx_server.ctx
[26-jan-25] merged examples/server/utils.hpp, llama.cpp is starting to support tools in the chat template, resolved rebase conflict by merging our logic of preformatted messages with new feature
adding the github workflow to create the artifacts based on platform and gpu architecture, for macos-cpu+metal, ubuntu-cpu+cuda, windows-cpu+cuda-12.4+11.6
also added for other ubuntu+windows cpu variants and gpu archs like vulkan etc., but disabled for now

server-bodhiapp_20250126-b4559-2cfb3f6

27 Jan 12:25
Compare
Choose a tag to compare
[BodhiApp] changes for bodhiapp.

[21-dec-24] using prompt if passed in chat completions, and not using messages
- added python integration tests for changes in server
- having add_special as request param to pass to upstream, allows pre-formatted chat messages to not be formatted again
- modified workflow to download and cache the llama2-7b model used for integration testing
[17-jan-25] updated to latest llama.cpp
server.cpp had changes where ctx_server.vocab was used instead of the earlier ctx_server.ctx
[26-jan-25] merged examples/server/utils.hpp, llama.cpp is starting to support tools in the chat template, resolved rebase conflict by merging our logic of preformatted messages with new feature
adding the github workflow to create the artifacts based on platform and gpu architecture, for macos-cpu+metal, ubuntu-cpu+cuda, windows-cpu+cuda-12.4+11.6
also added for other ubuntu+windows cpu variants and gpu archs like vulkan etc., but disabled for now

server-bodhiapp_20250126-b4559-1a6ec68

27 Jan 13:32
Compare
Choose a tag to compare
[BodhiApp] changes for bodhiapp.

[21-dec-24] using prompt if passed in chat completions, and not using messages
- added python integration tests for changes in server
- having add_special as request param to pass to upstream, allows pre-formatted chat messages to not be formatted again
- modified workflow to download and cache the llama2-7b model used for integration testing
[17-jan-25] updated to latest llama.cpp
server.cpp had changes where ctx_server.vocab was used instead of the earlier ctx_server.ctx
[26-jan-25] merged examples/server/utils.hpp, llama.cpp is starting to support tools in the chat template, resolved rebase conflict by merging our logic of preformatted messages with new feature
adding the github workflow to create the artifacts based on platform and gpu architecture, for macos-cpu+metal, ubuntu-cpu+cuda, windows-cpu+cuda-12.4+11.6
also added for other ubuntu+windows cpu variants and gpu archs like vulkan etc., but disabled for now