Skip to content

Releases: containers/ramalama

v0.7.1

28 Mar 01:25

Choose a tag to compare

What's Changed

Full Changelog: v0.7.0...v0.7.1

v0.7.0

26 Mar 14:18

Choose a tag to compare

This is a big release, We now have working support for RAG inside of RamaLama.
Try out
ramalama rag XYZ.pdf ABC.doc quay.io/NAME/myrag
ramalama run --rag quay.io/NAME/myrag MYMODEL

What's Changed

New Contributors

Full Changelog: v0.6.4...v0.7.0

v0.6.4

17 Mar 18:21
e18d780

Choose a tag to compare

What's Changed

  • Print error when converting from an OCI Image by @rhatdan in #932
  • Make compatible with the macOS system python3 by @ericcurtin in #933
  • Bugfixes noticed while installing on Raspberry Pi by @ericcurtin in #935
  • Add note about updating nvidia.yaml file by @rhatdan in #938
  • Fix docker handling of GPUs. by @rhatdan in #941
  • macOS detection fix by @ericcurtin in #942
  • Add chat template support by @engelmi in #917
  • Consolidate gpu detection by @ericcurtin in #943
  • Implement RamaLama shell by @ericcurtin in #915
  • Add Linux x86-64 support for Ascend NPU accelerator in llama.cpp backend by @leo-pony in #950
  • Handle CNAI annotation deprecation by @s3rj1k in #939
  • Fix install.sh for OSTree system by @ericcurtin in #951
  • Lets run container in all tests, to make sure it does not explode. by @rhatdan in #946
  • Added --chat-template-file support to ramalama serve by @engelmi in #952
  • Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1741850090 by @renovate in #956
  • Add specified nvidia-oci runtime by @rhatdan in #953
  • python3 validator by @ericcurtin in #959
  • There must be at least one CDI device present to use CUDA by @ericcurtin in #954
  • [NPU][Fix] only specify device num, but without ascend-docker-runtime installed, running ramalama/cann container image will failing by @leo-pony in #962
  • Fix port rendering in README by @andreadecorte in #963
  • Update docker.io/nvidia/cuda Docker tag to v12.8.1 by @renovate in #960
  • Update llama.cpp to contain threads features by @ericcurtin in #967
  • Fix ENTRYPOINTS of whisper-server and llama-server by @rhatdan in #965
  • Add software to support using rag in RamaLama by @rhatdan in #968
  • Update llama.cpp for some Gemma features by @ericcurtin in #973
  • Only set this environment variable if we can resolve CDI by @ericcurtin in #971
  • feat(cpu): add --threads option to specify number of cpu threads by @antheas in #966
  • Asashi build is failing because of no python3-devel package by @rhatdan in #974
  • GPG Check is failing on the Intel Repo by @cgruver in #976
  • Add --runtime-arg option for run and serve by @edmcman in #949
  • Fix handling of whisper-server and llama-server entrypoints by @rhatdan in #975
  • Bump to v0.6.4 by @rhatdan in #978

New Contributors

Full Changelog: v0.6.3...v0.6.4

v0.6.3

10 Mar 17:50
72c5faf

Choose a tag to compare

What's Changed

  • Check if terminal is compatible with emojis before using them by @ericcurtin in #878
  • Use vllm-openai upstream image by @ericcurtin in #880
  • The package available via dnf is in a good place by @ericcurtin in #879
  • Add Ollama to CI and system tests for its caching by @kush-gupt in #881
  • Moved pruning protocol from model to factory by @engelmi in #882
  • Remove emoji usage until linenoise.cpp and llama-run are compatible by @ericcurtin in #884
  • Inject config to cli functions by @engelmi in #889
  • Switch from tiny to smollm:135m by @ericcurtin in #891
  • benchmark failing because of lack of flag by @ericcurtin in #888
  • Update the README.md to point people at ramalama.ai web site by @rhatdan in #894
  • fix: handling of date with python 3.8/3.9/3.10 by @benoitf in #897
  • readme: fix artifactory link by @alaviss in #903
  • Added support for mac cpu and clear warning message by @bmahabirbu in #902
  • Use python variable instead of environment variable by @ericcurtin in #907
  • Update llama.cpp by @ericcurtin in #908
  • Build a non-kompute Vulkan container image by @ericcurtin in #910
  • Reintroduce emoji prompts by @ericcurtin in #913
  • Add new ramalama-*-core executables by @ericcurtin in #909
  • Detect & get info on hugging face repos, fix sizing of symlinked directories by @kush-gupt in #901
  • Add ramalama image built on Fedora using Fedora's rocm packages by @maxamillion in #596
  • Add new model store by @engelmi in #905
  • Add support for llama.cpp engine to use ascend NPU device by @leo-pony in #911
  • Extend make validate check to do more by @ericcurtin in #916
  • Modify GPU detection to match against env var value instead of prefix by @cgruver in #919
  • Add Intel ARC 155H to list of supported hardware by @cgruver in #920
  • Try to choose a free port on serve if default one is not available by @andreadecorte in #898
  • Add passing of environment variables to ramalama commands by @rhatdan in #922
  • Allow user to specify the images to use per hardware by @rhatdan in #921
  • fix: CHAT_FORMAT variable should be expanded by @benoitf in #926
  • Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1741600006 by @renovate in #928
  • Bump to v0.6.3 by @rhatdan in #931

New Contributors

Full Changelog: v0.6.2...v0.6.3

v0.6.2

24 Feb 15:45
00839ee

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.6.1...v0.6.2

v0.6.1

18 Feb 14:49
042f035

Choose a tag to compare

What's Changed

  • chore: use absolute link for the RamaLama logo by @benoitf in #781
  • Reuse Ollama cached image when available by @kush-gupt in #782
  • Add env var RAMALAMA_GPU_DEVICE to allow for explicit declaration of the GPU device to use by @cgruver in #773
  • Change RAMALAMA_GPU_DEVICE to RAMALAMA_DEVICE for AI accelerator device override by @cgruver in #786
  • Add Security information to README.md by @rhatdan in #787
  • Fix exiting on llama-serve when user hits ^c by @rhatdan in #785
  • Check if file exists before sorting them into a list by @kush-gupt in #784
  • Add ramalama run --keepalive option by @rhatdan in #789
  • Stash output from container_manager by @rhatdan in #790
  • Install llama.cpp for mac and nocontainer tests by @rhatdan in #792
  • _engine is set to None or has a value by @ericcurtin in #793
  • Only run dnf commands on platforms that have dnf by @ericcurtin in #794
  • Add ramalama rag command by @rhatdan in #501
  • Attempt to use build_llama_and_whisper.sh by @rhatdan in #795
  • Change --network-mode to --network by @ericcurtin in #800
  • Add some more gfx values to the default list by @ericcurtin in #806
  • Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1739449058 by @renovate in #808
  • Prepare containers to run with ai-lab-recipes by @rhatdan in #803
  • If ngl is not specified by @ericcurtin in #802
  • feat: add ramalama labels about the execution on top of container by @benoitf in #810
  • Add run and serve arguments for --device and --privileged by @cgruver in #809
  • chore: rewrite readarray function to make it portable by @benoitf in #815
  • chore: replace RAMALAMA label by ai.ramalama by @benoitf in #814
  • Upgrade from 6.3.1 to 6.3.2 by @ericcurtin in #816
  • Removed error wrapping in urlopen by @engelmi in #818
  • Encountered a bug where this function was returning -1 by @ericcurtin in #817
  • Align runtime arguments with run, serve, bench, and perplexity by @cgruver in #820
  • README: fix inspect command description by @kush-gupt in #826
  • Pin dev dependencies to major version and improve formatting + linting by @engelmi in #824
  • README: Fix typo by @bupd in #827
  • Switch apt-get to apt by @ericcurtin in #832
  • Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1739751568 by @renovate in #834
  • Add entrypoint container images by @rhatdan in #819
  • HuggingFace Cache Implementation by @kush-gupt in #833
  • Make serve by default expose network by @ericcurtin in #830
  • Fix up man page help verifacation by @rhatdan in #835
  • Fix handling of --privileged flag by @rhatdan in #821
  • chore: fix links of llama.cpp repository by @benoitf in #841
  • Unify CLI options (verbosity, version) by @mkesper in #685
  • Add system tests to pull from the Hugging Face cache by @kush-gupt in #846
  • Just one add_argument call for --dryrun/--dry-run by @ericcurtin in #847
  • Fix ramalama info to display NVIDIA and amd GPU information by @rhatdan in #848
  • Remove LICENSE header from gpu_detector.py by @ericcurtin in #850
  • Allowing modification of pull policy by @rhatdan in #843
  • Include instructions for installing on Fedora 42+ by @stefwalter in #849
  • Bump to 0.6.1 by @rhatdan in #851

New Contributors

Full Changelog: v0.6.0...v0.6.1

v0.6.0

10 Feb 18:23
14c9d35

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.5.5...v0.6.0

v0.5.5

01 Feb 12:42
c2c955b

Choose a tag to compare

What's Changed

  • Add perplexity subcommand to RamaLama CLI by @ericcurtin in #637
  • throwing an exception with there is a failure in http_client.init by @jhjaggars in #647
  • Add container image to support Intel ARC GPU by @cgruver in #644
  • Guide users to install huggingface-cli to login to huggingface by @pbabinca in #645
  • Update intel-gpu Containerfile to reduce the size of the builder image by @cgruver in #657
  • Look for configs also in /usr/local/share/ramalama by @jistr in #672
  • remove ro as an option when mounting images by @kush-gupt in #676
  • Add generated man pages for section 7 into gitignore by @jistr in #673
  • Revert "Added --jinja to llama-run command" by @ericcurtin in #683
  • Pull the source model if it isn't already in local storage for the convert and push functions by @kush-gupt in #680
  • bump llama.cpp to latest release hash aa6fb13 by @maxamillion in #692
  • Introduce a mode so one call install from git by @ericcurtin in #690
  • Add ramalama gpu_detector by @dougsland in #670
  • Bump to v0.5.5 by @rhatdan in #701

New Contributors

Full Changelog: v0.5.4...v0.5.5

v0.5.4

28 Jan 21:36
7bc9164

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.5.3...v0.5.4

v0.5.3

23 Jan 12:39
1cbb5b0

Choose a tag to compare

What's Changed

New Contributors

Full Changelog: v0.5.2...v0.5.3