1
0
mirror of https://github.com/containers/ramalama.git synced 2026-02-05 06:46:39 +01:00
Commit Graph

2935 Commits

Author SHA1 Message Date
Daniel J Walsh
eca00b131e Merge pull request #1866 from rhatdan/VERSION
Bump to v0.12.1
v0.12.1
2025-08-26 09:07:43 -04:00
Daniel J Walsh
6ba9a13792 Bump to v0.12.1
Signed-off-by: Daniel J Walsh <dwalsh@redhat.com>
2025-08-26 07:03:52 -04:00
Daniel J Walsh
5cc9fa1cff Merge pull request #1865 from containers/konflux/references/main
chore(deps): update konflux references
2025-08-25 08:14:39 -04:00
Brian M
c84798fe46 Merge pull request #1761 from bmahabirbu/rag-port
feat: added wait logic for rag_framework fixed doc2rag file and added…
2025-08-24 02:03:27 -04:00
Brian
521f7dd121 feat: added wait logic for rag_framework fixed doc2rag file and added env port logic
Signed-off-by: Brian <bmahabir@bu.edu>
2025-08-23 23:42:51 -04:00
red-hat-konflux-kflux-prd-rh03[bot]
57c1099462 chore(deps): update konflux references
Signed-off-by: red-hat-konflux-kflux-prd-rh03 <206760901+red-hat-konflux-kflux-prd-rh03[bot]@users.noreply.github.com>
2025-08-23 12:05:19 +00:00
Daniel J Walsh
1215ccd369 Merge pull request #1861 from olliewalsh/cifix
Fix python environment in ci jobs
2025-08-21 08:09:00 -04:00
Daniel J Walsh
1e4b539fb7 Merge pull request #1862 from olliewalsh/cifixtest
DNM: ci test
2025-08-21 08:07:55 -04:00
Daniel J Walsh
48dbd28afd Merge pull request #1863 from containers/konflux/mintmaker/main/registry.access.redhat.com-ubi9-ubi-9.x
chore(deps): update registry.access.redhat.com/ubi9/ubi docker tag to v9.6-1755678605
2025-08-21 08:06:17 -04:00
red-hat-konflux-kflux-prd-rh03[bot]
c5480fce52 chore(deps): update registry.access.redhat.com/ubi9/ubi docker tag to v9.6-1755678605
Signed-off-by: red-hat-konflux-kflux-prd-rh03 <206760901+red-hat-konflux-kflux-prd-rh03[bot]@users.noreply.github.com>
2025-08-21 04:05:30 +00:00
Oliver Walsh
6bcac9ddd7 Fix python environment in ci jobs
Was installing uv 2 different ways (github action and make install-requirement).
Use `uv run` explicitly instead of activating the virtualenv in the uv
github action (it's docs advise not to use this option).

Signed-off-by: Oliver Walsh <owalsh@redhat.com>
2025-08-20 22:42:39 +01:00
Chad Schellenger
075e894e94 Use PyYAML to read Nvidia CDI config
Signed-off-by: Oliver Walsh <owalsh@redhat.com>
2025-08-20 21:59:10 +01:00
Oliver Walsh
c8485b91ea Fix python environment in ci jobs
Was installing uv 2 different ways (github action and make install-requirement).
Use `uv run` explicitly instead of activating the virtualenv in the uv
github action (it's docs advise not to use this option).

Signed-off-by: Oliver Walsh <owalsh@redhat.com>
2025-08-20 21:50:16 +01:00
Michael Engel
69345751de Merge pull request #1858 from olliewalsh/deepseek_chat_template
Fix deepseek-r1 chat template conversion
2025-08-20 15:50:27 +02:00
Daniel J Walsh
9162e77723 Merge pull request #1854 from kpouget/build_container
Make the container build system more flexible
2025-08-20 05:43:27 -04:00
Michael Engel
73017bad20 Merge pull request #1850 from booxter/errno-darwin
fix: Don't use non-portable errno.{EMEDIUMTYPE,ENOMEDIUM}
2025-08-20 09:32:12 +02:00
Michael Engel
1c87ce7575 Merge pull request #1856 from olliewalsh/ollama_chat_template
Use correct chat template file for ollama models
2025-08-20 09:30:33 +02:00
Ihar Hrachyshka
83390f1b20 fix: Don't use non-portable errno.{EMEDIUMTYPE,ENOMEDIUM}
They are not part of POSIX and are Linux specific. This code crashes on
Darwin because of missing symbols.

Signed-off-by: Ihar Hrachyshka <ihar.hrachyshka@gmail.com>
2025-08-19 20:24:10 -04:00
Oliver Walsh
0c17e2b271 Fix deepseek-r1 chat template conversion
From https://github.com/engelmi/go2jinja/pull/5

Fixes https://github.com/containers/ramalama/issues/1857

Signed-off-by: Oliver Walsh <owalsh@redhat.com>
2025-08-19 18:38:50 +01:00
Oliver Walsh
412ec37a93 Use correct chat template file for ollama models
ollama uses go-template for chat template files. llama.cpp uses jinja2.
The go-templates are converted to jinja2 but since the first
chat template file is always chosen this file is not used.
Pick the last file instead of the first to resolve this.

Fixes #1855

Signed-off-by: Oliver Walsh <owalsh@redhat.com>
2025-08-19 17:11:08 +01:00
Kevin Pouget
52a36cd2dc Call clone_and_build_llama_cpp from the right function
Signed-off-by: Kevin Pouget <kpouget@redhat.com>
2025-08-19 15:29:58 +02:00
Kevin Pouget
8a8ce838d0 container_build.sh: pass environment variables to the image builder
Signed-off-by: Kevin Pouget <kpouget@redhat.com>
2025-08-19 15:29:58 +02:00
Kevin Pouget
d81e91864d container-images: make git_clone_specific_commit more reliable
Signed-off-by: Kevin Pouget <kpouget@redhat.com>
2025-08-19 15:29:58 +02:00
Kevin Pouget
41e84c29ba Allow controlling the llama/whisper repo/commit used in the image
Signed-off-by: Kevin Pouget <kpouget@redhat.com>
2025-08-19 15:29:58 +02:00
Kevin Pouget
3ae4dccb53 build_llama_and_whisper.sh: allow building the image in debug mode
Signed-off-by: Kevin Pouget <kpouget@redhat.com>
2025-08-19 15:29:58 +02:00
Daniel J Walsh
a4390f5151 Merge pull request #1846 from olliewalsh/progress
Fix progress bar not reaching 100%
2025-08-18 08:50:06 -04:00
Oliver Walsh
62ee0b1135 Fix progress bar not reaching 100%
Signed-off-by: Oliver Walsh <owalsh@redhat.com>
2025-08-18 12:36:18 +01:00
Daniel J Walsh
7bf82d06c7 Merge pull request #1848 from olliewalsh/hf_split_models
Add support for huggingface split GGUF models
2025-08-18 06:27:22 -04:00
Oliver Walsh
2ce3389a65 Add support for hf split GGUF models
Build and pull the complete list of files for split GGUF models.
Also handle case where the model is stored in a subdirectory of
the repo.

Signed-off-by: Oliver Walsh <owalsh@redhat.com>
2025-08-17 21:35:21 +01:00
Daniel J Walsh
2c94e5f39c Merge pull request #1851 from containers/konflux/references/main
chore(deps): update konflux references
2025-08-16 08:01:39 -04:00
Daniel J Walsh
015cbf5c86 Merge pull request #1847 from olliewalsh/test_fixes
Bats test fixes
2025-08-16 08:00:55 -04:00
red-hat-konflux-kflux-prd-rh03[bot]
1c0c741e27 chore(deps): update konflux references
Signed-off-by: red-hat-konflux-kflux-prd-rh03 <206760901+red-hat-konflux-kflux-prd-rh03[bot]@users.noreply.github.com>
2025-08-16 08:03:34 +00:00
Oliver Walsh
bab6e29083 Bats test fixes
Fix model used in generate quadlet tests
Use smaller model for hf pull test

Signed-off-by: Oliver Walsh <owalsh@redhat.com>
2025-08-15 17:33:56 +01:00
Daniel J Walsh
4a2c2a0750 Merge pull request #1840 from containers/konflux-build-sa-migration
konflux: migrate to per-component service accounts
2025-08-15 05:44:29 -04:00
Daniel J Walsh
85c6940561 Merge pull request #1841 from containers/konflux-remove-ecosystem-checks
konflux: remove ecosystem tests
2025-08-15 05:43:32 -04:00
Mike Bonnet
eecdb86135 konflux: remove ecosystem tests
The tests can take a long time and/or fail, and are not applicable to
these images.

Signed-off-by: Mike Bonnet <mikeb@redhat.com>
2025-08-14 16:27:17 -07:00
Mike Bonnet
c7f933a38a konflux: migrate to per-component service accounts
Konflux is moving from a shared "appstudio-pipeline" service account
to a separate service account for each component, to increase security
and improve efficiency.

Signed-off-by: Mike Bonnet <mikeb@redhat.com>
2025-08-14 16:15:23 -07:00
Daniel J Walsh
566dcb8592 Merge pull request #1833 from rhatdan/VERSION
Bump to v0.12.0
v0.12.0
2025-08-14 13:53:45 -04:00
Daniel J Walsh
558effaf6a Bump to v0.12.0
Signed-off-by: Daniel J Walsh <dwalsh@redhat.com>
2025-08-14 10:11:35 -04:00
Daniel J Walsh
0483a580a0 Merge pull request #1838 from engelmi/improve-no-model-file-error-message
Added error message for GGUF not found and safetensor unsupported
2025-08-14 09:55:21 -04:00
Michael Engel
27f69bf905 Added error message for GGUF not found and safetensor unsupported
Relates to: https://github.com/containers/ramalama/issues/1816

Signed-off-by: Michael Engel <mengel@redhat.com>
2025-08-14 14:04:56 +02:00
Daniel J Walsh
d802f2d3f5 Merge pull request #1767 from pbabinca/huggingface_cli
Improve error message when huggingface-cli is needed and missing
2025-08-14 06:42:20 -04:00
Daniel J Walsh
a6aba3400c Merge pull request #1834 from containers/llama-stack-image-config
allow the llama-stack image to be configured via an env var
2025-08-14 06:32:55 -04:00
Daniel J Walsh
1e418b1668 Merge pull request #1837 from containers/konflux/mintmaker/main/pre-commit-pre-commit-hooks-6.x
chore(deps): update pre-commit hook pre-commit/pre-commit-hooks to v6
2025-08-14 06:31:38 -04:00
Pavol Babincak
0827f6769b Improve error message when modelscope is needed and missing
Before:

    Error:
    Optional: ModelScope models require the modelscope module.
    This module can be installed via PyPI tools like uv, pip, pip3, pipx, or via
    distribution package managers like dnf or apt. Example:
    uv pip install modelscope

After:

    Error: This operation requires modelscope which is not available.

    This tool can be installed via PyPI tools like uv, pip, pip3 or pipx. Example:

    pip install modelscope

In particular:

    - follows the same structure of error message as in `huggingface` module

    - separate error from the explanation and guidance by new line

    - explain that this particular operation requires the CLI tool -
      currently the error is displayed on login, logout or push. Implies
      that some other operations might not require this tool

    - provide `pip` as first example how to solve the issue (that's also
      the tool mention in the README)

    - there aren't any distribution deb nor rpm packages with modelscope
      so remove that guidance

    - `modelscope` is a command line tool, not a module as it was
      referred before

Fixes: #1766

Signed-off-by: Pavol Babincak <pbabinca@redhat.com>
2025-08-14 11:00:43 +02:00
Pavol Babincak
55df39864d Improve error message when huggingface-cli is needed and missing
Before:

    Error:
    Optional: Huggingface models require the huggingface-cli module.
    This module can be installed via PyPI tools like uv, pip, pip3, pipx, or via
    distribution package managers like dnf or apt. Example:
    uv pip install huggingface_hub

After:

    Error: This operation requires huggingface-cli which is not available.

    This tool can be installed via PyPI tools like uv, pip, pip3 or pipx. Example:

    pip install -U "huggingface_hub[cli]"

    Or via distribution package managers like dnf or apt. Example:

    sudo dnf install python3-huggingface-hub

In particular:

    - separate error from the explanation and guidance by new line

    - explain that this particular operation requires the CLI tool -
      currently the error is displayed on login, logout or push. Implies
      that some other operations might not require this tool

    - provide `pip` as first example how to solve the issue (that's also
      the tool mention in the README)

    - use same package extra as official documentation to install the
      CLI tool: https://huggingface.co/docs/huggingface_hub/en/guides/cli

    - provide example with dnf command to install the package (dnf is
      also mentioned in the README file)

    - `huggingface-cli` is a command line tool, not a module as it was
      referred before

Fixes: #1766

Signed-off-by: Pavol Babincak <pbabinca@redhat.com>
2025-08-14 11:00:43 +02:00
red-hat-konflux-kflux-prd-rh03[bot]
8e1fe3a72e chore(deps): update pre-commit hook pre-commit/pre-commit-hooks to v6
Signed-off-by: red-hat-konflux-kflux-prd-rh03 <206760901+red-hat-konflux-kflux-prd-rh03[bot]@users.noreply.github.com>
2025-08-14 08:02:47 +00:00
Mike Bonnet
ac42aae2f0 allow the llama-stack image to be configured via an env var
Previously, the location of the "llama-stack" image was hard-coded.
Allow it to be configured via an env var, in the same way that the
ramalama and gpu-specific images are.

Update the integration tests to use the newly-built llama-stack image.

Signed-off-by: Mike Bonnet <mikeb@redhat.com>
2025-08-13 09:50:50 -07:00
Daniel J Walsh
47babb31c3 Merge pull request #1830 from containers/healthcheck-no-systemd
reimplement health checks outside of podman
2025-08-13 05:42:48 -04:00
Mike Bonnet
e37d07346c reimplement health checks outside of podman
Podman health checks rely on systemd timers to periodically run the check. When
not running under systemd, the health check config is silently ignored and the
check never runs, leaving the status permanently in the "starting" state. This
causes wait_for_healthy() to time out.

Reimplement health checks in ramalama by polling the "/models" API endpoint and
parsing the result. This avoids the dependency on systemd, and works when running
in non-podman environments.

Signed-off-by: Mike Bonnet <mikeb@redhat.com>
2025-08-12 22:29:21 -07:00