1
0
mirror of https://github.com/containers/ramalama.git synced 2026-02-05 15:47:26 +01:00
Files
ramalama/docs/ramalama-convert.1.md
Daniel J Walsh 50222f9458 Add support for converting to OCI artifacts
Add Podman >= 5.7.0 version requirement for artifact tests

Artifact support requires Podman 5.7.0 or later. This commit adds version
checking to skip artifact tests on systems with older Podman versions.

Changes:
- Add get_podman_version() and skip_if_podman_too_old() functions to test/system/helpers.bash
- Add skip_if_podman_too_old decorator to test/conftest.py for e2e tests
- Apply version check to all artifact-related tests in:
  - test/system/056-artifact.bats (BATS tests)
  - test/e2e/test_artifact.py (pytest e2e tests)
- Tests will be skipped with clear message on systems with Podman < 5.7.0
- Docker and nocontainer tests are also skipped as they don't support artifacts

The version check extracts the Podman version and compares it numerically,
handling development versions like "5.7.0-dev" correctly.

Cursor-AI-Generated

Signed-off-by: Daniel J Walsh <dwalsh@redhat.com>
2025-12-15 08:16:39 -05:00

3.5 KiB

% ramalama-convert 1

NAME

ramalama-convert - convert AI Models from local storage to OCI Image

SYNOPSIS

ramalama convert [options] model [target]

DESCRIPTION

Convert specified AI Model to an OCI Formatted AI Model

The model can be from RamaLama model storage in Huggingface, Ollama, or a local model stored on disk. Converting from an OCI model is not supported.

Note: The convert command must be run with containers. Use of the --nocontainer option is not allowed.

OPTIONS

--gguf=Q2_K | Q3_K_S | Q3_K_M | Q3_K_L | Q4_0 | Q4_K_S | Q4_K_M | Q5_0 | Q5_K_S | Q5_K_M | Q6_K | Q8_0

Convert Safetensor models into a GGUF with the specified quantization format. To learn more about model quantization, read llama.cpp documentation: https://github.com/ggml-org/llama.cpp/blob/master/tools/quantize/README.md

--help, -h

Print usage message

--image=IMAGE

Image to use for model quantization when converting to GGUF format (when the --gguf option has been specified). The image must have the llama-quantize executable available on the PATH. Defaults to the appropriate ramalama image based on available accelerators. If no accelerators are available, the current quay.io/ramalama/ramalama image will be used.

--network=none

sets the configuration for network namespaces when handling RUN instructions

--pull=policy

Pull image policy. The default is missing.

--rag-image=IMAGE

Image to use when converting to GGUF format (when then --gguf option has been specified). The image must have the convert_hf_to_gguf.py script executable and available in the PATH. The script is available from the llama.cpp GitHub repo. Defaults to the current quay.io/ramalama/ramalama-rag image.

--type="artifact" | raw | car

Convert the MODEL to the specified OCI Object

Type Description
artifact Store AI Models as artifacts
car Traditional OCI image including base image with the model stored in a /models subdir
raw Traditional OCI image including only the model and a link file model.file pointed at it stored at /

EXAMPLE

Generate an oci model out of an Ollama model.

$ ramalama convert ollama://tinyllama:latest oci://quay.io/rhatdan/tiny:latest
Building quay.io/rhatdan/tiny:latest...
STEP 1/2: FROM scratch
STEP 2/2: COPY sha256:2af3b81862c6be03c769683af18efdadb2c33f60ff32ab6f83e42c043d6c7816 /model
--> Using cache 69db4a10191c976d2c3c24da972a2a909adec45135a69dbb9daeaaf2a3a36344
COMMIT quay.io/rhatdan/tiny:latest
--> 69db4a10191c
Successfully tagged quay.io/rhatdan/tiny:latest
69db4a10191c976d2c3c24da972a2a909adec45135a69dbb9daeaaf2a3a36344

Generate and run an oci model with a quantized GGUF converted from Safetensors.

$ ramalama convert --gguf Q4_K_M hf://ibm-granite/granite-3.2-2b-instruct oci://quay.io/kugupta/granite-3.2-q4-k-m:latest
Converting /Users/kugupta/.local/share/ramalama/models/huggingface/ibm-granite/granite-3.2-2b-instruct to quay.io/kugupta/granite-3.2-q4-k-m:latest...
Building quay.io/kugupta/granite-3.2-q4-k-m:latest...
$ ramalama run oci://quay.io/kugupta/granite-3.2-q4-k-m:latest

SEE ALSO

ramalama(1), ramalama-push(1)

HISTORY

Aug 2024, Originally compiled by Eric Curtin ecurtin@redhat.com