mirror of
https://github.com/containers/ramalama.git
synced 2026-02-05 06:46:39 +01:00
Add Podman >= 5.7.0 version requirement for artifact tests Artifact support requires Podman 5.7.0 or later. This commit adds version checking to skip artifact tests on systems with older Podman versions. Changes: - Add get_podman_version() and skip_if_podman_too_old() functions to test/system/helpers.bash - Add skip_if_podman_too_old decorator to test/conftest.py for e2e tests - Apply version check to all artifact-related tests in: - test/system/056-artifact.bats (BATS tests) - test/e2e/test_artifact.py (pytest e2e tests) - Tests will be skipped with clear message on systems with Podman < 5.7.0 - Docker and nocontainer tests are also skipped as they don't support artifacts The version check extracts the Podman version and compares it numerically, handling development versions like "5.7.0-dev" correctly. Cursor-AI-Generated Signed-off-by: Daniel J Walsh <dwalsh@redhat.com>
80 lines
3.5 KiB
Markdown
80 lines
3.5 KiB
Markdown
% ramalama-convert 1
|
|
|
|
## NAME
|
|
ramalama\-convert - convert AI Models from local storage to OCI Image
|
|
|
|
## SYNOPSIS
|
|
**ramalama convert** [*options*] *model* [*target*]
|
|
|
|
## DESCRIPTION
|
|
Convert specified AI Model to an OCI Formatted AI Model
|
|
|
|
The model can be from RamaLama model storage in Huggingface, Ollama, or a local model stored on disk. Converting from an OCI model is not supported.
|
|
|
|
Note: The convert command must be run with containers. Use of the --nocontainer option is not allowed.
|
|
|
|
## OPTIONS
|
|
|
|
#### **--gguf**=*Q2_K* | *Q3_K_S* | *Q3_K_M* | *Q3_K_L* | *Q4_0* | *Q4_K_S* | *Q4_K_M* | *Q5_0* | *Q5_K_S* | *Q5_K_M* | *Q6_K* | *Q8_0*
|
|
|
|
Convert Safetensor models into a GGUF with the specified quantization format. To learn more about model quantization, read llama.cpp documentation:
|
|
https://github.com/ggml-org/llama.cpp/blob/master/tools/quantize/README.md
|
|
|
|
#### **--help**, **-h**
|
|
Print usage message
|
|
|
|
#### **--image**=IMAGE
|
|
Image to use for model quantization when converting to GGUF format (when the `--gguf` option has been specified). The image must have the
|
|
`llama-quantize` executable available on the `PATH`. Defaults to the appropriate `ramalama` image based on available accelerators. If no
|
|
accelerators are available, the current `quay.io/ramalama/ramalama` image will be used.
|
|
|
|
#### **--network**=*none*
|
|
sets the configuration for network namespaces when handling RUN instructions
|
|
|
|
#### **--pull**=*policy*
|
|
Pull image policy. The default is **missing**.
|
|
|
|
#### **--rag-image**=IMAGE
|
|
Image to use when converting to GGUF format (when then `--gguf` option has been specified). The image must have the `convert_hf_to_gguf.py` script
|
|
executable and available in the `PATH`. The script is available from the `llama.cpp` GitHub repo. Defaults to the current
|
|
`quay.io/ramalama/ramalama-rag` image.
|
|
|
|
#### **--type**="artifact" | *raw* | *car*
|
|
|
|
Convert the MODEL to the specified OCI Object
|
|
|
|
| Type | Description |
|
|
| -------- | ------------------------------------------------------------- |
|
|
| artifact | Store AI Models as artifacts |
|
|
| car | Traditional OCI image including base image with the model stored in a /models subdir |
|
|
| raw | Traditional OCI image including only the model and a link file `model.file` pointed at it stored at / |
|
|
|
|
## EXAMPLE
|
|
|
|
Generate an oci model out of an Ollama model.
|
|
```
|
|
$ ramalama convert ollama://tinyllama:latest oci://quay.io/rhatdan/tiny:latest
|
|
Building quay.io/rhatdan/tiny:latest...
|
|
STEP 1/2: FROM scratch
|
|
STEP 2/2: COPY sha256:2af3b81862c6be03c769683af18efdadb2c33f60ff32ab6f83e42c043d6c7816 /model
|
|
--> Using cache 69db4a10191c976d2c3c24da972a2a909adec45135a69dbb9daeaaf2a3a36344
|
|
COMMIT quay.io/rhatdan/tiny:latest
|
|
--> 69db4a10191c
|
|
Successfully tagged quay.io/rhatdan/tiny:latest
|
|
69db4a10191c976d2c3c24da972a2a909adec45135a69dbb9daeaaf2a3a36344
|
|
```
|
|
|
|
Generate and run an oci model with a quantized GGUF converted from Safetensors.
|
|
```
|
|
$ ramalama convert --gguf Q4_K_M hf://ibm-granite/granite-3.2-2b-instruct oci://quay.io/kugupta/granite-3.2-q4-k-m:latest
|
|
Converting /Users/kugupta/.local/share/ramalama/models/huggingface/ibm-granite/granite-3.2-2b-instruct to quay.io/kugupta/granite-3.2-q4-k-m:latest...
|
|
Building quay.io/kugupta/granite-3.2-q4-k-m:latest...
|
|
$ ramalama run oci://quay.io/kugupta/granite-3.2-q4-k-m:latest
|
|
```
|
|
|
|
## SEE ALSO
|
|
**[ramalama(1)](ramalama.1.md)**, **[ramalama-push(1)](ramalama-push.1.md)**
|
|
|
|
## HISTORY
|
|
Aug 2024, Originally compiled by Eric Curtin <ecurtin@redhat.com>
|