1
0
mirror of https://github.com/containers/ramalama.git synced 2026-02-05 06:46:39 +01:00

updates docsite and adds docsite to the make docs process

Signed-off-by: Ian Eaves <ian.k.eaves@gmail.com>
This commit is contained in:
Ian Eaves
2025-10-01 12:12:48 -05:00
parent ccc6e6a61f
commit b32ce823f4
31 changed files with 1629 additions and 1109 deletions

View File

@@ -100,9 +100,14 @@ build_multi_arch:
install-docs: docs
make -C docs install
.PHONY: docs
docs:
make -C docs
.PHONY: docs docs-manpages docsite-docs
docs: docs-manpages docsite-docs
docs-manpages:
$(MAKE) -C docs
docsite-docs:
$(MAKE) -C docsite convert
.PHONY: lint
lint:

View File

@@ -26,13 +26,15 @@ def extract_title_and_description(content, filename):
# Generate title from filename pattern
base_name = os.path.basename(filename)
title = None
if base_name == 'ramalama.1.md':
title = "ramalama" # Base command page
if base_name.startswith('ramalama-') and base_name.endswith('.1.md'):
title = 'ramalama' # Base command page
elif base_name.startswith('ramalama-') and base_name.endswith('.1.md'):
# Command: ramalama-chat.1.md -> chat (just the subcommand name)
command = base_name.replace('ramalama-', '').replace('.1.md', '')
title = command
if base_name.startswith('ramalama-') and base_name.endswith('.7.md'):
elif base_name.startswith('ramalama-') and base_name.endswith('.7.md'):
# Platform guide: ramalama-cuda.7.md -> CUDA Setup
platform = base_name.replace('ramalama-', '').replace('.7.md', '')
@@ -42,7 +44,7 @@ def extract_title_and_description(content, filename):
platform = platform.upper()
title = f"{platform} Setup"
if base_name.endswith('.5.md'):
elif base_name.endswith('.5.md'):
# Config files with custom titles
if base_name == 'ramalama.conf.5.md':
title = 'Configuration File'
@@ -51,8 +53,9 @@ def extract_title_and_description(content, filename):
else:
# Fallback for other .5.md files
title = base_name.replace('.5.md', '')
else:
# Fallback
if title is None:
# Fallback for any other file types
title = base_name.replace('.md', '').replace('-', ' ')
# Find description from NAME section
@@ -105,7 +108,7 @@ def detect_code_language(content):
return 'text'
def convert_markdown_to_mdx(content, filename):
def convert_markdown_to_mdx(content, filename, current_output_path, output_map):
"""Convert manpage markdown to MDX format"""
# Extract title and description
@@ -118,8 +121,13 @@ def convert_markdown_to_mdx(content, filename):
content = '\n'.join(lines)
# Convert NAME section
content = re.sub(r'## NAME\n([^\n]+)', '', content, flags=re.MULTILINE)
# Remove NAME section (handles both H1 and H2 variants)
content = re.sub(
r'^#{1,2}\s+NAME\s*\n(?:.*?)(?=^#{1,6}\s|\Z)',
'',
content,
flags=re.MULTILINE | re.DOTALL,
)
# Convert SYNOPSIS to proper heading
content = re.sub(r'## SYNOPSIS', '## Synopsis', content)
@@ -151,6 +159,8 @@ def convert_markdown_to_mdx(content, filename):
content = re.sub(r'\*\*\[([^\]]+)\]\(([^)]+)\)\*\*', r'[\1](\2)', content)
# Convert internal manpage links to docsite links
current_dir = current_output_path.parent
def convert_link(match):
text = match.group(1)
link = match.group(2)
@@ -159,38 +169,24 @@ def convert_markdown_to_mdx(content, filename):
if link.startswith(('http://', 'https://')):
return f'[{text}]({link})'
# All files are now in subdirectories, so use relative paths appropriately
if filename.endswith('.1.md'):
# From commands/ramalama/ directory
base_path = "../../"
elif filename.endswith('.5.md'):
# From configuration/ directory
base_path = "../"
elif filename.endswith('.7.md'):
# From platform-guides/ directory
base_path = "../"
else:
base_path = "./"
base_link = os.path.basename(link)
target_rel_path = output_map.get(base_link)
if link.endswith('.1.md'):
# Command reference
command_name = link.replace('ramalama-', '').replace('.1.md', '')
if command_name == 'ramalama':
if filename == 'ramalama.1.md':
return f'[{text}](#)' # Self-reference
else:
return f'[{text}](/docs/commands/ramalama/)' # Link to ramalama category index
return f'[{text}]({base_path}commands/ramalama/{command_name})'
if link.endswith('.5.md'):
# Configuration file
config_name = link.replace('ramalama.', '').replace('.5.md', '')
return f'[{text}]({base_path}configuration/{config_name})'
if link.endswith('.7.md'):
# Platform guide
guide_name = link.replace('ramalama-', '').replace('.7.md', '')
return f'[{text}]({base_path}platform-guides/{guide_name})'
if not target_rel_path:
return f'[{text}]({link})'
return f'[{text}]({link})'
target_path = target_rel_path
if target_path == current_output_path:
return f'[{text}](#)' # Self-reference
if target_path == Path('commands/ramalama/ramalama.mdx'):
return f'[{text}](/docs/commands/ramalama/)'
# Use absolute doc URL (prefix with /docs)
target_route = '/docs/' + target_path.with_suffix('').as_posix()
return f'[{text}]({target_route})'
content = re.sub(r'\[([^\]]+)\]\(([^)]+\.md)\)', convert_link, content)
@@ -324,30 +320,29 @@ def main():
print(f"\nFound {len(manpage_files)} manpage files to convert")
manpage_entries = []
for input_file in manpage_files:
filename = os.path.basename(input_file)
output_filename = get_output_filename(filename)
_, subdir = get_category_info(filename)
relative_output_path = Path(subdir) / output_filename
manpage_entries.append((input_file, filename, relative_output_path))
output_map = {filename: relative_path for _, filename, relative_path in manpage_entries}
for input_file, filename, relative_output_path in manpage_entries:
print(f"Converting {filename}...")
# Convert all files (overwriting existing ones)
# Read input file
with open(input_file, 'r', encoding='utf-8') as f:
content = f.read()
# Convert to MDX
mdx_content = convert_markdown_to_mdx(content, filename)
mdx_content = convert_markdown_to_mdx(content, filename, relative_output_path, output_map)
# Determine output directory and filename
output_filename = get_output_filename(filename)
# All manpage files go to their category directories
category, subdir = get_category_info(filename)
output_dir = docsite_docs_dir / subdir
output_path = output_dir / output_filename
# Create output directory if it doesn't exist
output_dir = (docsite_docs_dir / relative_output_path).parent
output_dir.mkdir(parents=True, exist_ok=True)
# Write output file
output_path = docsite_docs_dir / relative_output_path
with open(output_path, 'w', encoding='utf-8') as f:
f.write(mdx_content)

View File

@@ -1,11 +1,11 @@
---
title: ramalama bench.1
title: bench
description: benchmark specified AI Model
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-bench.1.md
---
# ramalama bench.1
# bench
## Synopsis
**ramalama bench** [*options*] *model* [arg ...]
@@ -63,7 +63,7 @@ OCI container image to run with specified AI model. RamaLama defaults to using
images based on the accelerator it discovers. For example:
`quay.io/ramalama/ramalama`. See the table below for all default images.
The default image tag is based on the minor version of the RamaLama package.
Version 0.12.1 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
Version 0.12.3 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
The default can be overridden in the ramalama.conf file or via the
RAMALAMA_IMAGE environment variable. `export RAMALAMA_IMAGE=quay.io/ramalama/aiimage:1.2` tells

View File

@@ -1,11 +1,11 @@
---
title: ramalama chat.1
title: chat
description: OpenAI chat with the specified REST API URL
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-chat.1.md
---
# ramalama chat.1
# chat
## Synopsis
**ramalama chat** [*options*] [arg...]
@@ -21,7 +21,7 @@ Chat with an OpenAI Rest API
#### **--api-key**
OpenAI-compatible API key.
Can also be set via the API_KEY environment variable.
Can also be set via the RAMALAMA_API_KEY environment variable.
#### **--color**
Indicate whether or not to use color in the chat.
@@ -33,6 +33,11 @@ Show this help message and exit
#### **--list**
List the available models at an endpoint
#### **--mcp**=SERVER_URL
MCP (Model Context Protocol) servers to use for enhanced tool calling capabilities.
Can be specified multiple times to connect to multiple MCP servers.
Each server provides tools that can be automatically invoked during chat conversations.
#### **--model**=MODEL
Model for inferencing (may not be required for endpoints that only serve one model)

View File

@@ -1,11 +1,11 @@
---
title: ramalama containers.1
title: containers
description: list all RamaLama containers
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-containers.1.md
---
# ramalama containers.1
# containers
## Synopsis
**ramalama containers** [*options*]

View File

@@ -1,11 +1,11 @@
---
title: ramalama convert.1
title: convert
description: convert AI Models from local storage to OCI Image
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-convert.1.md
---
# ramalama convert.1
# convert
## Synopsis
**ramalama convert** [*options*] *model* [*target*]
@@ -65,7 +65,7 @@ $ ramalama run oci://quay.io/kugupta/granite-3.2-q4-k-m:latest
```
## See Also
[ramalama(1)](/docs/commands/ramalama/), [ramalama-push(1)](../../commands/ramalama/push)
[ramalama(1)](/docs/commands/ramalama/), [ramalama-push(1)](/docs/commands/ramalama/push)
---

View File

@@ -1,11 +1,11 @@
---
title: ramalama daemon.1
title: daemon
description: run a RamaLama REST server
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-daemon.1.md
---
# ramalama daemon.1
# daemon
## Synopsis
**ramalama daemon** [*options*] [start|run]

View File

@@ -1,11 +1,11 @@
---
title: ramalama info.1
title: info
description: display RamaLama configuration information
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-info.1.md
---
# ramalama info.1
# info
## Synopsis
**ramalama info** [*options*]

View File

@@ -1,11 +1,11 @@
---
title: ramalama inspect.1
title: inspect
description: inspect the specified AI Model
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-inspect.1.md
---
# ramalama inspect.1
# inspect
## Synopsis
**ramalama inspect** [*options*] *model*

View File

@@ -1,11 +1,11 @@
---
title: ramalama list.1
title: list
description: list all downloaded AI Models
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-list.1.md
---
# ramalama list.1
# list
## Synopsis
**ramalama list** [*options*]

View File

@@ -1,11 +1,11 @@
---
title: ramalama login.1
title: login
description: login to remote registry
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-login.1.md
---
# ramalama login.1
# login
## Synopsis
**ramalama login** [*options*] [*registry*]

View File

@@ -1,11 +1,11 @@
---
title: ramalama logout.1
title: logout
description: logout from remote registry
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-logout.1.md
---
# ramalama logout.1
# logout
## Synopsis
**ramalama logout** [*options*] [*registry*]

View File

@@ -1,11 +1,11 @@
---
title: ramalama perplexity.1
title: perplexity
description: calculate the perplexity value of an AI Model
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-perplexity.1.md
---
# ramalama perplexity.1
# perplexity
## Synopsis
**ramalama perplexity** [*options*] *model* [arg ...]
@@ -66,7 +66,7 @@ OCI container image to run with specified AI model. RamaLama defaults to using
images based on the accelerator it discovers. For example:
`quay.io/ramalama/ramalama`. See the table below for all default images.
The default image tag is based on the minor version of the RamaLama package.
Version 0.12.1 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
Version 0.12.3 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
The default can be overridden in the ramalama.conf file or via the
RAMALAMA_IMAGE environment variable. `export RAMALAMA_IMAGE=quay.io/ramalama/aiimage:1.2` tells

View File

@@ -1,11 +1,11 @@
---
title: ramalama pull.1
title: pull
description: pull AI Models from Model registries to local storage
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-pull.1.md
---
# ramalama pull.1
# pull
## Synopsis
**ramalama pull** [*options*] *model*
@@ -24,6 +24,9 @@ Print usage message
#### **--tls-verify**=*true*
require HTTPS and verify certificates when contacting OCI registries
#### **--verify**=*true*
verify the model after pull, disable to allow pulling of models with different endianness
## See Also
[ramalama(1)](/docs/commands/ramalama/)

View File

@@ -1,11 +1,11 @@
---
title: ramalama push.1
title: push
description: push AI Models from local storage to remote registries
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-push.1.md
---
# ramalama push.1
# push
## Synopsis
**ramalama push** [*options*] *model* [*target*]
@@ -74,7 +74,7 @@ Writing manifest to image destination
```
## See Also
[ramalama(1)](/docs/commands/ramalama/), [ramalama-convert(1)](../../commands/ramalama/convert)
[ramalama(1)](/docs/commands/ramalama/), [ramalama-convert(1)](/docs/commands/ramalama/convert)
---

View File

@@ -1,11 +1,11 @@
---
title: ramalama rag.1
title: rag
description: generate and convert Retrieval Augmented Generation (RAG) data from provided documents into an OCI Image
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-rag.1.md
---
# ramalama rag.1
# rag
## Synopsis
**ramalama rag** [options] [path ...] image
@@ -56,7 +56,7 @@ OCI container image to run with specified AI model. RamaLama defaults to using
images based on the accelerator it discovers. For example:
`quay.io/ramalama/ramalama-rag`. See the table below for all default images.
The default image tag is based on the minor version of the RamaLama package.
Version 0.12.1 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
Version 0.12.3 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
The default can be overridden in the ramalama.conf file or via the
RAMALAMA_IMAGE environment variable. `export RAMALAMA_IMAGE=quay.io/ramalama/aiimage:1.2` tells

View File

@@ -1,11 +1,11 @@
---
title: ramalama.1
title: ramalama
description: Simple management tool for working with AI Models
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama.1.md
---
# ramalama.1
# ramalama
## Synopsis
**ramalama** [*options*] *command*
@@ -32,14 +32,14 @@ RamaLama pulls AI Models from model registries. Starting a chatbot or a rest API
When both Podman and Docker are installed, RamaLama defaults to Podman, The `RAMALAMA_CONTAINER_ENGINE=docker` environment variable can override this behaviour. When neither are installed RamaLama attempts to run the model with software on the local system.
:::note
On MacOS systems that use Podman for containers, configure the Podman machine to use the `libkrun` machine provider. The `libkrun` provider enables containers within the Podman Machine access to the Mac's GPU. See [ramalama-macos(7)](../../platform-guides/macos) for further information.
On MacOS systems that use Podman for containers, configure the Podman machine to use the `libkrun` machine provider. The `libkrun` provider enables containers within the Podman Machine access to the Mac's GPU. See [ramalama-macos(7)](/docs/platform-guides/macos) for further information.
:::
:::note
On systems with NVIDIA GPUs, see [ramalama-cuda(7)](../../platform-guides/cuda) to correctly configure the host system.
On systems with NVIDIA GPUs, see [ramalama-cuda(7)](/docs/platform-guides/cuda) to correctly configure the host system.
:::
RamaLama CLI defaults can be modified via ramalama.conf files. Default settings for flags are defined in [ramalama.conf(5)](../../configuration/conf).
RamaLama CLI defaults can be modified via ramalama.conf files. Default settings for flags are defined in [ramalama.conf(5)](/docs/configuration/conf).
## SECURITY
@@ -146,25 +146,25 @@ The default can be overridden in the ramalama.conf file.
| Command | Description |
| ------------------------------------------------- | ---------------------------------------------------------- |
| [ramalama-bench(1)](../../commands/ramalama/bench) |benchmark specified AI Model|
| [ramalama-chat(1)](../../commands/ramalama/chat) |OpenAI chat with the specified REST API URL|
| [ramalama-containers(1)](../../commands/ramalama/containers)|list all RamaLama containers|
| [ramalama-convert(1)](../../commands/ramalama/convert) |convert AI Models from local storage to OCI Image|
| [ramalama-daemon(1)](../../commands/ramalama/daemon) |run a RamaLama REST server|
| [ramalama-info(1)](../../commands/ramalama/info) |display RamaLama configuration information|
| [ramalama-inspect(1)](../../commands/ramalama/inspect) |inspect the specified AI Model|
| [ramalama-list(1)](../../commands/ramalama/list) |list all downloaded AI Models|
| [ramalama-login(1)](../../commands/ramalama/login) |login to remote registry|
| [ramalama-logout(1)](../../commands/ramalama/logout) |logout from remote registry|
| [ramalama-perplexity(1)](../../commands/ramalama/perplexity)|calculate the perplexity value of an AI Model|
| [ramalama-pull(1)](../../commands/ramalama/pull) |pull AI Models from Model registries to local storage|
| [ramalama-push(1)](../../commands/ramalama/push) |push AI Models from local storage to remote registries|
| [ramalama-rag(1)](../../commands/ramalama/rag) |generate and convert Retrieval Augmented Generation (RAG) data from provided documents into an OCI Image|
| [ramalama-rm(1)](../../commands/ramalama/rm) |remove AI Models from local storage|
| [ramalama-run(1)](../../commands/ramalama/run) |run specified AI Model as a chatbot|
| [ramalama-serve(1)](../../commands/ramalama/serve) |serve REST API on specified AI Model|
| [ramalama-stop(1)](../../commands/ramalama/stop) |stop named container that is running AI Model|
| [ramalama-version(1)](../../commands/ramalama/version) |display version of RamaLama|
| [ramalama-bench(1)](/docs/commands/ramalama/bench) |benchmark specified AI Model|
| [ramalama-chat(1)](/docs/commands/ramalama/chat) |OpenAI chat with the specified REST API URL|
| [ramalama-containers(1)](/docs/commands/ramalama/containers)|list all RamaLama containers|
| [ramalama-convert(1)](/docs/commands/ramalama/convert) |convert AI Models from local storage to OCI Image|
| [ramalama-daemon(1)](/docs/commands/ramalama/daemon) |run a RamaLama REST server|
| [ramalama-info(1)](/docs/commands/ramalama/info) |display RamaLama configuration information|
| [ramalama-inspect(1)](/docs/commands/ramalama/inspect) |inspect the specified AI Model|
| [ramalama-list(1)](/docs/commands/ramalama/list) |list all downloaded AI Models|
| [ramalama-login(1)](/docs/commands/ramalama/login) |login to remote registry|
| [ramalama-logout(1)](/docs/commands/ramalama/logout) |logout from remote registry|
| [ramalama-perplexity(1)](/docs/commands/ramalama/perplexity)|calculate the perplexity value of an AI Model|
| [ramalama-pull(1)](/docs/commands/ramalama/pull) |pull AI Models from Model registries to local storage|
| [ramalama-push(1)](/docs/commands/ramalama/push) |push AI Models from local storage to remote registries|
| [ramalama-rag(1)](/docs/commands/ramalama/rag) |generate and convert Retrieval Augmented Generation (RAG) data from provided documents into an OCI Image|
| [ramalama-rm(1)](/docs/commands/ramalama/rm) |remove AI Models from local storage|
| [ramalama-run(1)](/docs/commands/ramalama/run) |run specified AI Model as a chatbot|
| [ramalama-serve(1)](/docs/commands/ramalama/serve) |serve REST API on specified AI Model|
| [ramalama-stop(1)](/docs/commands/ramalama/stop) |stop named container that is running AI Model|
| [ramalama-version(1)](/docs/commands/ramalama/version) |display version of RamaLama|
## CONFIGURATION FILES
@@ -195,7 +195,7 @@ although the recommended way is to use the ramalama.conf file.
| TMPDIR | directory for temporary files. Defaults to /var/tmp if unset.|
## See Also
[podman(1)](https://github.com/containers/podman/blob/main/docs/source/markdown/podman.1.md), **docker(1)**, [ramalama.conf(5)](../../configuration/conf), [ramalama-cuda(7)](../../platform-guides/cuda), [ramalama-macos(7)](../../platform-guides/macos)
[podman(1)](https://github.com/containers/podman/blob/main/docs/source/markdown/podman.1.md), **docker(1)**, [ramalama.conf(5)](/docs/configuration/conf), [ramalama-cuda(7)](/docs/platform-guides/cuda), [ramalama-macos(7)](/docs/platform-guides/macos)
---

View File

@@ -1,11 +1,11 @@
---
title: ramalama rm.1
title: rm
description: remove AI Models from local storage
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-rm.1.md
---
# ramalama rm.1
# rm
## Synopsis
**ramalama rm** [*options*] *model* [...]

View File

@@ -1,11 +1,11 @@
---
title: ramalama run.1
title: run
description: run specified AI Model as a chatbot
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-run.1.md
---
# ramalama run.1
# run
## Synopsis
**ramalama run** [*options*] *model* [arg ...]
@@ -77,7 +77,7 @@ OCI container image to run with specified AI model. RamaLama defaults to using
images based on the accelerator it discovers. For example:
`quay.io/ramalama/ramalama`. See the table below for all default images.
The default image tag is based on the minor version of the RamaLama package.
Version 0.12.1 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
Version 0.12.3 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
The default can be overridden in the ramalama.conf file or via the
RAMALAMA_IMAGE environment variable. `export RAMALAMA_IMAGE=quay.io/ramalama/aiimage:1.2` tells
@@ -102,6 +102,11 @@ If GPU device on host system is accessible to user via group access, this option
#### **--keepalive**
duration to keep a model loaded (e.g. 5m)
#### **--mcp**=SERVER_URL
MCP (Model Context Protocol) servers to use for enhanced tool calling capabilities.
Can be specified multiple times to connect to multiple MCP servers.
Each server provides tools that can be automatically invoked during chat conversations.
#### **--name**, **-n**
name of the container to run the Model in
@@ -229,10 +234,10 @@ $ ramalama run granite
## NVIDIA CUDA Support
See [ramalama-cuda(7)](../../platform-guides/cuda) for setting up the host Linux system for CUDA support.
See [ramalama-cuda(7)](/docs/platform-guides/cuda) for setting up the host Linux system for CUDA support.
## See Also
[ramalama(1)](/docs/commands/ramalama/), [ramalama-cuda(7)](../../platform-guides/cuda)
[ramalama(1)](/docs/commands/ramalama/), [ramalama-cuda(7)](/docs/platform-guides/cuda)
---

View File

@@ -1,11 +1,11 @@
---
title: ramalama serve.1
title: serve
description: serve REST API on specified AI Model
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-serve.1.md
---
# ramalama serve.1
# serve
## Synopsis
**ramalama serve** [*options*] _model_
@@ -106,6 +106,7 @@ Generate specified configuration format for running the AI Model as a service
| quadlet | Podman supported container definition for running AI Model under systemd |
| kube | Kubernetes YAML definition for running the AI Model as a service |
| quadlet/kube | Kubernetes YAML definition for running the AI Model as a service and Podman supported container definition for running the Kube YAML specified pod under systemd|
| compose | Compose YAML definition for running the AI Model as a service |
Optionally, an output directory for the generated files can be specified by
appending the path to the type, e.g. `--generate kube:/etc/containers/systemd`.
@@ -121,7 +122,7 @@ OCI container image to run with specified AI model. RamaLama defaults to using
images based on the accelerator it discovers. For example:
`quay.io/ramalama/ramalama`. See the table above for all default images.
The default image tag is based on the minor version of the RamaLama package.
Version 0.12.1 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
Version 0.12.3 of RamaLama pulls an image with a `:0.12` tag from the quay.io/ramalama OCI repository. The --image option overrides this default.
The default can be overridden in the ramalama.conf file or via the
RAMALAMA_IMAGE environment variable. `export RAMALAMA_IMAGE=quay.io/ramalama/aiimage:1.2` tells
@@ -402,6 +403,30 @@ spec:
name: dri
```
### Generate Compose file
```bash
$ ramalama serve --name=my-smollm-server --port 1234 --generate=compose smollm:135m
Generating Compose YAML file: docker-compose.yaml
$ cat docker-compose.yaml
version: '3.8'
services:
my-smollm-server:
image: quay.io/ramalama/ramalama:latest
container_name: my-smollm-server
command: ramalama serve --host 0.0.0.0 --port 1234 smollm:135m
ports:
- "1234:1234"
volumes:
- ~/.local/share/ramalama/models/smollm-135m-instruct:/mnt/models/model.file:ro
environment:
- HOME=/tmp
cap_drop:
- ALL
security_opt:
- no-new-privileges
- label=disable
```
### Generate a Llama Stack Kubernetes YAML file named MyLamaStack
```bash
$ ramalama serve --api llama-stack --name MyLamaStack --generate=kube oci://quay.io/rhatdan/granite:latest
@@ -520,7 +545,7 @@ WantedBy=multi-user.target default.target
## NVIDIA CUDA Support
See [ramalama-cuda(7)](../../platform-guides/cuda) for setting up the host Linux system for CUDA support.
See [ramalama-cuda(7)](/docs/platform-guides/cuda) for setting up the host Linux system for CUDA support.
## MLX Support
@@ -544,7 +569,7 @@ ramalama --runtime=mlx serve hf://mlx-community/Unsloth-Phi-4-4bit
```
## See Also
[ramalama(1)](/docs/commands/ramalama/), [ramalama-stop(1)](../../commands/ramalama/stop), **quadlet(1)**, **systemctl(1)**, **podman(1)**, **podman-ps(1)**, [ramalama-cuda(7)](../../platform-guides/cuda)
[ramalama(1)](/docs/commands/ramalama/), [ramalama-stop(1)](/docs/commands/ramalama/stop), **quadlet(1)**, **systemctl(1)**, **podman(1)**, **podman-ps(1)**, [ramalama-cuda(7)](/docs/platform-guides/cuda)
---

View File

@@ -1,11 +1,11 @@
---
title: ramalama stop.1
title: stop
description: stop named container that is running AI Model
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-stop.1.md
---
# ramalama stop.1
# stop
## Synopsis
**ramalama stop** [*options*] *name*
@@ -38,7 +38,7 @@ $ ramalama stop --all
```
## See Also
[ramalama(1)](/docs/commands/ramalama/), [ramalama-run(1)](../../commands/ramalama/run), [ramalama-serve(1)](../../commands/ramalama/serve)
[ramalama(1)](/docs/commands/ramalama/), [ramalama-run(1)](/docs/commands/ramalama/run), [ramalama-serve(1)](/docs/commands/ramalama/serve)
---

View File

@@ -1,11 +1,11 @@
---
title: ramalama version.1
title: version
description: display version of RamaLama
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-version.1.md
---
# ramalama version.1
# version
## Synopsis
**ramalama version**
@@ -22,9 +22,9 @@ Print usage message
```bash
$ ramalama version
ramalama version 0.12.1
ramalama version 0.12.3
$ ramalama -q version
0.12.1
0.12.3
>
```
## See Also

View File

@@ -7,10 +7,6 @@ description: Configuration file reference
# Configuration File
# NAME
ramalama.conf - These configuration files specifies default
configuration options and command-line flags for RamaLama.
# DESCRIPTION
RamaLama reads all ramalama.conf files, if they exists
and modify the defaults for running RamaLama on the host. ramalama.conf uses
@@ -71,6 +67,10 @@ The ramalama table contains settings to configure and manage the OCI runtime.
Unified API layer for Inference, RAG, Agents, Tools, Safety, Evals, and Telemetry.
Options: llama-stack, none
**api_key**=""
OpenAI-compatible API key. Can also be set via the RAMALAMA_API_KEY environment variable.
**carimage**="registry.access.redhat.com/ubi10-micro:latest"
OCI model car image

View File

@@ -7,9 +7,6 @@ description: Configuration file reference
# OCI Spec
# NAME
ramalama-oci - RamaLama oci:// Image Format
# DESCRIPTION
RamaLamas `oci://` transport uses [OpenContainers image registries](https://github.com/opencontainers/distribution-spec) to store AI models.

View File

@@ -1,11 +1,11 @@
---
title: ramalama cann.7
title: CANN Setup
description: Platform-specific setup guide
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-cann.7.md
---
# ramalama cann.7
# CANN Setup
# Setting Up RamaLama with Ascend NPU Support on Linux systems

View File

@@ -1,11 +1,11 @@
---
title: ramalama cuda.7
title: CUDA Setup
description: Platform-specific setup guide
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-cuda.7.md
---
# ramalama cuda.7
# CUDA Setup
# Setting Up RamaLama with CUDA Support on Linux systems

View File

@@ -1,11 +1,11 @@
---
title: ramalama macos.7
title: macOS Setup
description: Platform-specific setup guide
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-macos.7.md
---
# ramalama macos.7
# macOS Setup
# Configure Podman Machine on Mac for GPU Acceleration

View File

@@ -1,11 +1,11 @@
---
title: ramalama musa.7
title: MUSA Setup
description: Platform-specific setup guide
# This file is auto-generated from manpages. Do not edit manually.
# Source: ramalama-musa.7.md
---
# ramalama musa.7
# MUSA Setup
# Setting Up RamaLama with MUSA Support on Linux systems

View File

@@ -26,7 +26,12 @@ const config: Config = {
projectName: 'ramalama', // Usually your repo name.
onBrokenLinks: 'throw',
onBrokenMarkdownLinks: 'warn',
markdown: {
hooks: {
onBrokenMarkdownLinks: 'warn',
},
},
// Even if you don't use internationalization, you can use this field to set
// useful metadata like html lang. For example, if your site is Chinese, you

2373
docsite/package-lock.json generated

File diff suppressed because it is too large Load Diff

View File

@@ -5,6 +5,7 @@
"scripts": {
"docusaurus": "docusaurus",
"start": "docusaurus start",
"dev": "docusaurus start",
"build": "docusaurus build",
"swizzle": "docusaurus swizzle",
"deploy": "docusaurus deploy",
@@ -15,20 +16,20 @@
"typecheck": "tsc"
},
"dependencies": {
"@docusaurus/core": "3.8.1",
"@docusaurus/plugin-client-redirects": "^3.8.1",
"@docusaurus/preset-classic": "3.8.1",
"@mdx-js/react": "^3.0.0",
"clsx": "^2.0.0",
"prism-react-renderer": "^2.3.0",
"react": "^19.0.0",
"react-dom": "^19.0.0"
"@docusaurus/core": "3.9.1",
"@docusaurus/plugin-client-redirects": "^3.9.1",
"@docusaurus/preset-classic": "3.9.1",
"@mdx-js/react": "^3.1.1",
"clsx": "^2.1.1",
"prism-react-renderer": "^2.4.1",
"react": "^19.1.1",
"react-dom": "^19.1.1"
},
"devDependencies": {
"@docusaurus/module-type-aliases": "3.8.1",
"@docusaurus/tsconfig": "3.8.1",
"@docusaurus/types": "3.8.1",
"typescript": "~5.9.0"
"@docusaurus/module-type-aliases": "3.9.1",
"@docusaurus/tsconfig": "3.9.1",
"@docusaurus/types": "3.9.1",
"typescript": "~5.9.3"
},
"browserslist": {
"production": [