mirror of
https://github.com/containers/ramalama.git
synced 2026-02-05 06:46:39 +01:00
Fix typo in llama.cpp engine spec
Signed-off-by: Oliver Walsh <owalsh@redhat.com>
This commit is contained in:
@@ -58,7 +58,7 @@ commands:
|
||||
- name: "--flash-attn"
|
||||
description: "Set Flash Attention use"
|
||||
value: "on"
|
||||
if: "{{ host.uses_nvidia or host.uses_metal() }}"
|
||||
if: "{{ host.uses_nvidia or host.uses_metal }}"
|
||||
- name: "-ngl"
|
||||
description: "Number of layers to offload to the GPU if available"
|
||||
value: "{{ 999 if args.ngl < 0 else args.ngl }}"
|
||||
|
||||
@@ -58,7 +58,7 @@ commands:
|
||||
- name: "--flash-attn"
|
||||
description: "Set Flash Attention use"
|
||||
value: "on"
|
||||
if: "{{ host.uses_nvidia or host.uses_metal() }}"
|
||||
if: "{{ host.uses_nvidia or host.uses_metal }}"
|
||||
- name: "-ngl"
|
||||
description: "Number of layers to offload to the GPU if available"
|
||||
value: "{{ 999 if args.ngl < 0 else args.ngl }}"
|
||||
|
||||
Reference in New Issue
Block a user