From 732100bf4d44fee0f80d619d3486e8b6e75c4f88 Mon Sep 17 00:00:00 2001 From: Ashleigh Brennan Date: Wed, 28 Jan 2026 11:49:09 -0600 Subject: [PATCH] CNV-74970: Removing callouts and fixing other misc doc issues --- .../virt-removing-interface-from-nodes.adoc | 36 +++++----- ...ving-mediated-device-from-cluster-cli.adoc | 8 +-- .../virt-restoring-vm-from-snapshot-cli.adoc | 69 ++++++++++--------- modules/virt-running-ssp-pipeline-cli.adoc | 9 +-- modules/virt-schedule-cpu-host-model-vms.adoc | 4 +- ...virt-schedule-supported-cpu-model-vms.adoc | 5 +- .../virt-setting-cpu-allocation-ratio.adoc | 5 +- modules/virt-setting-policy-attributes.adoc | 8 +-- modules/virt-subscribing-cli.adoc | 16 +++-- .../virt-troubleshoot-storage-checkup.adoc | 8 +-- 10 files changed, 85 insertions(+), 83 deletions(-) diff --git a/modules/virt-removing-interface-from-nodes.adoc b/modules/virt-removing-interface-from-nodes.adoc index 2ed66d4355..1d0c3f24a7 100644 --- a/modules/virt-removing-interface-from-nodes.adoc +++ b/modules/virt-removing-interface-from-nodes.adoc @@ -31,36 +31,32 @@ Deleting the node network policy that added an interface does not change the con apiVersion: nmstate.io/v1 kind: NodeNetworkConfigurationPolicy metadata: - name: <1> + name: spec: - nodeSelector: <2> - node-role.kubernetes.io/worker: "" <3> + nodeSelector: + node-role.kubernetes.io/worker: "" desiredState: interfaces: - name: br1 type: linux-bridge - state: absent <4> - - name: eth1 <5> - type: ethernet <6> - state: up <7> + state: absent + - name: eth1 + type: ethernet + state: up ipv4: - dhcp: true <8> - enabled: true <9> + dhcp: true + enabled: true ---- -<1> Name of the policy. -<2> Optional: If you do not include the `nodeSelector` parameter, the policy applies to all nodes in the cluster. -<3> This example uses the `node-role.kubernetes.io/worker: ""` node selector to select all worker nodes in the cluster. -<4> Changing the state to `absent` removes the interface. -<5> The name of the interface that is to be unattached from the bridge interface. -<6> The type of interface. This example creates an Ethernet networking interface. -<7> The requested state for the interface. -<8> Optional: If you do not use `dhcp`, you can either set a static IP or leave the interface without an IP address. -<9> Enables `ipv4` in this example. +** `metadata.name` defines the name of the policy. +** `spec.nodeSelector` defines the `nodeSelector` parameter. This parameter is optional. If you do not include the `nodeSelector` parameter, the policy applies to all nodes in the cluster. This example uses the `node-role.kubernetes.io/worker: ""` node selector to select all worker nodes in the cluster. +** `spec.desiredState.interfaces` defines the name, type, and desired state of an interface. This example creates both Linux bridge and Ethernet networking interfaces. Setting `state: absent` removes the interface. +** `spec.desiredState.interfaces.ipv4` defines `ipv4` settings for the interface. These settings are optional. If you do not use `dhcp`, you can either set a static IP or leave the interface without an IP address. Setting `enabled: true` enables `ipv4` in this example. . Update the policy on the node and remove the interface: + [source,terminal] ---- -$ oc apply -f <1> +$ oc apply -f ---- -<1> File name of the policy manifest. ++ +Where `` is the filename of the policy manifest. diff --git a/modules/virt-removing-mediated-device-from-cluster-cli.adoc b/modules/virt-removing-mediated-device-from-cluster-cli.adoc index 02c6a9b9d7..cd61aa4f98 100644 --- a/modules/virt-removing-mediated-device-from-cluster-cli.adoc +++ b/modules/virt-removing-mediated-device-from-cluster-cli.adoc @@ -33,14 +33,14 @@ metadata: namespace: {CNVNamespace} spec: mediatedDevicesConfiguration: - mediatedDeviceTypes: <1> + mediatedDeviceTypes: - nvidia-231 permittedHostDevices: - mediatedDevices: <2> + mediatedDevices: - mdevNameSelector: GRID T4-2Q resourceName: nvidia.com/GRID_T4-2Q ---- -<1> To remove the `nvidia-231` device type, delete it from the `mediatedDeviceTypes` array. -<2> To remove the `GRID T4-2Q` device, delete the `mdevNameSelector` field and its corresponding `resourceName` field. +** To remove the `nvidia-231` device type, delete it from the `mediatedDeviceTypes` array. +** To remove the `GRID T4-2Q` device, delete the `mdevNameSelector` field and its corresponding `resourceName` field. . Save your changes and exit the editor. diff --git a/modules/virt-restoring-vm-from-snapshot-cli.adoc b/modules/virt-restoring-vm-from-snapshot-cli.adoc index 7c2eb943be..232aeb4e8e 100644 --- a/modules/virt-restoring-vm-from-snapshot-cli.adoc +++ b/modules/virt-restoring-vm-from-snapshot-cli.adoc @@ -50,7 +50,7 @@ The snapshot controller updates the status fields of the `VirtualMachineRestore` .Verification -* Verify that the VM is restored to the previous state represented by the snapshot and that the `complete` flag is set to `true`: +* Verify that the VM is restored to the previous state represented by the snapshot and that the `status.complete` flag is set to `true`: + [source,terminal] ---- @@ -64,48 +64,49 @@ Example output: apiVersion: snapshot.kubevirt.io/v1beta1 kind: VirtualMachineRestore metadata: -creationTimestamp: "2020-09-30T14:46:27Z" -generation: 5 -name: my-vmrestore -namespace: default -ownerReferences: -- apiVersion: kubevirt.io/v1 - blockOwnerDeletion: true - controller: true - kind: VirtualMachine - name: my-vm - uid: 355897f3-73a0-4ec4-83d3-3c2df9486f4f + creationTimestamp: "2020-09-30T14:46:27Z" + generation: 5 + name: my-vmrestore + namespace: default + ownerReferences: + - apiVersion: kubevirt.io/v1 + blockOwnerDeletion: true + controller: true + kind: VirtualMachine + name: my-vm + uid: 355897f3-73a0-4ec4-83d3-3c2df9486f4f resourceVersion: "5512" - selfLink: /apis/snapshot.kubevirt.io/v1beta1/namespaces/default/virtualmachinerestores/my-vmrestore uid: 71c679a8-136e-46b0-b9b5-f57175a6a041 - spec: - target: - apiGroup: kubevirt.io - kind: VirtualMachine - name: my-vm +spec: + target: + apiGroup: kubevirt.io + kind: VirtualMachine + name: my-vm virtualMachineSnapshotName: my-vmsnapshot - status: - complete: true <1> +status: + complete: true conditions: - lastProbeTime: null - lastTransitionTime: "2020-09-30T14:46:28Z" - reason: Operation complete - status: "False" <2> - type: Progressing + lastTransitionTime: "2020-09-30T14:46:28Z" + reason: Operation complete + status: "False" + type: Progressing - lastProbeTime: null - lastTransitionTime: "2020-09-30T14:46:28Z" - reason: Operation complete - status: "True" <3> - type: Ready + lastTransitionTime: "2020-09-30T14:46:28Z" + reason: Operation complete + status: "True" + type: Ready deletedDataVolumes: - test-dv1 restoreTime: "2020-09-30T14:46:28Z" restores: - dataVolumeName: restore-71c679a8-136e-46b0-b9b5-f57175a6a041-datavolumedisk1 - persistentVolumeClaim: restore-71c679a8-136e-46b0-b9b5-f57175a6a041-datavolumedisk1 - volumeName: datavolumedisk1 - volumeSnapshotName: vmsnapshot-28eedf08-5d6a-42c1-969c-2eda58e2a78d-volume-datavolumedisk1 + persistentVolumeClaim: restore-71c679a8-136e-46b0-b9b5-f57175a6a041-datavolumedisk1 + volumeName: datavolumedisk1 + volumeSnapshotName: vmsnapshot-28eedf08-5d6a-42c1-969c-2eda58e2a78d-volume-datavolumedisk1 ---- -<1> Specifies if the process of restoring the VM to the state represented by the snapshot is complete. -<2> The `status` field of the `Progressing` condition specifies if the VM is still being restored. -<3> The `status` field of the `Ready` condition specifies if the VM restoration process is complete. ++ +[NOTE] +==== +If the `Progressing` condition has `status: "True"`, the VM is still being restored. +==== diff --git a/modules/virt-running-ssp-pipeline-cli.adoc b/modules/virt-running-ssp-pipeline-cli.adoc index acb99ab803..64b8766086 100644 --- a/modules/virt-running-ssp-pipeline-cli.adoc +++ b/modules/virt-running-ssp-pipeline-cli.adoc @@ -28,9 +28,9 @@ metadata: spec: params: - name: winImageDownloadURL - value: # <1> + value: - name: acceptEula - value: false # <2> + value: false pipelineRef: params: - name: catalog @@ -51,8 +51,9 @@ spec: fsGroup: 107 runAsUser: 107 ---- -<1> Specify the URL for the Windows 11 64-bit ISO file. The product's language must be English (United States). -<2> Example `PipelineRun` objects have a special parameter, `acceptEula`. By setting this parameter, you are agreeing to the applicable Microsoft user license agreements for each deployment or installation of the Microsoft products. If you set it to false, the pipeline exits at the first task. +** For ``, specify the URL for the Windows 11 64-bit ISO file. The product's language must be English (United States). +** Example `PipelineRun` objects have a special parameter, `acceptEula`. By setting this parameter, you are agreeing to the applicable Microsoft user license agreements for each deployment or installation of the Microsoft products. If you set it to false, the pipeline exits at the first task. + . Apply the `PipelineRun` manifest: + [source,terminal] diff --git a/modules/virt-schedule-cpu-host-model-vms.adoc b/modules/virt-schedule-cpu-host-model-vms.adoc index e70d6f9c95..a68ca7edb9 100644 --- a/modules/virt-schedule-cpu-host-model-vms.adoc +++ b/modules/virt-schedule-cpu-host-model-vms.adoc @@ -24,6 +24,6 @@ spec: spec: domain: cpu: - model: host-model <1> + model: host-model ---- -<1> The VM that inherits the CPU model of the node where it is scheduled. +** `spec.template.spec.domain.cpu.model` defines the VM that inherits the CPU model of the node where it is scheduled. diff --git a/modules/virt-schedule-supported-cpu-model-vms.adoc b/modules/virt-schedule-supported-cpu-model-vms.adoc index 35d7695221..9afa4dca11 100644 --- a/modules/virt-schedule-supported-cpu-model-vms.adoc +++ b/modules/virt-schedule-supported-cpu-model-vms.adoc @@ -24,6 +24,7 @@ spec: spec: domain: cpu: - model: Conroe <1> + model: Conroe +# ... ---- -<1> CPU model for the VM. +** `spec.template.spec.domain.cpu.model` defines the CPU model for the VM. diff --git a/modules/virt-setting-cpu-allocation-ratio.adoc b/modules/virt-setting-cpu-allocation-ratio.adoc index b1f91a94bb..be87b7cfaa 100644 --- a/modules/virt-setting-cpu-allocation-ratio.adoc +++ b/modules/virt-setting-cpu-allocation-ratio.adoc @@ -34,7 +34,8 @@ $ oc edit hyperconverged kubevirt-hyperconverged -n openshift-cnv ... spec: resourceRequirements: - vmiCPUAllocationRatio: 1 <1> + vmiCPUAllocationRatio: 1 # ... ---- -<1> When `vmiCPUAllocationRatio` is set to `1`, the maximum amount of vCPUs are requested for the pod. ++ +When `vmiCPUAllocationRatio` is set to `1`, the maximum amount of vCPUs are requested for the pod. diff --git a/modules/virt-setting-policy-attributes.adoc b/modules/virt-setting-policy-attributes.adoc index 769b312b6e..59b80bc22f 100644 --- a/modules/virt-setting-policy-attributes.adoc +++ b/modules/virt-setting-policy-attributes.adoc @@ -25,8 +25,8 @@ spec: domain: cpu: features: - - name: apic <1> - policy: require <2> + - name: apic + policy: require ---- -<1> Name of the CPU feature for the VM. -<2> Policy attribute for the VM. +** `spec.template.spec.domain.cpu.features.name` defines the name of the CPU feature for the VM. +** `spec.template.spec.domain.cpu.features.policy` defines the policy attribute for the VM. diff --git a/modules/virt-subscribing-cli.adoc b/modules/virt-subscribing-cli.adoc index 91b0b25025..572ed62968 100644 --- a/modules/virt-subscribing-cli.adoc +++ b/modules/virt-subscribing-cli.adoc @@ -52,9 +52,10 @@ spec: sourceNamespace: openshift-marketplace name: {CNVSubscriptionSpecName} startingCSV: kubevirt-hyperconverged-operator.v{HCOVersion} - channel: "stable" <1> + channel: "stable" ---- -<1> Using the `stable` channel ensures that you install the version of ++ +Using the `stable` channel ensures that you install the version of {VirtProductName} that is compatible with your {product-title} version. endif::openshift-enterprise,openshift-rosa,openshift-dedicated,openshift-rosa-hcp[] @@ -87,9 +88,10 @@ spec: sourceNamespace: openshift-marketplace name: {CNVSubscriptionSpecName} startingCSV: kubevirt-hyperconverged-operator.v{HCOVersion} - channel: "stable" <1> + channel: "stable" ---- -<1> Using the `stable` channel ensures that you install the version of ++ +Using the `stable` channel ensures that you install the version of {VirtProductName} that is compatible with your {product-title} version. endif::openshift-origin[] @@ -115,7 +117,7 @@ $ oc get csv -n {CNVNamespace} + If the CSV was created successfully, the output shows an entry that contains a `NAME` value of `kubevirt-hyperconverged-operator-*`, a `DISPLAY` value of `{VirtProductName}`, and a `PHASE` value of `Succeeded`, as shown in the following example output: + -*Example output* +Example output: + [source,terminal,subs="attributes+"] ---- @@ -130,7 +132,7 @@ kubevirt-hyperconverged-operator.v{HCOVersion} {VirtProductName} {HCOVersion $ oc get hco -n {CNVNamespace} kubevirt-hyperconverged -o json | jq .status.versions ---- + -*Example output* +Example output: + [source,terminal,subs="attributes+"] ---- @@ -147,7 +149,7 @@ $ oc get hco -n {CNVNamespace} kubevirt-hyperconverged -o json | jq .status.vers $ oc get hco kubevirt-hyperconverged -n {CNVNamespace} -o json | jq -r '.status.conditions[] | {type,status}' ---- + -*Example output* +Example output: + [source,terminal] ---- diff --git a/modules/virt-troubleshoot-storage-checkup.adoc b/modules/virt-troubleshoot-storage-checkup.adoc index d667583fef..d850f1f9d2 100644 --- a/modules/virt-troubleshoot-storage-checkup.adoc +++ b/modules/virt-troubleshoot-storage-checkup.adoc @@ -35,11 +35,11 @@ metadata: kiagnose/checkup-type: kubevirt-storage data: spec.timeout: 10m - status.succeeded: "false" # <1> - status.failureReason: "ErrNoDefaultStorageClass" # <2> + status.succeeded: "false" + status.failureReason: "ErrNoDefaultStorageClass" # ... ---- -<1> If the checkup has failed, the `status.succeeded` value is `false`. -<2> If the checkup has failed, the `status.failureReason` field contains an error message. In this example output, the `ErrNoDefaultStorageClass` error message means that no default storage class is configured. +** If the checkup has failed, the `status.succeeded` value is `false`. +** If the checkup has failed, the `status.failureReason` field contains an error message. In this example output, the `ErrNoDefaultStorageClass` error message means that no default storage class is configured. . Search the directory provided by the `must-gather` tool for logs, events, or terms related to the error in the `data.status.failureReason` field value.