1
0
mirror of https://github.com/openshift/openshift-docs.git synced 2026-02-05 12:46:18 +01:00

OSDOCS-15679: Added missing underscore to {context} attributes

This commit is contained in:
dfitzmau
2025-08-05 17:00:13 +01:00
committed by openshift-cherrypick-robot
parent 7dd16b2fed
commit 2d9ae493e9
54 changed files with 63 additions and 58 deletions

View File

@@ -19,7 +19,7 @@ include::modules/update-service-delete-service-web-console.adoc[leveloffset=+2]
// Deleting an OpenShift Update Service application by using the CLI
include::modules/update-service-delete-service-cli.adoc[leveloffset=+2]
[id="update-service-uninstall{context}"]
[id="update-service-uninstall_{context}"]
== Uninstalling the OpenShift Update Service Operator
You can uninstall the OpenShift Update Service Operator by using the {product-title} web console or CLI.

View File

@@ -2,7 +2,7 @@
//
// *scalability_and_performance/cnf-provisioning-and-installing-a-distributed-unit.adoc
[id="cnf-du-configuring-a-performance-profile-to-support-workload-partitioning.adoc_{context}"]
[id="cnf-du-configuring-a-performance-profile-to-support-workload-partitioning_{context}"]
= Configuring a performance profile to support workload partitioning

View File

@@ -2,7 +2,7 @@
//
// *scalability_and_performance/cnf-provisioning-and-deploying-a-distributed-unit.adoc
[id="cnf-du-management-pods.adoc_{context}"]
[id="cnf-du-management-pods_{context}"]
= Cluster Management pods

View File

@@ -3,7 +3,7 @@
// * rosa_cluster_admin/rosa_nodes/rosa-managing-worker-nodes.adoc
:_mod-docs-content-type: PROCEDURE
[id="configuring_machine_pool_disk_volume_cli{context}"]
[id="configuring-machine-pool-disk-volume-cli_{context}"]
= Configuring machine pool disk volume using the ROSA CLI
.Prerequisite for cluster creation

View File

@@ -4,7 +4,7 @@
:_mod-docs-content-type: PROCEDURE
ifdef::openshift-rosa[]
[id="configuring_machine_pool_disk_volume_ocm{context}"]
[id="configuring-machine-pool-disk-volume-ocm_{context}"]
= Configuring machine pool disk volume using OpenShift Cluster Manager
endif::openshift-rosa[]
.Prerequisite for cluster creation

View File

@@ -3,7 +3,7 @@
// * installing/installing_vsphere/installing-vsphere-post-installation-configuration.adoc
:_mod-docs-content-type: PROCEDURE
[id="configuring-vSphere-monitoring-configuration-completion{context}"]
[id="configuring-vSphere-monitoring-configuration-completion_{context}"]
= Verifying the configuration
The connection configuration process updates operator statuses and control plane nodes. It takes approximately an hour to complete. During the configuration process, the nodes will reboot. Previously bound `PersistentVolumeClaims` objects might become disconnected.

View File

@@ -3,7 +3,7 @@
// * rosa_cluster_admin/rosa_nodes/rosa-managing-worker-nodes.adoc
:_mod-docs-content-type: PROCEDURE
[id="deleting-machine-pools-cli{context}"]
[id="deleting-machine-pools-cli_{context}"]
= Deleting a machine pool using the ROSA CLI
You can delete a machine pool for your {product-title} cluster by using the ROSA CLI.

View File

@@ -5,7 +5,7 @@
// * osd_cluster_admin/osd_nodes/osd-managing-worker-nodes.adoc
:_mod-docs-content-type: PROCEDURE
[id="deleting-machine-pools-ocm{context}"]
[id="deleting-machine-pools-ocm_{context}"]
ifndef::openshift-rosa,openshift-rosa-hcp[]
= Deleting a machine pool
endif::openshift-rosa,openshift-rosa-hcp[]

View File

@@ -3,7 +3,7 @@
// * rosa_cluster_admin/rosa_nodes/rosa-managing-worker-nodes.adoc
:_mod-docs-content-type: PROCEDURE
[id="deleting-machine-pools{context}"]
[id="deleting-machine-pools_{context}"]
= Deleting a machine pool
You can delete a machine pool in the event that your workload requirements have changed and your current machine pools no longer meet your needs.

View File

@@ -21,7 +21,7 @@ Issue:: The Self Node Remediation Operator is installed but the daemon set is no
Resolution:: Check the Operator logs for errors or warnings.
[id="unsuccessful_remediation{context}"]
[id="unsuccessful-remediation_{context}"]
== Unsuccessful remediation
Issue:: An unhealthy node was not remediated.

View File

@@ -3,7 +3,7 @@
// * gitops/configuring_argo_cd_to_recursively_sync_a_git_repository_with_your_application/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.adoc
:_mod-docs-content-type: PROCEDURE
[id="default-permissions-of-an-argocd-instance.adoc{context}"]
[id="default-permissions-of-an-argocd-instance_{context}"]
= Default permissions of an Argocd instance

View File

@@ -3,7 +3,7 @@
// * gitops/configuring_argo_cd_to_recursively_sync_a_git_repository_with_your_application/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.adoc
:_mod-docs-content-type: PROCEDURE
[id="using-argo-cd-instance-to-manage-cluster-scoped-resources{context}"]
[id="using-argo-cd-instance-to-manage-cluster-scoped-resources_{context}"]
= Using an Argo CD instance to manage cluster-scoped resources

View File

@@ -3,7 +3,7 @@
// * applications/working_with_helm_charts/configuring-custom-helm-chart-repositories.adoc
:_mod-docs-content-type: PROCEDURE
[id="adding-namespace-scoped-helm-chart-repositories.adoc_{context}"]
[id="adding-namespace-scoped-helm-chart-repositories_{context}"]
= Adding namespace-scoped custom Helm chart repositories
[role="_abstract"]

View File

@@ -1,7 +1,8 @@
//module included in logging-5-8-release-notes.adoc
:content-type: REFERENCE
:_mod-docs-content-type: REFERENCE
[id="logging-release-notes-5-8-0_{context}"]
= Logging 5.8.0
This release includes link:https://access.redhat.com/errata/RHBA-2023:6139[OpenShift Logging Bug Fix Release 5.8.0] and link:https://access.redhat.com/errata/RHBA-2023:6134[OpenShift Logging Bug Fix Release 5.8.0 Kibana].
[id="logging-release-notes-5-8-0-deprecation-notice"]

View File

@@ -1,7 +1,8 @@
//module included in logging-5-8-release-notes.adoc
:content-type: REFERENCE
:_mod-docs-content-type: REFERENCE
[id="logging-release-notes-5-8-1_{context}"]
= Logging 5.8.1
This release includes link:https://access.redhat.com/errata/RHSA-2023:7720[OpenShift Logging Bug Fix Release 5.8.1] and link:https://access.redhat.com/errata/RHBA-2023:7717[OpenShift Logging Bug Fix Release 5.8.1 Kibana].
[id="logging-release-notes-5-8-1-enhancements"]

View File

@@ -1,7 +1,8 @@
//module included in logging-5-8-release-notes.adoc
:content-type: REFERENCE
:_mod-docs-content-type: REFERENCE
[id="logging-release-notes-5-8-2"]
= Logging 5.8.2
This release includes link:https://access.redhat.com/errata/RHSA-2024:0271[OpenShift Logging Bug Fix Release 5.8.2].
[id="logging-release-notes-5-8-2-bug-fixes"]

View File

@@ -1,7 +1,8 @@
//module included in logging-5-9-release-notes.adoc
:content-type: REFERENCE
:_mod-docs-content-type: REFERENCE
[id="logging-release-notes-5-9-0_{context}"]
= Logging 5.9.0
This release includes link:https://access.redhat.com/errata/RHBA-2024:1591[OpenShift Logging Bug Fix Release 5.9.0]
[id="logging-release-notes-5-9-0-removal-notice"]

View File

@@ -1,5 +1,5 @@
//module included in logging-5-9-release-notes.adoc
:content-type: REFERENCE
:_mod-docs-content-type: REFERENCE
[id="logging-release-notes-5-9-1_{context}"]
= Logging 5.9.1
This release includes link:https://access.redhat.com/errata/RHSA-2024:2096[OpenShift Logging Bug Fix Release 5.9.1]

View File

@@ -1,7 +1,8 @@
//module included in logging-5-7-release-notes.adoc
:content-type: REFERENCE
[id="logging-release-notes-5-7-0{context}"]
:_mod-docs-content-type: REFERENCE
[id="logging-release-notes-5-7-0_{context}"]
= Logging 5.7.0
This release includes link:https://access.redhat.com/errata/RHBA-2023:2133[OpenShift Logging Bug Fix Release 5.7.0].
[id="logging-5-7-enhancements"]

View File

@@ -15,7 +15,7 @@ Typically this means that the device does not have an attached network interface
Configuring networking settings for disconnected environments is necessary because {microshift-short} requires a network device to support cluster communication. To meet this requirement, you must configure {microshift-short} networking settings to use the "fake" IP address you assign to the system loopback device during setup.
====
[id="microshift-disconnected-host-procedure-summary{context}"]
[id="microshift-disconnected-host-procedure-summary_{context}"]
== Procedure summary
To run {microshift-short} on a disconnected host, the following steps are required:

View File

@@ -3,7 +3,7 @@
// * observability/monitoring/shiftstack-prometheus-configuration.adoc
:_mod-docs-content-type: CONCEPT
[id="monitoring-shiftstack-metrics.adoc_{context}"]
[id="monitoring-shiftstack-metrics_{context}"]
= Available metrics for clusters that run on RHOSO
To query metrics and identifying resources across the stack, there are helper metrics that establish a correlation between {rhoso-first} infrastructure resources and their representations in the tenant {product-title} cluster.

View File

@@ -9,7 +9,7 @@
You can customize and export the view by using *Show advanced options*.
You can set the row size by using the *Display options* drop-down menu. The default value is *Normal*.
[id="network-observability-cao-managing-columns-trafficflow{context}"]
[id="network-observability-cao-managing-columns-trafficflow_{context}"]
== Managing columns
You can select the required columns to be displayed, and reorder them. To manage columns, click *Manage columns*.

View File

@@ -2,7 +2,7 @@
// * networking/network_observability/installing-operators.adoc
:_mod-docs-content-type: CONCEPT
[id="network-observability-lokistack-configuring-ingestion{context}"]
[id="network-observability-lokistack-configuring-ingestion_{context}"]
= LokiStack ingestion limits and health alerts
The LokiStack instance comes with default settings according to the configured size. It is possible to override some of these settings, such as the ingestion and query limits. An automatic alert in the web console notifies you when these limits are reached.

View File

@@ -3,7 +3,7 @@
// network_observability/network-observability-scheduling-resources.adoc
:_mod-docs-content-type: CONCEPT
[id="network-observability-multi-tenancy{context}"]
[id="network-observability-multi-tenancy_{context}"]
= Network observability deployment in specific nodes
You can configure the `FlowCollector` to control the deployment of network observability components in specific nodes. The `spec.agent.ebpf.advanced.scheduling`, `spec.processor.advanced.scheduling`, and `spec.consolePlugin.advanced.scheduling` specifications have the following configurable settings:

View File

@@ -3,7 +3,7 @@
// network_observability/observing-network-traffic.adoc
:_mod-docs-content-type: REFERENCE
[id="network-observability-quickfilter{context}"]
[id="network-observability-quickfilter_{context}"]
= Filtering the network traffic
By default, the Network Traffic page displays the traffic flow data in the cluster based on the default filters configured in the `FlowCollector` instance. You can use the filter options to observe the required data by changing the preset filter.

View File

@@ -27,7 +27,7 @@ You can edit the `FlowCollector` to view information about network traffic event
. Select *cluster*, and then select the *YAML* tab.
. Configure the `FlowCollector` CR to enable viewing `NetworkEvents`, for example:
+
[id="network-observability-flowcollector-configuring-networkevents{context}"]
[id="network-observability-flowcollector-configuring-networkevents_{context}"]
.Example `FlowCollector` configuration
[source, yaml]
----

View File

@@ -3,7 +3,7 @@
// network_observability/observing-network-traffic.adoc
:_mod-docs-content-type: PROCEDURE
[id="network-observability-zones{context}"]
[id="network-observability-zones_{context}"]
= Working with availability zones
You can configure the `FlowCollector` to collect information about the cluster availability zones. This allows you to enrich network flow data with the link:https://kubernetes.io/docs/reference/labels-annotations-taints/#topologykubernetesiozone[`topology.kubernetes.io/zone`] label value applied to the nodes.

View File

@@ -3,7 +3,7 @@
// * backup_and_restore/application_backup_and_restore/installing/overriding-kopia-algorithms.adoc
:_mod-docs-content-type: PROCEDURE
[id="oadp-usecase-kopia-override-algorithms.adoc_{context}"]
[id="oadp-usecase-kopia-override-algorithms_{context}"]
= Use case for overriding Kopia hashing, encryption, and splitter algorithms
[role="_abstract"]

View File

@@ -3,7 +3,7 @@
// *cicd/pipelines/using-tekton-hub-with-openshift-pipelines.adoc
:_mod-docs-content-type: PROCEDURE
[id="disabling-tekton-hub-authorization-after-upgrade.adoc_{context}"]
[id="disabling-tekton-hub-authorization-after-upgrade_{context}"]
= Disabling {tekton-hub} authorization after upgrading the {pipelines-title} Operator from 1.7 to 1.8
[role="_abstract"]

View File

@@ -3,7 +3,7 @@
// *cicd/pipelines/using-tekton-hub-with-openshift-pipelines.adoc
:_mod-docs-content-type: PROCEDURE
[id="installing-tekton-hub-without-login-and-rating.adoc_{context}"]
[id="installing-tekton-hub-without-login-and-rating_{context}"]
= Installing {tekton-hub} without login and rating
[role="_abstract"]

View File

@@ -3,7 +3,7 @@
// *openshift_pipelines/working-with-pipelines-web-console.adoc
:_mod-docs-content-type: CONCEPT
[id="op-interacting-pipelines_from_topology_view{context}"]
[id="op-interacting-pipelines_from_topology_view_{context}"]
= Interacting with pipelines from Topology view
The side panel of the application node in the *Topology* page displays the status of a pipeline run and you can interact with it.

View File

@@ -3,7 +3,7 @@
// *openshift_pipelines/working-with-pipelines-web-console.adoc
:_mod-docs-content-type: PROCEDURE
[id="op-starting-pipelines_from_topology_view{context}"]
[id="op-starting-pipelines-from-topology-view_{context}"]
= Starting pipelines from Topology view
For pipelines created using the *From Git* option, you can use the *Topology* view to interact with pipelines after you start them:

View File

@@ -2,7 +2,7 @@
//
// * rosa_architecture/rosa_architecture_sub/rosa-architecture-models.adoc
[id="osd-aws-privatelink-architecture.adoc_{context}"]
[id="osd-aws-privatelink-architecture_{context}"]
= AWS PrivateLink architecture
The Red{nbsp}Hat managed infrastructure that creates AWS PrivateLink clusters is hosted on private subnets. The connection between Red{nbsp}Hat and the customer-provided infrastructure is created through AWS PrivateLink VPC endpoints.
@@ -18,7 +18,7 @@ The following diagram shows network connectivity of a PrivateLink cluster.
image::156_OpenShift_ROSA_Arch_1221_privatelink.png[Multi-AZ AWS PrivateLink cluster deployed on private subnets]
[id="osd-aws-reference-architecture.adoc_{context}"]
[id="osd-aws-reference-architecture_{context}"]
== AWS reference architectures
AWS provides multiple reference architectures that can be useful to customers when planning how to set up a configuration that uses AWS PrivateLink. Here are three examples:

View File

@@ -3,7 +3,7 @@
// * storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
//
:content-type: PROCEDURE
:_mod-docs-content-type: PROCEDURE
[id="persistent-storage-csi-vsphere-encryption-datastore-url_{context}"]
= Using datastore URL

View File

@@ -3,7 +3,7 @@
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
//
:content-type: PROCEDURE
:_mod-docs-content-type: PROCEDURE
[id="persistent-storage-csi-vsphere-encryption-tag-based_{context}"]
= Using tag-based placement

View File

@@ -3,7 +3,7 @@
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
//
:content-type: CONCEPT
:_mod-docs-content-type: CONCEPT
[id="persistent-storage-csi-vsphere-limitations_{context}"]
= vSphere CSI limitations

View File

@@ -3,7 +3,7 @@
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
//
:content-type: PROCEDURE
:_mod-docs-content-type: PROCEDURE
[id="persistent-storage-csi-vsphere-top-aware-during-install_{context}"]
= Creating vSphere storage topology during installation

View File

@@ -3,7 +3,7 @@
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
//
:content-type: PROCEDURE
:_mod-docs-content-type: PROCEDURE
[id="persistent-storage-csi-vsphere-top-aware-infra-top_{context}"]
= Creating vSphere storage topology without an infra topology

View File

@@ -3,7 +3,7 @@
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
//
:content-type: CONCEPT
:_mod-docs-content-type: CONCEPT
[id="persistent-storage-csi-vsphere-top-aware-overview_{context}"]
= vSphere CSI topology overview

View File

@@ -3,7 +3,7 @@
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
//
:content-type: PROCEDURE
:_mod-docs-content-type: PROCEDURE
[id="persistent-storage-csi-vsphere-top-aware-post-install_{context}"]
= Creating vSphere storage topology postinstallation

View File

@@ -3,7 +3,7 @@
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
//
:content-type: PROCEDURE
:_mod-docs-content-type: PROCEDURE
[id="persistent-storage-csi-vsphere-top-aware-results_{context}"]
= Results

View File

@@ -3,7 +3,7 @@
// * rosa_cluster_admin/rosa_nodes/rosa-managing-worker-nodes.adoc
:_mod-docs-content-type: PROCEDURE
[id="rosa-adding-tags-cli{context}"]
[id="rosa-adding-tags-cli_{context}"]
= Adding tags to a machine pool using the ROSA CLI
You can add tags to a machine pool for your {product-title} cluster by using the ROSA command-line interface (CLI).

View File

@@ -5,7 +5,7 @@
// * osd_cluster_admin/osd_nodes/osd-managing-worker-nodes.adoc
:_mod-docs-content-type: PROCEDURE
[id="rosa-adding-taints-cli{context}"]
[id="rosa-adding-taints-cli_{context}"]
= Adding taints to a machine pool using the ROSA CLI
You can add taints to a machine pool for your {product-title} cluster by using the ROSA CLI.

View File

@@ -5,7 +5,7 @@
// * osd_cluster_admin/osd_nodes/osd-managing-worker-nodes.adoc
:_mod-docs-content-type: PROCEDURE
[id="rosa-adding-taints-ocm{context}"]
[id="rosa-adding-taints-ocm_{context}"]
= Adding taints to a machine pool using {cluster-manager}
You can add taints to a machine pool for your {product-title} cluster by using {cluster-manager-first}.

View File

@@ -3,7 +3,7 @@
// * rosa_install_access_delete_clusters/rosa-sts-creating-a-cluster-with-customizations.adoc
:_mod-docs-content-type: CONCEPT
[id="rosa-imds{context}"]
[id="rosa-imds_{context}"]
= Instance Metadata Service
There are two types of ways to access instance metadata from a running instance:

View File

@@ -3,7 +3,7 @@
// * rosa_release_notes/rosa-release-notes.adoc
:_mod-docs-content-type: PROCEDURE
[id="updating_rosa_cli{context}"]
[id="updating-rosa-cli_{context}"]
== Updating the ROSA CLI tool
To use the latest version of the {product-title} (ROSA) CLI, `rosa`, download the ROSA CLI (`rosa`) from the Hybrid Cloud Console. If you already have this tool, the procedure is the same for updates.

View File

@@ -38,7 +38,7 @@ Custom domains and subdomains are not available for the platform service routes,
* Optional: Secondary router/ingress load balancer that is a secondary application load balancer, denoted by `apps2` in the URL. The secondary load balancer can be configured in {cluster-manager} to be either publicly accessible over the internet, or only privately accessible over a pre-existing private connection. If a 'Label match' is configured for this router load balancer, then only application routes matching this label will be exposed on this router load balancer, otherwise all application routes are also exposed on this router load balancer.
* Optional: Load balancers for services that can be mapped to a service running on {product-title} to enable advanced ingress features, such as non-HTTP/SNI traffic or the use of non-standard ports. These can be purchased in groups of 4 for non-CCS clusters, or they can be provisioned through the cloud provider console in Customer Cloud Subscription (CCS) clusters; however, each AWS account has a quota that link:https://docs.aws.amazon.com/elasticloadbalancing/latest/classic/elb-limits.html[limits the number of Classic Load Balancers] that can be used within each cluster.
[id="network-usage{context}"]
[id="network-usage_{context}"]
== Network usage
For non-CCS {product-title} clusters, network usage is measured based on data transfer between inbound, VPC peering, VPN, and AZ traffic. On a non-CCS {product-title} base cluster, 12 TB of network I/O is provided. Additional network I/O can be purchased in 12 TB increments. For CCS {product-title} clusters, network usage is not monitored, and is billed directly by the cloud provider.

View File

@@ -84,7 +84,7 @@ All activities performed by SREs arrive from Red Hat IP addresses and are logged
For a full list of permissions, see `sts_support_permission_policy.json` in the link:https://docs.openshift.com/rosa/rosa_architecture/rosa-sts-about-iam-resources.html[About IAM resources] user guide.
[id="rosa-sre-access-privatelink-vpc.adoc_{context}"]
[id="rosa-sre-access-privatelink-vpc_{context}"]
== SRE access through PrivateLink VPC endpoint service
PrivateLink VPC endpoint service is created as part of the ROSA cluster creation.

View File

@@ -3,7 +3,7 @@
// * updating/updating_a_cluster/updating-cluster-cli.adoc
:_mod-docs-content-type: PROCEDURE
[id="update-conditional-upgrade-path{context}"]
[id="update-conditional-upgrade-path_{context}"]
= Updating along a conditional update path
You can update along a recommended conditional update path using the web console or the OpenShift CLI (`oc`).

View File

@@ -3,7 +3,7 @@
// * virt/support/virt-troubleshooting.adoc
:_mod-docs-content-type: CONCEPT
[id="virt-about-dv-conditions-and-events.adoc_{context}"]
[id="virt-about-dv-conditions-and-events_{context}"]
= About data volume conditions and events
You can diagnose data volume issues by examining the output of the `Conditions` and `Events` sections

View File

@@ -3,7 +3,7 @@
// * virt/virtual_machines/virt-deprecated-tasks.adoc
:_mod-docs-content-type: PROCEDURE
[id="virt-deprecated-tasks.web.adoc_{context}"]
[id="virt-deprecated-tasks.web_{context}"]
= Removing deprecated or unused resources
You can clean up deprecated or unused resources associated with the {pipelines-title} Operator.

View File

@@ -402,7 +402,7 @@ With the duplicated flows update, the *Interface* and *Interface Direction* fiel
=====
For more information, see xref:../../observability/network_observability/network-observability-operator-monitoring.adoc#network-observability-netobserv-dashboard-ebpf-agent-alerts_network_observability[Using the eBPF agent alert]
and xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-quickfilternw-observe-network-traffic[Quick filters].
and xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-quickfilter_nw-observe-network-traffic[Quick filters].
[id="network-observability-ebpf-collection-filtering-1.6_{context}"]
@@ -464,7 +464,7 @@ You can create Prometheus alerts for the *Netobserv* dashboard using DNS, Packet
[id="network-observability-zones"]
==== Availability zones
You can configure the `FlowCollector` resource to collect information about the cluster availability zones. This configuration enriches the network flow data with the link:https://kubernetes.io/docs/reference/labels-annotations-taints/#topologykubernetesiozone[`topology.kubernetes.io/zone`] label value applied to the nodes. For more information, see xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-zonesnw-observe-network-traffic[Working with availability zones].
You can configure the `FlowCollector` resource to collect information about the cluster availability zones. This configuration enriches the network flow data with the link:https://kubernetes.io/docs/reference/labels-annotations-taints/#topologykubernetesiozone[`topology.kubernetes.io/zone`] label value applied to the nodes. For more information, see xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-zones_nw-observe-network-traffic[Working with availability zones].
[id="network-observability-enhanced-configuration-and-ui-1.5"]
==== Notable enhancements
@@ -583,7 +583,7 @@ The 1.4 release of the Network Observability Operator adds improvements and new
** The *NetObserv / Health* dashboard shows flows overhead as well as top flow rates per nodes, namespaces, and workloads.
** Infrastructure and Application metrics are shown in a split-view for namespaces and workloads.
For more information, see xref:../../observability/network_observability/network-observability-overview.adoc#network-observability-dashboards[Network Observability metrics] and xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-quickfilternw-observe-network-traffic[Quick filters].
For more information, see xref:../../observability/network_observability/network-observability-overview.adoc#network-observability-dashboards[Network Observability metrics] and xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-quickfilter_nw-observe-network-traffic[Quick filters].
[discrete]
[id="configuration-enhancements-1.4_{context}"]

View File

@@ -56,5 +56,5 @@ include::modules/understanding-update-channels.adoc[leveloffset=+1]
[role="_additional-resources"]
.Additional resources
* xref:../../updating/updating_a_cluster/updating-cluster-cli.adoc#update-conditional-upgrade-pathupdating-cluster-cli[Updating along a conditional upgrade path]
* xref:../../updating/updating_a_cluster/updating-cluster-cli.adoc#update-conditional-upgrade-path_updating-cluster-cli[Updating along a conditional upgrade path]
* xref:../../updating/understanding_updates/understanding-update-channels-release.adoc#fast-stable-channel-strategies_understanding-update-channels-releases[Choosing the correct channel for your cluster]

View File

@@ -69,7 +69,7 @@ include::modules/update-upgrading-oc-adm-upgrade-status.adoc[leveloffset=+1]
ifndef::openshift-origin[]
* xref:../../updating/updating_a_cluster/control-plane-only-update.adoc#control-plane-only-update[Performing a Control Plane Only update]
endif::openshift-origin[]
* xref:../../updating/updating_a_cluster/updating-cluster-cli.adoc#update-conditional-upgrade-pathupdating-cluster-cli[Updating along a conditional update path]
* xref:../../updating/updating_a_cluster/updating-cluster-cli.adoc#update-conditional-upgrade-path_updating-cluster-cli[Updating along a conditional update path]
ifndef::openshift-origin[]
* xref:../../updating/understanding_updates/understanding-update-channels-release.adoc#understanding-update-channels-releases[Understanding update channels and releases]
endif::openshift-origin[]