mirror of
https://github.com/openshift/openshift-docs.git
synced 2026-02-05 12:46:18 +01:00
OSDOCS-15679: Added missing underscore to {context} attributes
This commit is contained in:
committed by
openshift-cherrypick-robot
parent
7dd16b2fed
commit
2d9ae493e9
@@ -19,7 +19,7 @@ include::modules/update-service-delete-service-web-console.adoc[leveloffset=+2]
|
||||
// Deleting an OpenShift Update Service application by using the CLI
|
||||
include::modules/update-service-delete-service-cli.adoc[leveloffset=+2]
|
||||
|
||||
[id="update-service-uninstall{context}"]
|
||||
[id="update-service-uninstall_{context}"]
|
||||
== Uninstalling the OpenShift Update Service Operator
|
||||
|
||||
You can uninstall the OpenShift Update Service Operator by using the {product-title} web console or CLI.
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
//
|
||||
// *scalability_and_performance/cnf-provisioning-and-installing-a-distributed-unit.adoc
|
||||
|
||||
[id="cnf-du-configuring-a-performance-profile-to-support-workload-partitioning.adoc_{context}"]
|
||||
[id="cnf-du-configuring-a-performance-profile-to-support-workload-partitioning_{context}"]
|
||||
|
||||
= Configuring a performance profile to support workload partitioning
|
||||
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
//
|
||||
// *scalability_and_performance/cnf-provisioning-and-deploying-a-distributed-unit.adoc
|
||||
|
||||
[id="cnf-du-management-pods.adoc_{context}"]
|
||||
[id="cnf-du-management-pods_{context}"]
|
||||
|
||||
= Cluster Management pods
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * rosa_cluster_admin/rosa_nodes/rosa-managing-worker-nodes.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="configuring_machine_pool_disk_volume_cli{context}"]
|
||||
[id="configuring-machine-pool-disk-volume-cli_{context}"]
|
||||
= Configuring machine pool disk volume using the ROSA CLI
|
||||
|
||||
.Prerequisite for cluster creation
|
||||
|
||||
@@ -4,7 +4,7 @@
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
ifdef::openshift-rosa[]
|
||||
[id="configuring_machine_pool_disk_volume_ocm{context}"]
|
||||
[id="configuring-machine-pool-disk-volume-ocm_{context}"]
|
||||
= Configuring machine pool disk volume using OpenShift Cluster Manager
|
||||
endif::openshift-rosa[]
|
||||
.Prerequisite for cluster creation
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * installing/installing_vsphere/installing-vsphere-post-installation-configuration.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="configuring-vSphere-monitoring-configuration-completion{context}"]
|
||||
[id="configuring-vSphere-monitoring-configuration-completion_{context}"]
|
||||
= Verifying the configuration
|
||||
|
||||
The connection configuration process updates operator statuses and control plane nodes. It takes approximately an hour to complete. During the configuration process, the nodes will reboot. Previously bound `PersistentVolumeClaims` objects might become disconnected.
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * rosa_cluster_admin/rosa_nodes/rosa-managing-worker-nodes.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="deleting-machine-pools-cli{context}"]
|
||||
[id="deleting-machine-pools-cli_{context}"]
|
||||
= Deleting a machine pool using the ROSA CLI
|
||||
You can delete a machine pool for your {product-title} cluster by using the ROSA CLI.
|
||||
|
||||
|
||||
@@ -5,7 +5,7 @@
|
||||
// * osd_cluster_admin/osd_nodes/osd-managing-worker-nodes.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="deleting-machine-pools-ocm{context}"]
|
||||
[id="deleting-machine-pools-ocm_{context}"]
|
||||
ifndef::openshift-rosa,openshift-rosa-hcp[]
|
||||
= Deleting a machine pool
|
||||
endif::openshift-rosa,openshift-rosa-hcp[]
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * rosa_cluster_admin/rosa_nodes/rosa-managing-worker-nodes.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="deleting-machine-pools{context}"]
|
||||
[id="deleting-machine-pools_{context}"]
|
||||
= Deleting a machine pool
|
||||
|
||||
You can delete a machine pool in the event that your workload requirements have changed and your current machine pools no longer meet your needs.
|
||||
|
||||
@@ -21,7 +21,7 @@ Issue:: The Self Node Remediation Operator is installed but the daemon set is no
|
||||
|
||||
Resolution:: Check the Operator logs for errors or warnings.
|
||||
|
||||
[id="unsuccessful_remediation{context}"]
|
||||
[id="unsuccessful-remediation_{context}"]
|
||||
== Unsuccessful remediation
|
||||
Issue:: An unhealthy node was not remediated.
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * gitops/configuring_argo_cd_to_recursively_sync_a_git_repository_with_your_application/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="default-permissions-of-an-argocd-instance.adoc{context}"]
|
||||
[id="default-permissions-of-an-argocd-instance_{context}"]
|
||||
|
||||
= Default permissions of an Argocd instance
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * gitops/configuring_argo_cd_to_recursively_sync_a_git_repository_with_your_application/configuring-an-openshift-cluster-by-deploying-an-application-with-cluster-configurations.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="using-argo-cd-instance-to-manage-cluster-scoped-resources{context}"]
|
||||
[id="using-argo-cd-instance-to-manage-cluster-scoped-resources_{context}"]
|
||||
|
||||
= Using an Argo CD instance to manage cluster-scoped resources
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * applications/working_with_helm_charts/configuring-custom-helm-chart-repositories.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="adding-namespace-scoped-helm-chart-repositories.adoc_{context}"]
|
||||
[id="adding-namespace-scoped-helm-chart-repositories_{context}"]
|
||||
= Adding namespace-scoped custom Helm chart repositories
|
||||
|
||||
[role="_abstract"]
|
||||
|
||||
@@ -1,7 +1,8 @@
|
||||
//module included in logging-5-8-release-notes.adoc
|
||||
:content-type: REFERENCE
|
||||
:_mod-docs-content-type: REFERENCE
|
||||
[id="logging-release-notes-5-8-0_{context}"]
|
||||
= Logging 5.8.0
|
||||
|
||||
This release includes link:https://access.redhat.com/errata/RHBA-2023:6139[OpenShift Logging Bug Fix Release 5.8.0] and link:https://access.redhat.com/errata/RHBA-2023:6134[OpenShift Logging Bug Fix Release 5.8.0 Kibana].
|
||||
|
||||
[id="logging-release-notes-5-8-0-deprecation-notice"]
|
||||
|
||||
@@ -1,7 +1,8 @@
|
||||
//module included in logging-5-8-release-notes.adoc
|
||||
:content-type: REFERENCE
|
||||
:_mod-docs-content-type: REFERENCE
|
||||
[id="logging-release-notes-5-8-1_{context}"]
|
||||
= Logging 5.8.1
|
||||
|
||||
This release includes link:https://access.redhat.com/errata/RHSA-2023:7720[OpenShift Logging Bug Fix Release 5.8.1] and link:https://access.redhat.com/errata/RHBA-2023:7717[OpenShift Logging Bug Fix Release 5.8.1 Kibana].
|
||||
|
||||
[id="logging-release-notes-5-8-1-enhancements"]
|
||||
|
||||
@@ -1,7 +1,8 @@
|
||||
//module included in logging-5-8-release-notes.adoc
|
||||
:content-type: REFERENCE
|
||||
:_mod-docs-content-type: REFERENCE
|
||||
[id="logging-release-notes-5-8-2"]
|
||||
= Logging 5.8.2
|
||||
|
||||
This release includes link:https://access.redhat.com/errata/RHSA-2024:0271[OpenShift Logging Bug Fix Release 5.8.2].
|
||||
|
||||
[id="logging-release-notes-5-8-2-bug-fixes"]
|
||||
|
||||
@@ -1,7 +1,8 @@
|
||||
//module included in logging-5-9-release-notes.adoc
|
||||
:content-type: REFERENCE
|
||||
:_mod-docs-content-type: REFERENCE
|
||||
[id="logging-release-notes-5-9-0_{context}"]
|
||||
= Logging 5.9.0
|
||||
|
||||
This release includes link:https://access.redhat.com/errata/RHBA-2024:1591[OpenShift Logging Bug Fix Release 5.9.0]
|
||||
|
||||
[id="logging-release-notes-5-9-0-removal-notice"]
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
//module included in logging-5-9-release-notes.adoc
|
||||
:content-type: REFERENCE
|
||||
:_mod-docs-content-type: REFERENCE
|
||||
[id="logging-release-notes-5-9-1_{context}"]
|
||||
= Logging 5.9.1
|
||||
This release includes link:https://access.redhat.com/errata/RHSA-2024:2096[OpenShift Logging Bug Fix Release 5.9.1]
|
||||
|
||||
@@ -1,7 +1,8 @@
|
||||
//module included in logging-5-7-release-notes.adoc
|
||||
:content-type: REFERENCE
|
||||
[id="logging-release-notes-5-7-0{context}"]
|
||||
:_mod-docs-content-type: REFERENCE
|
||||
[id="logging-release-notes-5-7-0_{context}"]
|
||||
= Logging 5.7.0
|
||||
|
||||
This release includes link:https://access.redhat.com/errata/RHBA-2023:2133[OpenShift Logging Bug Fix Release 5.7.0].
|
||||
|
||||
[id="logging-5-7-enhancements"]
|
||||
|
||||
@@ -15,7 +15,7 @@ Typically this means that the device does not have an attached network interface
|
||||
Configuring networking settings for disconnected environments is necessary because {microshift-short} requires a network device to support cluster communication. To meet this requirement, you must configure {microshift-short} networking settings to use the "fake" IP address you assign to the system loopback device during setup.
|
||||
====
|
||||
|
||||
[id="microshift-disconnected-host-procedure-summary{context}"]
|
||||
[id="microshift-disconnected-host-procedure-summary_{context}"]
|
||||
== Procedure summary
|
||||
|
||||
To run {microshift-short} on a disconnected host, the following steps are required:
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * observability/monitoring/shiftstack-prometheus-configuration.adoc
|
||||
|
||||
:_mod-docs-content-type: CONCEPT
|
||||
[id="monitoring-shiftstack-metrics.adoc_{context}"]
|
||||
[id="monitoring-shiftstack-metrics_{context}"]
|
||||
= Available metrics for clusters that run on RHOSO
|
||||
|
||||
To query metrics and identifying resources across the stack, there are helper metrics that establish a correlation between {rhoso-first} infrastructure resources and their representations in the tenant {product-title} cluster.
|
||||
|
||||
@@ -9,7 +9,7 @@
|
||||
You can customize and export the view by using *Show advanced options*.
|
||||
You can set the row size by using the *Display options* drop-down menu. The default value is *Normal*.
|
||||
|
||||
[id="network-observability-cao-managing-columns-trafficflow{context}"]
|
||||
[id="network-observability-cao-managing-columns-trafficflow_{context}"]
|
||||
== Managing columns
|
||||
You can select the required columns to be displayed, and reorder them. To manage columns, click *Manage columns*.
|
||||
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
|
||||
// * networking/network_observability/installing-operators.adoc
|
||||
:_mod-docs-content-type: CONCEPT
|
||||
[id="network-observability-lokistack-configuring-ingestion{context}"]
|
||||
[id="network-observability-lokistack-configuring-ingestion_{context}"]
|
||||
|
||||
= LokiStack ingestion limits and health alerts
|
||||
The LokiStack instance comes with default settings according to the configured size. It is possible to override some of these settings, such as the ingestion and query limits. An automatic alert in the web console notifies you when these limits are reached.
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// network_observability/network-observability-scheduling-resources.adoc
|
||||
|
||||
:_mod-docs-content-type: CONCEPT
|
||||
[id="network-observability-multi-tenancy{context}"]
|
||||
[id="network-observability-multi-tenancy_{context}"]
|
||||
= Network observability deployment in specific nodes
|
||||
|
||||
You can configure the `FlowCollector` to control the deployment of network observability components in specific nodes. The `spec.agent.ebpf.advanced.scheduling`, `spec.processor.advanced.scheduling`, and `spec.consolePlugin.advanced.scheduling` specifications have the following configurable settings:
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// network_observability/observing-network-traffic.adoc
|
||||
|
||||
:_mod-docs-content-type: REFERENCE
|
||||
[id="network-observability-quickfilter{context}"]
|
||||
[id="network-observability-quickfilter_{context}"]
|
||||
= Filtering the network traffic
|
||||
|
||||
By default, the Network Traffic page displays the traffic flow data in the cluster based on the default filters configured in the `FlowCollector` instance. You can use the filter options to observe the required data by changing the preset filter.
|
||||
|
||||
@@ -27,7 +27,7 @@ You can edit the `FlowCollector` to view information about network traffic event
|
||||
. Select *cluster*, and then select the *YAML* tab.
|
||||
. Configure the `FlowCollector` CR to enable viewing `NetworkEvents`, for example:
|
||||
+
|
||||
[id="network-observability-flowcollector-configuring-networkevents{context}"]
|
||||
[id="network-observability-flowcollector-configuring-networkevents_{context}"]
|
||||
.Example `FlowCollector` configuration
|
||||
[source, yaml]
|
||||
----
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// network_observability/observing-network-traffic.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="network-observability-zones{context}"]
|
||||
[id="network-observability-zones_{context}"]
|
||||
= Working with availability zones
|
||||
|
||||
You can configure the `FlowCollector` to collect information about the cluster availability zones. This allows you to enrich network flow data with the link:https://kubernetes.io/docs/reference/labels-annotations-taints/#topologykubernetesiozone[`topology.kubernetes.io/zone`] label value applied to the nodes.
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * backup_and_restore/application_backup_and_restore/installing/overriding-kopia-algorithms.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="oadp-usecase-kopia-override-algorithms.adoc_{context}"]
|
||||
[id="oadp-usecase-kopia-override-algorithms_{context}"]
|
||||
= Use case for overriding Kopia hashing, encryption, and splitter algorithms
|
||||
|
||||
[role="_abstract"]
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// *cicd/pipelines/using-tekton-hub-with-openshift-pipelines.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="disabling-tekton-hub-authorization-after-upgrade.adoc_{context}"]
|
||||
[id="disabling-tekton-hub-authorization-after-upgrade_{context}"]
|
||||
= Disabling {tekton-hub} authorization after upgrading the {pipelines-title} Operator from 1.7 to 1.8
|
||||
|
||||
[role="_abstract"]
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// *cicd/pipelines/using-tekton-hub-with-openshift-pipelines.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="installing-tekton-hub-without-login-and-rating.adoc_{context}"]
|
||||
[id="installing-tekton-hub-without-login-and-rating_{context}"]
|
||||
= Installing {tekton-hub} without login and rating
|
||||
|
||||
[role="_abstract"]
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// *openshift_pipelines/working-with-pipelines-web-console.adoc
|
||||
|
||||
:_mod-docs-content-type: CONCEPT
|
||||
[id="op-interacting-pipelines_from_topology_view{context}"]
|
||||
[id="op-interacting-pipelines_from_topology_view_{context}"]
|
||||
= Interacting with pipelines from Topology view
|
||||
|
||||
The side panel of the application node in the *Topology* page displays the status of a pipeline run and you can interact with it.
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// *openshift_pipelines/working-with-pipelines-web-console.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="op-starting-pipelines_from_topology_view{context}"]
|
||||
[id="op-starting-pipelines-from-topology-view_{context}"]
|
||||
= Starting pipelines from Topology view
|
||||
|
||||
For pipelines created using the *From Git* option, you can use the *Topology* view to interact with pipelines after you start them:
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
//
|
||||
// * rosa_architecture/rosa_architecture_sub/rosa-architecture-models.adoc
|
||||
|
||||
[id="osd-aws-privatelink-architecture.adoc_{context}"]
|
||||
[id="osd-aws-privatelink-architecture_{context}"]
|
||||
= AWS PrivateLink architecture
|
||||
|
||||
The Red{nbsp}Hat managed infrastructure that creates AWS PrivateLink clusters is hosted on private subnets. The connection between Red{nbsp}Hat and the customer-provided infrastructure is created through AWS PrivateLink VPC endpoints.
|
||||
@@ -18,7 +18,7 @@ The following diagram shows network connectivity of a PrivateLink cluster.
|
||||
|
||||
image::156_OpenShift_ROSA_Arch_1221_privatelink.png[Multi-AZ AWS PrivateLink cluster deployed on private subnets]
|
||||
|
||||
[id="osd-aws-reference-architecture.adoc_{context}"]
|
||||
[id="osd-aws-reference-architecture_{context}"]
|
||||
== AWS reference architectures
|
||||
|
||||
AWS provides multiple reference architectures that can be useful to customers when planning how to set up a configuration that uses AWS PrivateLink. Here are three examples:
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
|
||||
//
|
||||
|
||||
:content-type: PROCEDURE
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="persistent-storage-csi-vsphere-encryption-datastore-url_{context}"]
|
||||
= Using datastore URL
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
|
||||
//
|
||||
|
||||
:content-type: PROCEDURE
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="persistent-storage-csi-vsphere-encryption-tag-based_{context}"]
|
||||
= Using tag-based placement
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
|
||||
//
|
||||
|
||||
:content-type: CONCEPT
|
||||
:_mod-docs-content-type: CONCEPT
|
||||
[id="persistent-storage-csi-vsphere-limitations_{context}"]
|
||||
= vSphere CSI limitations
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
|
||||
//
|
||||
|
||||
:content-type: PROCEDURE
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="persistent-storage-csi-vsphere-top-aware-during-install_{context}"]
|
||||
= Creating vSphere storage topology during installation
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
|
||||
//
|
||||
|
||||
:content-type: PROCEDURE
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="persistent-storage-csi-vsphere-top-aware-infra-top_{context}"]
|
||||
= Creating vSphere storage topology without an infra topology
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
|
||||
//
|
||||
|
||||
:content-type: CONCEPT
|
||||
:_mod-docs-content-type: CONCEPT
|
||||
[id="persistent-storage-csi-vsphere-top-aware-overview_{context}"]
|
||||
= vSphere CSI topology overview
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
|
||||
//
|
||||
|
||||
:content-type: PROCEDURE
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="persistent-storage-csi-vsphere-top-aware-post-install_{context}"]
|
||||
= Creating vSphere storage topology postinstallation
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// storage/container_storage_interface/persistent-storage-csi-vsphere.adoc
|
||||
//
|
||||
|
||||
:content-type: PROCEDURE
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="persistent-storage-csi-vsphere-top-aware-results_{context}"]
|
||||
= Results
|
||||
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * rosa_cluster_admin/rosa_nodes/rosa-managing-worker-nodes.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="rosa-adding-tags-cli{context}"]
|
||||
[id="rosa-adding-tags-cli_{context}"]
|
||||
= Adding tags to a machine pool using the ROSA CLI
|
||||
|
||||
You can add tags to a machine pool for your {product-title} cluster by using the ROSA command-line interface (CLI).
|
||||
|
||||
@@ -5,7 +5,7 @@
|
||||
// * osd_cluster_admin/osd_nodes/osd-managing-worker-nodes.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="rosa-adding-taints-cli{context}"]
|
||||
[id="rosa-adding-taints-cli_{context}"]
|
||||
= Adding taints to a machine pool using the ROSA CLI
|
||||
|
||||
You can add taints to a machine pool for your {product-title} cluster by using the ROSA CLI.
|
||||
|
||||
@@ -5,7 +5,7 @@
|
||||
// * osd_cluster_admin/osd_nodes/osd-managing-worker-nodes.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="rosa-adding-taints-ocm{context}"]
|
||||
[id="rosa-adding-taints-ocm_{context}"]
|
||||
= Adding taints to a machine pool using {cluster-manager}
|
||||
|
||||
You can add taints to a machine pool for your {product-title} cluster by using {cluster-manager-first}.
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * rosa_install_access_delete_clusters/rosa-sts-creating-a-cluster-with-customizations.adoc
|
||||
|
||||
:_mod-docs-content-type: CONCEPT
|
||||
[id="rosa-imds{context}"]
|
||||
[id="rosa-imds_{context}"]
|
||||
= Instance Metadata Service
|
||||
|
||||
There are two types of ways to access instance metadata from a running instance:
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * rosa_release_notes/rosa-release-notes.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="updating_rosa_cli{context}"]
|
||||
[id="updating-rosa-cli_{context}"]
|
||||
== Updating the ROSA CLI tool
|
||||
|
||||
To use the latest version of the {product-title} (ROSA) CLI, `rosa`, download the ROSA CLI (`rosa`) from the Hybrid Cloud Console. If you already have this tool, the procedure is the same for updates.
|
||||
|
||||
@@ -38,7 +38,7 @@ Custom domains and subdomains are not available for the platform service routes,
|
||||
* Optional: Secondary router/ingress load balancer that is a secondary application load balancer, denoted by `apps2` in the URL. The secondary load balancer can be configured in {cluster-manager} to be either publicly accessible over the internet, or only privately accessible over a pre-existing private connection. If a 'Label match' is configured for this router load balancer, then only application routes matching this label will be exposed on this router load balancer, otherwise all application routes are also exposed on this router load balancer.
|
||||
* Optional: Load balancers for services that can be mapped to a service running on {product-title} to enable advanced ingress features, such as non-HTTP/SNI traffic or the use of non-standard ports. These can be purchased in groups of 4 for non-CCS clusters, or they can be provisioned through the cloud provider console in Customer Cloud Subscription (CCS) clusters; however, each AWS account has a quota that link:https://docs.aws.amazon.com/elasticloadbalancing/latest/classic/elb-limits.html[limits the number of Classic Load Balancers] that can be used within each cluster.
|
||||
|
||||
[id="network-usage{context}"]
|
||||
[id="network-usage_{context}"]
|
||||
== Network usage
|
||||
For non-CCS {product-title} clusters, network usage is measured based on data transfer between inbound, VPC peering, VPN, and AZ traffic. On a non-CCS {product-title} base cluster, 12 TB of network I/O is provided. Additional network I/O can be purchased in 12 TB increments. For CCS {product-title} clusters, network usage is not monitored, and is billed directly by the cloud provider.
|
||||
|
||||
|
||||
@@ -84,7 +84,7 @@ All activities performed by SREs arrive from Red Hat IP addresses and are logged
|
||||
|
||||
For a full list of permissions, see `sts_support_permission_policy.json` in the link:https://docs.openshift.com/rosa/rosa_architecture/rosa-sts-about-iam-resources.html[About IAM resources] user guide.
|
||||
|
||||
[id="rosa-sre-access-privatelink-vpc.adoc_{context}"]
|
||||
[id="rosa-sre-access-privatelink-vpc_{context}"]
|
||||
== SRE access through PrivateLink VPC endpoint service
|
||||
|
||||
PrivateLink VPC endpoint service is created as part of the ROSA cluster creation.
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * updating/updating_a_cluster/updating-cluster-cli.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="update-conditional-upgrade-path{context}"]
|
||||
[id="update-conditional-upgrade-path_{context}"]
|
||||
= Updating along a conditional update path
|
||||
|
||||
You can update along a recommended conditional update path using the web console or the OpenShift CLI (`oc`).
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * virt/support/virt-troubleshooting.adoc
|
||||
|
||||
:_mod-docs-content-type: CONCEPT
|
||||
[id="virt-about-dv-conditions-and-events.adoc_{context}"]
|
||||
[id="virt-about-dv-conditions-and-events_{context}"]
|
||||
= About data volume conditions and events
|
||||
|
||||
You can diagnose data volume issues by examining the output of the `Conditions` and `Events` sections
|
||||
|
||||
@@ -3,7 +3,7 @@
|
||||
// * virt/virtual_machines/virt-deprecated-tasks.adoc
|
||||
|
||||
:_mod-docs-content-type: PROCEDURE
|
||||
[id="virt-deprecated-tasks.web.adoc_{context}"]
|
||||
[id="virt-deprecated-tasks.web_{context}"]
|
||||
= Removing deprecated or unused resources
|
||||
|
||||
You can clean up deprecated or unused resources associated with the {pipelines-title} Operator.
|
||||
|
||||
@@ -402,7 +402,7 @@ With the duplicated flows update, the *Interface* and *Interface Direction* fiel
|
||||
=====
|
||||
|
||||
For more information, see xref:../../observability/network_observability/network-observability-operator-monitoring.adoc#network-observability-netobserv-dashboard-ebpf-agent-alerts_network_observability[Using the eBPF agent alert]
|
||||
and xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-quickfilternw-observe-network-traffic[Quick filters].
|
||||
and xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-quickfilter_nw-observe-network-traffic[Quick filters].
|
||||
|
||||
|
||||
[id="network-observability-ebpf-collection-filtering-1.6_{context}"]
|
||||
@@ -464,7 +464,7 @@ You can create Prometheus alerts for the *Netobserv* dashboard using DNS, Packet
|
||||
|
||||
[id="network-observability-zones"]
|
||||
==== Availability zones
|
||||
You can configure the `FlowCollector` resource to collect information about the cluster availability zones. This configuration enriches the network flow data with the link:https://kubernetes.io/docs/reference/labels-annotations-taints/#topologykubernetesiozone[`topology.kubernetes.io/zone`] label value applied to the nodes. For more information, see xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-zonesnw-observe-network-traffic[Working with availability zones].
|
||||
You can configure the `FlowCollector` resource to collect information about the cluster availability zones. This configuration enriches the network flow data with the link:https://kubernetes.io/docs/reference/labels-annotations-taints/#topologykubernetesiozone[`topology.kubernetes.io/zone`] label value applied to the nodes. For more information, see xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-zones_nw-observe-network-traffic[Working with availability zones].
|
||||
|
||||
[id="network-observability-enhanced-configuration-and-ui-1.5"]
|
||||
==== Notable enhancements
|
||||
@@ -583,7 +583,7 @@ The 1.4 release of the Network Observability Operator adds improvements and new
|
||||
** The *NetObserv / Health* dashboard shows flows overhead as well as top flow rates per nodes, namespaces, and workloads.
|
||||
** Infrastructure and Application metrics are shown in a split-view for namespaces and workloads.
|
||||
|
||||
For more information, see xref:../../observability/network_observability/network-observability-overview.adoc#network-observability-dashboards[Network Observability metrics] and xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-quickfilternw-observe-network-traffic[Quick filters].
|
||||
For more information, see xref:../../observability/network_observability/network-observability-overview.adoc#network-observability-dashboards[Network Observability metrics] and xref:../../observability/network_observability/observing-network-traffic.adoc#network-observability-quickfilter_nw-observe-network-traffic[Quick filters].
|
||||
|
||||
[discrete]
|
||||
[id="configuration-enhancements-1.4_{context}"]
|
||||
|
||||
@@ -56,5 +56,5 @@ include::modules/understanding-update-channels.adoc[leveloffset=+1]
|
||||
|
||||
[role="_additional-resources"]
|
||||
.Additional resources
|
||||
* xref:../../updating/updating_a_cluster/updating-cluster-cli.adoc#update-conditional-upgrade-pathupdating-cluster-cli[Updating along a conditional upgrade path]
|
||||
* xref:../../updating/updating_a_cluster/updating-cluster-cli.adoc#update-conditional-upgrade-path_updating-cluster-cli[Updating along a conditional upgrade path]
|
||||
* xref:../../updating/understanding_updates/understanding-update-channels-release.adoc#fast-stable-channel-strategies_understanding-update-channels-releases[Choosing the correct channel for your cluster]
|
||||
|
||||
@@ -69,7 +69,7 @@ include::modules/update-upgrading-oc-adm-upgrade-status.adoc[leveloffset=+1]
|
||||
ifndef::openshift-origin[]
|
||||
* xref:../../updating/updating_a_cluster/control-plane-only-update.adoc#control-plane-only-update[Performing a Control Plane Only update]
|
||||
endif::openshift-origin[]
|
||||
* xref:../../updating/updating_a_cluster/updating-cluster-cli.adoc#update-conditional-upgrade-pathupdating-cluster-cli[Updating along a conditional update path]
|
||||
* xref:../../updating/updating_a_cluster/updating-cluster-cli.adoc#update-conditional-upgrade-path_updating-cluster-cli[Updating along a conditional update path]
|
||||
ifndef::openshift-origin[]
|
||||
* xref:../../updating/understanding_updates/understanding-update-channels-release.adoc#understanding-update-channels-releases[Understanding update channels and releases]
|
||||
endif::openshift-origin[]
|
||||
|
||||
Reference in New Issue
Block a user