mirror of
https://github.com/openshift/openshift-docs.git
synced 2026-02-06 15:46:57 +01:00
42 lines
1.7 KiB
Plaintext
42 lines
1.7 KiB
Plaintext
[id="serverless-kafka-developer"]
|
|
= Knative Kafka
|
|
include::modules/serverless-document-attributes.adoc[]
|
|
include::modules/common-attributes.adoc[]
|
|
:context: serverless-kafka-developer
|
|
|
|
toc::[]
|
|
|
|
Knative Kafka functionality is available in an {ServerlessProductName} installation xref:../../serverless/admin_guide/serverless-kafka-admin.adoc#serverless-kafka-admin[if a cluster administrator has installed the `KnativeKafka` custom resource].
|
|
|
|
Knative Kafka provides additional options, such as:
|
|
|
|
* Kafka event source
|
|
* xref:../../serverless/develop/serverless-creating-channels.adoc#serverless-creating-channels[Kafka channel]
|
|
// * Kafka broker
|
|
|
|
[NOTE]
|
|
====
|
|
Knative Kafka is not currently supported for IBM Z and IBM Power Systems.
|
|
====
|
|
|
|
include::modules/serverless-kafka-event-delivery.adoc[leveloffset=+1]
|
|
|
|
[id="serverless-kafka-developer-event-source"]
|
|
== Using a Kafka event source
|
|
|
|
You can create a Knative Kafka event source that reads events from an Apache Kafka cluster and passes these events to a sink.
|
|
|
|
// dev console
|
|
include::modules/serverless-kafka-source-odc.adoc[leveloffset=+2]
|
|
// kn commands
|
|
include::modules/serverless-kafka-source-kn.adoc[leveloffset=+2]
|
|
include::modules/specifying-sink-flag-kn.adoc[leveloffset=+3]
|
|
// YAML
|
|
include::modules/serverless-kafka-source-yaml.adoc[leveloffset=+2]
|
|
|
|
[id="additional-resources_serverless-kafka-developer"]
|
|
== Additional resources
|
|
|
|
* See the link:https://access.redhat.com/documentation/en-us/red_hat_amq/7.6/html/amq_streams_on_openshift_overview/kafka-concepts_str#kafka-concepts-key_str[Red Hat AMQ Streams] documentation for more information about Kafka concepts.
|
|
* See xref:../../serverless/discover/knative-event-sources.adoc#knative-event-sources[Event sources].
|