Skip to content

Latest commit

 

History

History
57 lines (41 loc) · 4.16 KB

overview.rst

File metadata and controls

57 lines (41 loc) · 4.16 KB

|cp| Demo (cp-demo)

The cp-demo example builds a full |cp| deployment with an |ak-tm| event streaming application using ksqlDB and Kafka Streams for stream processing, and all the components have security enabled end-to-end. The tutorial includes a module to extend it into a hybrid deployment that runs Cluster Linking and Schema Linking to copy data and schemas from a local on-prem |ak| cluster to |ccloud|, a fully-managed service for |ak-tm|. Follow the accompanying guided tutorial, broken down step-by-step, to learn how |ak| and |ccloud| work with |kconnect|, |sr-long|, |c3|, Cluster Linking, and security enabled end-to-end.

Overview

Use Case

The use case is an |ak-tm| event streaming application that processes real-time edits to real Wikipedia pages.

image

The full event streaming platform based on |cp| is described as follows. Wikimedia's EventStreams publishes a continuous stream of real-time edits happening to real wiki pages. A Kafka source connector kafka-connect-sse streams the server-sent events (SSE) from https://stream.wikimedia.org/v2/stream/recentchange, and a custom |kconnect| transform kafka-connect-json-schema extracts the JSON from these messages and then are written to a |ak| cluster. This example uses ksqlDB and a :ref:`Kafka Streams <kafka_streams>` application for data processing. Then a Kafka sink connector kafka-connect-elasticsearch streams the data out of Kafka and is materialized into Elasticsearch for analysis by Kibana. All data is using |sr-long| and Avro, and Confluent Control Center is managing and monitoring the deployment.

Data Pattern

Data pattern is as follows:

Components Consumes From Produces To
SSE source connector Wikipedia wikipedia.parsed
ksqlDB wikipedia.parsed ksqlDB streams and tables
Kafka Streams application wikipedia.parsed wikipedia.parsed.count-by-domain
Elasticsearch sink connector WIKIPEDIABOT (from ksqlDB) Elasticsearch/Kibana

How to use this tutorial

We suggest following the cp-demo tutorial in order:

  1. :ref:`cp-demo-on-prem-tutorial`: bring up the on-prem |ak| cluster and explore the different technical areas of |cp|
  2. :ref:`cp-demo-hybrid`: create a cluster link to copy data from a local on-prem |ak| cluster to |ccloud|, and use the Metrics API to monitor both
  3. :ref:`cp-demo-teardown`: clean up your on-prem and |ccloud| environment