Apache Kafka® Transaction Data Streaming For Dummies
Apache Kafka® Transaction Data Streaming For Dummies
Transaction
Data Streaming
Confluent & Qlik® Special Edition
by Thornton J. Craig,
Kevin Petrie, Tim Berglund
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Apache Kafka® Transaction Data Streaming For Dummies®,
Confluent & Qlik® Special Edition
Published by
John Wiley & Sons, Inc.
111 River St.
Hoboken, NJ 07030-5774
www.wiley.com
Copyright © 2022 by John Wiley & Sons, Inc.
No part of this publication may be reproduced, stored in a retrieval system or transmitted in any
form or by any means, electronic, mechanical, photocopying, recording, scanning or otherwise,
except as permitted under Sections 107 or 108 of the 1976 United States Copyright Act, without
the prior written permission of the Publisher. Requests to the Publisher for permission should
be addressed to the Permissions Department, John Wiley & Sons, Inc., 111 River Street, Hoboken,
NJ 07030, (201) 748-6011, fax (201) 748-6008, or online at http://www.wiley.com/go/permissions.
Trademarks: Wiley, For Dummies, the Dummies Man logo, The Dummies Way, Dummies.com,
Making Everything Easier, and related trade dress are trademarks or registered trademarks of John
Wiley & Sons, Inc. and/or its affiliates in the United States and other countries, and may not be
used without written permission. Confluent and the Confluent logo are registered trademarks of
Confluent. Qlik and the Qlik logo are registered trademarks of Qlik. Apache, Apache Kafka, Kafka,
and associated open source project names are trademarks of the Apache Software Foundation.
All other trademarks are the property of their respective owners. John Wiley & Sons, Inc., is not
associated with any product or vendor mentioned in this book.
For general information on our other products and services, or how to create a custom For
Dummies book for your business or organization, please contact our Business Development
Department in the U.S. at 877-409-4177, contact info@dummies.biz, or visit www.wiley.com/go/
custompub. For information about licensing the For Dummies brand for products or services,
contact BrandedRights&Licenses@Wiley.com.
ISBN: 978-1-119-62601-5 (pbk); ISBN: 978-1-119-62603-9 (ebk)
Manufactured in the United States of America
10 9 8 7 6 5 4 3 2 1
Publisher’s Acknowledgments
Some of the people who helped bring this book to market include the
following:
Project Editor: Acquisitions Editor: Steve Hayes
Carrie Burchfield-Leighton Business Development
Editorial Manager: Rev Mengle Representative: Molly Daugherty
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
01_9781119626015-ffirs.indd ii Trim size: 5.5 in × 8.5 in September 20, 2022 3:09 PM
Table of Contents
INTRODUCTION................................................................................................ 1
About This Book.................................................................................... 1
Icons Used in This Book........................................................................ 1
Beyond the Book................................................................................... 2
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
CHAPTER 6: Starting Your Journey: Effective Planning.............. 33
Planning an Effective Transaction Data Streaming System........... 33
Transaction Consistency..................................................................... 35
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Introduction
A
pache Kafka is a highly flexible streaming platform that
supports multiple, key use cases in modern data architec-
tures. One critical use for Kafka, and the focus of this book,
is building scalable, real-time data pipelines. Streams of unbounded
data are effectively ingested, persisted, and delivered using Kafka as
a framework. High-volume, high-velocity data from social media
feeds, Internet of Things (IoT) sources, and any database transac-
tions are written, replicated, and read in real-time. Change data
capture (CDC) database replication technologies leverage Kafka to
enable a highly effective platform supporting real-time transac-
tional database streams.
The Tip icon points out helpful suggestions and useful nuggets of
information.
Introduction 1
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Beyond the Book
This book can help you discover more about implementing Kafka
transaction streaming, but if you want resources beyond what
this book offers, we have some insight for you:
»» https://www.qlik.com/us/products/qlik-replicate:
Qlik Replicate® empowers organizations to accelerate data
replication, ingestion and streaming across a wide variety of
heterogeneous databases, data warehouses, and big data
platforms.
»» https://cnfl.io/express-scripts: An on-demand
webinar with joint customer, Express Scripts
»» https://www.qlik.com/us/resource-library/streaming-
change-data-capture: This book serves as a practical guide
for enterprise architects, data managers and CIOs as they
enable modern data lake, streaming and cloud architectures
with CDC.
»» www.confluent.io/product/confluent-platform:
Confluent’s streaming platform based on Kafka
»» www.confluent.io/confluent-cloud: Kafka for the cloud
»» https://kafka-tutorials.confluent.io: A collection of
event streaming use cases, with each tutorial featuring an
example scenario and several complete code solutions
»» https://docs.confluent.io/current/tutorials/
index.html: Detailed demos of Confluent and Kafka
»» http://kafka.apache.org: Especially for developers,
documentation that offers deeper dives into implementation
and code details
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Chapter 1
Understanding
Transaction Streaming
with Apache Kafka
M
odern data requirements are real time. This is the case
with processing online transactions or social media
feeds, capturing Internet of Things (IoT) data, or gener-
ating “just-in-time” or instant analytics on any type of database
transaction. Rapid, incremental, and high-volume changes in
data require ultra-fast replication or processing. Traditional batch
processes operate on a schedule using a fixed time window to
process database transactions, often slowing or pausing produc-
tion operations during off-hours given the high impact of its full-
load replication on source databases. Batch processes struggle to
support real-time and continuous changes in databases.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
then create a stream. The stream of changes is then collected,
persisted, and delivered using a streaming platform like Kafka.
The sources for batch and streaming data are often the same,
although newer sources like IoT or other real-time feeds are usu-
ally associated with streaming data. Even non-real-time data can
be a stream, like CDC database transactions. The key differen-
tiator between batch and streaming is that the stream data isn’t
bounded or captured in a specific unit or group to be processed.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
FIGURE 1-1: Comparing streaming data and traditional batch data processes.
Geolocation analysis
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Real time and near real time differ and depend on use cases. For
some requirements, a 15-minute delay is real time; for others it
needs to be sub-second.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
TABLE 1-2 Database Types and Functions
Database Type Examples Roles/Use Cases
MySQL
PostgreSQL
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
method is tried and tested, highly effective, and low impact to
capture and replicate data changes into target systems. But what
if you need to persist the stream of transactions? Or if the stream
needs to be consumed by multiple targets? Perhaps some con-
sumers require near real time, while others retrieve infrequently
using intermittent jobs.
Machine learning Stream processing (see Chapter 5 for more info) enables
analytics using AI/ML operations.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Kafka Use Cases Description
Data persistence By default, Kafka persists data records for seven days. This
timeframe is configurable down to minutes or up to
forever, based on demands and use cases.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
IN THIS CHAPTER
»» Studying the Kafka architecture
Chapter 2
Looking Deeper into
Apache Kafka
A
pache Kafka is a streaming platform. What exactly does
that mean? Essentially, a streaming platform ingests, per-
sists, and presents streams of data for consumption. The
stream can be any type of data — any source that can be serialized
into a record. In this chapter, we dig a little deeper into the anat-
omy of Kafka, so you can better understand its moving parts and
why Kafka is such a powerful streaming platform.
With a deeper view of Kafka, you see that topics resemble data-
base transaction logs. Transaction data, captured for example via
change data capture (CDC), is essentially a log of records that’s
persisted and made consumable at scale. These records can be
consumed for multiple use cases (databases, microservices,
immediate stream analytics, and so on), and because they’re
stored for longer periods, they can be replayed or consumed at
slower velocities, and in order. Kafka offers a solution for trans-
action data through its log-based architecture. For even more
information on Kafka, visit https://kafka.apache.org.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Looking at the Kafka Architecture
Processing data with Kafka relies on three basic components: pro-
ducers, brokers, and consumers. Figure 2-1 shows a basic Kafka
architecture.
Kafka producers
The producer is a piece of code or process that writes data to Kafka.
Producers use a serializer and an optional key to create a record
containing the source data. The record is then written to Kafka.
The producer’s components are detailed as follows:
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
»» De-serializer: This is used by the consumer to convert
records back into the original data.
»» Key: A key is an optional identifier used to hash (select)
which partition will receive a record. This function is used to
target records to specific partitions.
Kafka cluster
A Kafka cluster is a collection of brokers. Brokers are the basic
functional unit in Kafka. A broker runs the Kafka process and
responds to requests from producers and consumers. Brokers rely
on, manage, and/or interact with the following components:
Kafka consumers
The consumer is an application program that reads records from
a topic. Topics can be read by a single consumer, or if the topic is
partitioned, a group of consumers can read from its partitions in
parallel. This enables faster throughput and lower latency, which
allow scaling of reads — an important concept we talk about more
with streaming transactions in Chapter 4.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
An offset is a marked position that Kafka consumers use to keep
track of the most recent message they’ve consumed from each
partition. This enables consumers to pause read operations and
then pick up where they left off and resume in sequence.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
ORDER, SEMANTICS, AND
CONSISTENCY — OH MY!
Kafka captures database events as transactions by using a producer
to create a record containing the details of the event. Each record is
written to a Kafka topic. Because records of the same key always land
in the same topic, transactions related to the same key are kept
strictly in order. Database transactions often require transactional
consistency — defined here as being holistic, ordered, and committed
only one time (exactly once semantics). You can find more about
“exactly once” semantics and options with Kafka Streams at www.
confluent.io/blog/enabling-exactly-once-kafka-streams.
We also want to make the distinction between database transaction
consistency and exactly once semantics: They overlap, and you need
both, but they’re different things. We cover this more in Chapter 5.
Events (records in Kafka) can also be grouped by topic into one parti-
tion, guaranteeing the correct order. Other methods include sharing
transactional metadata by using headers that are attached to Kafka
records. You can read more about this in Chapter 6.
Figure 2-2 shows the anatomy of how Kafka uses topics, parti-
tions, and records.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
IN THIS CHAPTER
»» Integrating data with Apache Kafka
Connect API
Chapter 3
Ingesting Data with
Apache Kafka
T
raditional data integration pipelines focus on Extract,
Transform, and Load (ETL) or, as is now common with big
data architectures, Extract, Load, and Transform (ELT) pro-
cessing. Data is extracted using a process like change data capture
(CDC) and/or by writing data to another database or target. Data
is then either transformed before it is loaded (for example, ETL)
or loaded into an analytical or target system and then transformed
(ELT). The emergence of data lakes and other big data architec-
tures has blurred the line between ETL and ELT, and both pat-
terns are now commonly used.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Kafka supports common data integration methods, and we cover
two of them here: Kafka Connect and CDC. We also cover common
data integration architectural patterns in the last section of this
chapter.
Kafka Connect
Kafka Connect is an open-source component of Kafka that sup-
ports connections to common databases. Connect creates a com-
mon framework to develop, deploy, and manage connectors to
Kafka. Common scenarios implemented with Connect include
database ingestion, collecting metrics from application servers,
or database extraction processes. Each of these workflows can
be processed incrementally, in near real time, or in batches. For
more info on Kafka Connect, visit https://docs.confluent.io/
current/connect/index.html or https://kafka.apache.org/
documentation/#connect.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
(see Chapter 7 for more info), to provide prebuilt connectors for
common databases and other data sources. Using a managed ser-
vice for Kafka Connect also helps provide a fully scalable Kafka
Connect infrastructure.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Table 3-1 describes the impact and requirements for CDC capture
methods, which sometimes depend on data source requirements.
Query-based Often used if the source does not have change Low
logs, as is generally the case with data
warehouses. This method has similar use cases
as triggers, but with less impact. The CDC engine
queries the database at regular intervals,
although the time between intervals can
significantly increase latency.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
TABLE 3-2 Data Integration Architectural Patterns
Pattern Description
Data lake integration This entails data ingestion into a data lake or other
analytics platform. Many sources may feed into a data
lake and require CDC + Kafka to effectively ingest and
transform data.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
IN THIS CHAPTER
»» Using CDC and Kafka to deliver real-time
results
Chapter 4
Applying CDC
to Apache Kafka
T
he complete change data capture (CDC) process consists of
two functions: capturing the changed data and enabling
replication of the changed data. Both functions grouped
together are commonly referred to as CDC or replication (although
data replication is technically speaking a distinct process from
CDC). Understanding this two-step process is important as you
look at CDC and Apache Kafka in this chapter. CDC captures and
presents data changes to Kafka. Kafka becomes the persistence
and transit layer in which CDC changes are replicated and
delivered.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
producers to Kafka, eliminating the need for scripting and greatly
simplifying management.
Topic schemas and partition usage (see the next section “Topics
and Partitions” for more info) should be aligned with throughput
and latency requirements (more on this in Chapter 6). Producers
can use optimized schema serializers, such as Apache Avro, to
ensure consistency in data stream pipelines. Many organizations
leverage dedicated schema registries such as Confluent’s Schema
Registry. These registries allow for version control and updates of
topic schemas while enforcing record consistency.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Getting Organized: Topics and Partitions
Two key goals of a streaming platform are to optimize data
throughput and ensure consistency. CDC inserts, updates, deletes,
and DDL changes are most commonly captured from database
sources. Kafka’s architecture leverages topics to provide structure,
partitions for scalability (high throughput), and replication for
redundancy. Transaction streaming use cases are often a tradeoff
between high throughput, guaranteed ordering, and low latency.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
TABLE 4-2 Kafka Partition Configuration Options
Option Benefits Considerations
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
IN THIS CHAPTER
»» Defining the core components that
support stream processing
Chapter 5
Understanding Stream
Processing with Apache
Kafka and Its Benefits
S
tream processing is a method of performing transforma-
tions or generating analytics on transactional data inside a
stream. Traditional Extract, Transform, Load (ETL)-based
data integration functions are performed on data that will be ana-
lyzed in a database, data lake, or data warehouse. Analytics are
typically run against a data warehouse with structured and
cleansed data. In contrast, streaming platforms like Apache Kafka
enable both integration and in-stream analytics against data as it
moves through the stream.
CHAPTER 5 Understanding Stream Processing with Apache Kafka and Its Benefits 27
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
In this chapter, you discover the components that support stream
processing and learn why stream processing is beneficial.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
In Figure 5-2, we show you a simple example of the relationship
between tables and streams.
KSQL
KSQL is a SQL-like language for describing stream process-
ing operations on data in Kafka. It uses similar stream and table
abstractions as the KStream and KTable classes found in the Kafka
Streams API (see the preceding section).
CHAPTER 5 Understanding Stream Processing with Apache Kafka and Its Benefits 29
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
TABLE 5-1 Windowing Options in Kafka
Window Type Description
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
TABLE 5-2 Stream Processing Use Cases
Stream Processing Use Case Description
Data integration The Streams API and KSQL both support basic
to more-complex data integration scenarios,
using simple SQL-like commands and queries.
CHAPTER 5 Understanding Stream Processing with Apache Kafka and Its Benefits 31
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
IN THIS CHAPTER
»» Planning transaction data streaming
systems
Chapter 6
Starting Your Journey:
Effective Planning
I
n this chapter, we explain planning and implementing your
CDC and Apache Kafka streaming architecture. You explore
common characteristics of effective streaming approaches,
including careful pipeline design, standardized process, pooled
resources, granular monitoring, data governance policies, and
productive collaboration. You dive into implementation best prac-
tices, including configuration methods to ensure transaction con-
sistency. We also give you a case study from Generali, a Fortune
100 financial services organization, in implementing CDC and
Apache Kafka.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
attributes of effective streaming approaches to transaction data
that span people, process, and technology are as follows:
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
explicitly authorized by the original owners of that data. The
central IT organization should define, monitor usage, and
enforce policies, and then provide reports to the appropriate
internal compliance officer.
Transaction Consistency
Most analytics use cases for transaction data require records to
be accurate, and this requires transaction consistency, which is
the guarantee that committed transactions relate to one another
according to consistent rules imposed by databases or other repos-
itories. Transactions must be committed accurately, atomically, in
order, and only once. Ideally, you’ll maintain the same level of con-
sistency as records travel from source database through each point
in the Kafka streaming architecture. Two methods to consider for
maintaining transaction consistency include the following:
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
UNIFYING THE CUSTOMER
EXPERIENCE WITH CDC
AND KAFKA
Generali Group is one of the leading insurers in the world, with
insurance and investment services to 61 million customers in over
50 countries. Generali Switzerland like its Italian parent, believes its
business is fundamentally based on data. And its IT organization says
unlocking the value of that data is akin to awakening a “sleeping
beauty.” Generali Switzerland seeks to achieve this by implementing a
new data streaming architecture based on CDC and Kafka. With this
architecture, the company intends to create a simple, flexible, and
unified omni-channel customer experience that reduces churn,
increases revenue per customer, and improves operational efficiency.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
siloed scripted procedures. Meanwhile, Qlik Replicate worked with
Confluent Platform to provide efficient transaction data integration
across applications, with easy connectors to enable agile development
of new software components. You can read more about Confluent
Platform in Chapter 7.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
IN THIS CHAPTER
»» Understanding Confluent Platform
Chapter 7
Ten Reasons to Choose
Confluent and Qlik
E
very For Dummies book ends in a chapter called the Part of
Tens. In this chapter, we’ve chosen ten reasons why
Confluent and Qlik products are good choices for transaction
streaming with Apache Kafka. We’ve divided this chapter into two
sections, focusing on five reasons for each company’s products.
We start with Confluent and end with Qlik.
These materials are © 2022John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Development and integration flexibility
Confluent Platform easily integrates with many existing systems
as part of a broad and expanding partner ecosystem. The Kafka
Connect API and connectors provide developers, data engineers,
and operators a simple way to stream records between popular
applications, sources, and targets. Some connectors are included
in the platform (as JDBC, S3, HDFS, and Elasticsearch) and
included as part of Confluent Platform. Many more can be down-
loaded from Confluent Hub.
Comprehensive management,
monitoring, and control
A key requirement of transaction streaming is to understand and
control Kafka cluster operations. Confluent Control Center provides
a curated, GUI-based dashboard that reveals insights about the
inner workings of your Kafka clusters and the data flowing through
them. You gain key operational and monitoring capabilities, as well
as new confidence to meet service level agreements (SLAs). You
can assess cluster health, availability, and scalability while track-
ing KPIs for end-to-end performance and broker/cluster resource
utilization across environments based on any Kafka client and any
programming language. These KPIs and threshold-based alerts
address record delivery status, latency, throughput/consumer lag,
and missing/duplicate/delayed events.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
availability, transient high-throughput, or infinite retention. The
results are higher confidence and lower risk.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Stream processing interface
Confluent products use a stream processing interface called
KSQL. This streaming SQL engine helps developers build powerful
continuous stream processing queries against Kafka with famil-
iar SQL-like semantics. KSQL provides an intuitive interactive
SQL interface for stream processing on Kafka, without the need
to write code in a programming language. You simply perform
stream processing tasks using SQL-like statements. KSQL is scal-
able, elastic, and fault-tolerant, and it supports a wide range of
streaming operations, including data filtering, transformations,
aggregations, joins, and windowing. It enables use cases such as
streaming ETL, scoring, and anomaly detection. (Find out more
about KSQL in Chapter 5.)
The same flexible process applies to all end point types, helping
you add or remove producers in a modular fashion. Qlik Replicate
fully supports and automates all topic and partition configuration
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
options available with Kafka, executing and monitoring through a
centralized, intuitive GUI.
Fan-out capabilities
Enterprises often have multiple uses for production transactional
data. The same customer purchase table might need to be con-
sumed real-time by dozens or more systems for fraud prevention,
next best offer recommendations and supply-chain optimization.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
Qlik Replicate enables data engineers to automatically provision a
single database producer, once, that Kafka can then map to many
streams and consumers, eliminating the need for duplicative con-
figuration processes. These fan-out capabilities have the added
benefit of minimizing impact on production workloads.
Metadata integration
Qlik Replicate CDC automatically propagates source schema
changes to any supported target, including Kafka. Data engineers
ensure that consumers are synchronized with all relevant source
structures. For example, Qlik Replicate automatically captures
and propagates all the metadata associated with source schema/
DDL changes. This includes database level metadata, for example,
to correctly identify date or time fields that databases may not
otherwise make available to Kafka running on the Kafka Avro and
JSON formats.
These materials are © 2022 John Wiley & Sons, Inc. Any dissemination, distribution, or unauthorized use is strictly prohibited.
WILEY END USER LICENSE AGREEMENT
Go to www.wiley.com/go/eula to access Wiley’s ebook EULA.