Change data feed
Note
This feature is available in Delta Lake 2.0.0 and above. This feature is in experimental support mode.
Change Data Feed (CDF) feature allows Delta tables to track row-level changes between versions of a Delta table. When enabled on a Delta table, the runtime records “change events” for all the data written into the table. This includes the row data along with metadata indicating whether the specified row was inserted, deleted, or updated.
You can read the change events in batch queries using DataFrame APIs (that is, df.read
) and in streaming queries using DataFrame APIs (that is, df.readStream
).
Use cases
Change Data Feed is not enabled by default. The following use cases should drive when you enable the change data feed.
Silver and Gold tables: Improve Delta performance by processing only row-level changes following initial
MERGE
,UPDATE
, orDELETE
operations to accelerate and simplify ETL and ELT operations.Transmit changes: Send a change data feed to downstream systems such as Kafka or RDBMS that can use it to incrementally process in later stages of data pipelines.
Audit trail table: Capture the change data feed as a Delta table provides perpetual storage and efficient query capability to see all changes over time, including when deletes occur and what updates were made.
Enable change data feed
You must explicitly enable the change data feed option using one of the following methods:
New table: Set the table property
delta.enableChangeDataFeed = true
in theCREATE TABLE
command.CREATE TABLE student (id INT, name STRING, age INT) TBLPROPERTIES (delta.enableChangeDataFeed = true)
Existing table: Set the table property
delta.enableChangeDataFeed = true
in theALTER TABLE
command.ALTER TABLE myDeltaTable SET TBLPROPERTIES (delta.enableChangeDataFeed = true)
All new tables:
set spark.databricks.delta.properties.defaults.enableChangeDataFeed = true;
Important
Once you enable the change data feed option for a table, you can no longer write to the table using Delta Lake 1.2.1 or below. You can always read the table.
Only changes made after you enable the change data feed are recorded; past changes to a table are not captured.
Change data storage
Delta Lake records change data for UPDATE
, DELETE
, and MERGE
operations in the _change_data
folder under the Delta table directory.
These records may be skipped when Delta Lake detects it can efficiently compute the change data feed directly from the transaction log.
In particular, insert-only operations and full partition deletes will not generate data in the _change_data
directory.
The files in the _change_data
folder follow the retention policy of the table. Therefore, if you run the VACUUM command, change data feed data is also deleted.
Read changes in batch queries
You can provide either version or timestamp for the start and end. The start and end versions and timestamps are inclusive in the queries. To read the changes from a particular start version to the latest version of the table, specify only the starting version or timestamp.
You specify a version as an integer and a timestamps as a string in the format yyyy-MM-dd[ HH:mm:ss[.SSS]]
.
If you provide a version lower or timestamp older than one that has recorded change events, that is, when the change data feed was enabled, an error is thrown indicating that the change data feed was not enabled.
# version as ints or longs
spark.read.format("delta") \
.option("readChangeFeed", "true") \
.option("startingVersion", 0) \
.option("endingVersion", 10) \
.table("myDeltaTable")
# timestamps as formatted timestamp
spark.read.format("delta") \
.option("readChangeFeed", "true") \
.option("startingTimestamp", '2021-04-21 05:45:46') \
.option("endingTimestamp", '2021-05-21 12:00:00') \
.table("myDeltaTable")
# providing only the startingVersion/timestamp
spark.read.format("delta") \
.option("readChangeFeed", "true") \
.option("startingVersion", 0) \
.table("myDeltaTable")
# path based tables
spark.read.format("delta") \
.option("readChangeFeed", "true") \
.option("startingTimestamp", '2021-04-21 05:45:46') \
.load("pathToMyDeltaTable")
// version as ints or longs
spark.read.format("delta")
.option("readChangeFeed", "true")
.option("startingVersion", 0)
.option("endingVersion", 10)
.table("myDeltaTable")
// timestamps as formatted timestamp
spark.read.format("delta")
.option("readChangeFeed", "true")
.option("startingTimestamp", "2021-04-21 05:45:46")
.option("endingTimestamp", "2021-05-21 12:00:00")
.table("myDeltaTable")
// providing only the startingVersion/timestamp
spark.read.format("delta")
.option("readChangeFeed", "true")
.option("startingVersion", 0)
.table("myDeltaTable")
// path based tables
spark.read.format("delta")
.option("readChangeFeed", "true")
.option("startingTimestamp", "2021-04-21 05:45:46")
.load("pathToMyDeltaTable")
Read changes in streaming queries
# providing a starting version
spark.readStream.format("delta") \
.option("readChangeFeed", "true") \
.option("startingVersion", 0) \
.table("myDeltaTable")
# providing a starting timestamp
spark.readStream.format("delta") \
.option("readChangeFeed", "true") \
.option("startingTimestamp", "2021-04-21 05:35:43") \
.load("/pathToMyDeltaTable")
# not providing a starting version/timestamp will result in the latest snapshot being fetched first
spark.readStream.format("delta") \
.option("readChangeFeed", "true") \
.table("myDeltaTable")
// providing a starting version
spark.readStream.format("delta")
.option("readChangeFeed", "true")
.option("startingVersion", 0)
.table("myDeltaTable")
// providing a starting timestamp
spark.readStream.format("delta")
.option("readChangeFeed", "true")
.option("startingVersion", "2021-04-21 05:35:43")
.load("/pathToMyDeltaTable")
// not providing a starting version/timestamp will result in the latest snapshot being fetched first
spark.readStream.format("delta")
.option("readChangeFeed", "true")
.table("myDeltaTable")
To get the change data while reading the table, set the option readChangeFeed
to true
.
The startingVersion
or startingTimestamp
are optional and if not provided the stream returns the latest
snapshot of the table at the time of streaming as an INSERT
and future changes as change data.
Options like rate limits (maxFilesPerTrigger
, maxBytesPerTrigger
) and excludeRegex
are also supported when reading change data.
Note
Rate limiting can be atomic for versions other than the starting snapshot version. That is, the entire commit version will be rate limited or the entire commit will be returned.
By default if a user passes in a version or timestamp exceeding the last commit on a table, the error timestampGreaterThanLatestCommit
will be thrown. CDF can handle the out of range version case, if the user sets the following configuration to true
.
set spark.databricks.delta.changeDataFeed.timestampOutOfRange.enabled = true;
If you provide a start version greater than the last commit on a table or a start timestamp newer than the last commit on a table, then when the preceding configuration is enabled, an empty read result is returned.
If you provide an end version greater than the last commit on a table or an end timestamp newer than the last commit on a table, then when the preceding configuration is enabled in batch read mode, all changes between the start version and the last commit are be returned.
Change data event schema
In addition to the data columns, change data contains metadata columns that identify the type of change event:
Column name |
Type |
Values |
---|---|---|
|
String |
|
|
Long |
The Delta log or table version containing the change. |
|
Timestamp |
The timestamp associated when the commit was created. |
(1) preimage
is the value before the update, postimage
is the value after the update.
Frequently asked questions (FAQ)
What is the overhead of enabling the change data feed?
There is no significant impact. The change data records are generated in line during the query execution process, and are generally much smaller than the total size of rewritten files.
Known limitations
In Delta Lake 2.0.0, CDF reads for both batch and streaming are explicitly blocked on tables with column mapping enabled. In Delta Lake 2.1.0, CDF batch reads are supported on tables with column mapping enabled when DROP COLUMN
and RENAME COLUMN
have not been used.