Export DuckLake catalog tables to Delta Lake format
Maintainer(s):
djouallah
Installing and Loading
INSTALL delta_export FROM community;
LOAD delta_export;
Example
-- Load required extensions
LOAD ducklake;
LOAD delta_export;
-- Attach a DuckLake catalog
ATTACH 'ducklake:metadata.ducklake' AS lake (DATA_PATH '/path/to/data');
USE lake;
-- Create and populate a table
CREATE TABLE orders (id INTEGER, amount DOUBLE);
INSERT INTO orders VALUES (1, 99.99), (2, 149.50);
-- Export all DuckLake tables to Delta Lake format
CALL delta_export();
About delta_export
Delta Export reads DuckLake internal metadata and writes Delta Lake checkpoint files (Parquet, JSON transaction log, and _last_checkpoint) for each table in the catalog.
This enables interoperability between DuckLake-managed data and any system that reads Delta Lake format, such as Apache Spark, Databricks, Microsoft Fabric, and other Delta-compatible engines.
Works with both DuckDB and SQLite-backed DuckLake catalogs, and supports all storage backends including local filesystem and ABFSS.
Added Functions
| function_name | function_type | description | comment | examples |
|---|---|---|---|---|
| delta_export | table | NULL | NULL |