Apache XTable™ (Incubating) is a cross-table converter for table formats that facilitates omni-directional interoperability across data processing systems and query engines. Currently, Apache XTable™ supports widely adopted open-source table formats such as Apache Hudi, Apache Iceberg, and Delta Lake.
Apache XTable™ simplifies data lake operations by leveraging a common model for table representation. This allows users to write data in one format while still benefiting from integrations and features available in other formats. For instance, Apache XTable™ enables existing Hudi users to seamlessly work with Databricks's Photon Engine or query Iceberg tables with Snowflake. Creating transformations from one format to another is straightforward and only requires the implementation of a few interfaces, which we believe will facilitate the expansion of supported source and target formats in the future.
- Use Java11 for building the project. If you are using some other java version, you can use jenv to use multiple java versions locally.
- Build the project using
mvn clean package
. Usemvn clean package -DskipTests
to skip tests while building. - Use
mvn clean test
ormvn test
to run all unit tests. If you need to run only a specific test you can do this by something likemvn test -Dtest=TestDeltaSync -pl core
. - Similarly, use
mvn clean verify
ormvn verify
to run integration tests.
- We use Maven Spotless plugin and Google java format for code style.
- Use
mvn spotless:check
to find out code style violations andmvn spotless:apply
to fix them. Code style check is tied to compile phase by default, so code style violations will lead to build failures.
- Get a pre-built bundled jar or create the jar with
mvn install -DskipTests
- create a yaml file that follows the format below:
sourceFormat: HUDI
targetFormats:
- DELTA
- ICEBERG
datasets:
-
tableBasePath: s3://tpc-ds-datasets/1GB/hudi/call_center
tableDataPath: s3://tpc-ds-datasets/1GB/hudi/call_center/data
tableName: call_center
namespace: my.db
-
tableBasePath: s3://tpc-ds-datasets/1GB/hudi/catalog_sales
tableName: catalog_sales
partitionSpec: cs_sold_date_sk:VALUE
-
tableBasePath: s3://hudi/multi-partition-dataset
tableName: multi_partition_dataset
partitionSpec: time_millis:DAY:yyyy-MM-dd,type:VALUE
-
tableBasePath: abfs://container@storage.dfs.core.windows.net/multi-partition-dataset
tableName: multi_partition_dataset
sourceFormat
is the format of the source table that you want to converttargetFormats
is a list of formats you want to create from your source tablestableBasePath
is the basePath of the tabletableDataPath
is an optional field specifying the path to the data files. If not specified, the tableBasePath will be used. For Iceberg source tables, you will need to specify the/data
path.namespace
is an optional field specifying the namespace of the table and will be used when syncing to a catalog.partitionSpec
is a spec that allows us to infer partition values. This is only required for Hudi source tables. If the table is not partitioned, leave it blank. If it is partitioned, you can specify a spec with a comma separated list with formatpath:type:format
path
is a dot separated path to the partition fieldtype
describes how the partition value was generated from the column valueVALUE
: an identity transform of field value to partition valueYEAR
: data is partitioned by a field representing a date and year granularity is usedMONTH
: same asYEAR
but with month granularityDAY
: same asYEAR
but with day granularityHOUR
: same asYEAR
but with hour granularity
format
: if your partition type isYEAR
,MONTH
,DAY
, orHOUR
specify the format for the date string as it appears in your file paths
- The default implementations of table format converters can be replaced with custom implementations by specifying a converter configs yaml file in the format below:
# conversionSourceProviderClass: The class name of a table format's converter factory, where the converter is
# used for reading from a table of this format. All user configurations, including hadoop config
# and converter specific configuration, will be available to the factory for instantiation of the
# converter.
# conversionTargetProviderClass: The class name of a table format's converter factory, where the converter is
# used for writing to a table of this format.
# configuration: A map of configuration values specific to this converter.
tableFormatConverters:
HUDI:
conversionSourceProviderClass: org.apache.xtable.hudi.HudiConversionSourceProvider
DELTA:
conversionTargetProviderClass: org.apache.xtable.delta.DeltaConversionTarget
configuration:
spark.master: local[2]
spark.app.name: xtable
- A catalog can be used when reading and updating Iceberg tables. The catalog can be specified in a yaml file and passed in with the
--icebergCatalogConfig
option. The format of the catalog config file is:
catalogImpl: io.my.CatalogImpl
catalogName: name
catalogOptions: # all other options are passed through in a map
key1: value1
key2: value2
- run with
java -jar xtable-utilities/target/xtable-utilities-0.1.0-SNAPSHOT-bundled.jar --datasetConfig my_config.yaml [--hadoopConfig hdfs-site.xml] [--convertersConfig converters.yaml] [--icebergCatalogConfig catalog.yaml]
The bundled jar includes hadoop dependencies for AWS, Azure, and GCP. Sample hadoop configurations for configuring the converters can be found in the xtable-hadoop-defaults.xml file. The custom hadoop configurations can be passed in with the--hadoopConfig [custom-hadoop-config-file]
option. The config in custom hadoop config file will override the default hadoop configurations. For an example of a custom hadoop config file, see hadoop.xml.
For setting up the repo on IntelliJ, open the project and change the java version to Java11 in File->ProjectStructure
You have found a bug, or have a cool idea you that want to contribute to the project ? Please file a GitHub issue here
Adding a new target format requires a developer implement ConversionTarget. Once you have implemented that interface, you can integrate it into the ConversionController. If you think others may find that target useful, please raise a Pull Request to add it to the project.