SQL DDL
The following are SparkSQL table management actions available:
Spark Create Table
Only SparkSQL needs an explicit Create Table command. No Create Table command is required in Spark when using Scala or Python. The first batch of a Write to a table will create the table if it does not exist.
Options
Users can set table options while creating a hudi table.
Parameter Name | Description | (Optional/Required) : Default Value |
---|---|---|
primaryKey | The primary key names of the table, multiple fields separated by commas. | (Optional) : id |
type | The type of table to create (read more). cow = COPY-ON-WRITE, mor = MERGE-ON-READ. | (Optional) : cow |
preCombineField | The Pre-Combine field of the table. | (Optional) : ts |
To set any custom hudi config(like index type, max parquet size, etc), see the "Set hudi config section" .
Table Type
Here is an example of creating a COW table.
-- create a non-primary key table
create table if not exists hudi_table2(
id int,
name string,
price double
) using hudi
options (
type = 'cow'
);
Primary Key
Here is an example of creating COW table with a primary key 'id'.
-- create a managed cow table
create table if not exists hudi_table0 (
id int,
name string,
price double
) using hudi
options (
type = 'cow',
primaryKey = 'id'
);
PreCombineField
Here is an example of creating an MOR external table. The preCombineField option is used to specify the preCombine field for merge.
-- create an external mor table
create table if not exists hudi_table1 (
id int,
name string,
price double,
ts bigint
) using hudi
options (
type = 'mor',
primaryKey = 'id,name',
preCombineField = 'ts'
);
Partitioned Table
Here is an example of creating a COW partitioned table.
create table if not exists hudi_table_p0 (
id bigint,
name string,
dt string,
hh string
) using hudi
options (
type = 'cow',
primaryKey = 'id'
)
partitioned by (dt, hh);
Create Table for an External Hudi Table
You can create an External table using the location
statement. If an external location is not specified it is considered a managed table. You can read more about external vs managed tables here.
An external table is useful if you need to read/write to/from a pre-existing hudi table.
create table h_p1 using hudi
location '/path/to/hudi';
Create Table AS SELECT
Hudi supports CTAS(Create table as select) on spark sql.
Note: For better performance to load data to hudi table, CTAS uses bulk insert as the write operation.
Example CTAS command to create a non-partitioned COW table.
create table h3 using hudi
as
select 1 as id, 'a1' as name, 10 as price;
Example CTAS command to create a partitioned, primary key COW table.
create table h2 using hudi
options (type = 'cow', primaryKey = 'id')
partitioned by (dt)
as
select 1 as id, 'a1' as name, 10 as price, 1000 as dt;
Example CTAS command to load data from another table.
# create managed parquet table
create table parquet_mngd using parquet location 'file:///tmp/parquet_dataset/*.parquet';
# CTAS by loading data into hudi table
create table hudi_tbl using hudi location 'file:/tmp/hudi/hudi_tbl/' options (
type = 'cow',
primaryKey = 'id',
preCombineField = 'ts'
)
partitioned by (datestr) as select * from parquet_mngd;
Set hoodie config options
You can also set the config with table options when creating table which will work for the table scope only and override the config set by the SET command.
create table if not exists h3(
id bigint,
name string,
price double
) using hudi
options (
primaryKey = 'id',
type = 'mor',
${hoodie.config.key1} = '${hoodie.config.value2}',
${hoodie.config.key2} = '${hoodie.config.value2}',
....
);
e.g.
create table if not exists h3(
id bigint,
name string,
price double
) using hudi
options (
primaryKey = 'id',
type = 'mor',
hoodie.cleaner.fileversions.retained = '20',
hoodie.keep.max.commits = '20'
);
Spark Alter Table
Syntax
-- Alter table name
ALTER TABLE oldTableName RENAME TO newTableName
-- Alter table add columns
ALTER TABLE tableIdentifier ADD COLUMNS(colAndType (,colAndType)*)
-- Alter table column type
ALTER TABLE tableIdentifier CHANGE COLUMN colName colName colType
ALTER TABLE ... RENAME TO ...
is not supported when using AWS Glue Data Catalog as hive metastore as Glue itself does
not support table renames.
Examples
alter table h0 rename to h0_1;
alter table h0_1 add columns(ext0 string);
alter table h0_1 change column id id bigint;
Alter hoodie config options
You can also alter the write config for a table by the ALTER SERDEPROPERTIES
Example:
alter table h3 set serdeproperties (hoodie.keep.max.commits = '10')
Use set command
You can use the set command to set any custom hudi's config, which will work for the whole spark session scope.
set hoodie.insert.shuffle.parallelism = 100;
set hoodie.upsert.shuffle.parallelism = 100;
set hoodie.delete.shuffle.parallelism = 100;
Flink
Create Catalog
The catalog helps to manage the SQL tables, the table can be shared among CLI sessions if the catalog persists the table DDLs.
For hms
mode, the catalog also supplements the hive syncing options.
HMS mode catalog SQL demo:
CREATE CATALOG hoodie_catalog
WITH (
'type'='hudi',
'catalog.path' = '${catalog default root path}',
'hive.conf.dir' = '${directory where hive-site.xml is located}',
'mode'='hms' -- supports 'dfs' mode that uses the DFS backend for table DDLs persistence
);