public final class HoodieSparkQuickstart extends Object
| Modifier and Type | Method and Description |
|---|---|
static org.apache.spark.sql.Dataset<org.apache.spark.sql.Row> |
delete(org.apache.spark.sql.SparkSession spark,
String tablePath,
String tableName)
Delete data based in data information.
|
static void |
deleteByPartition(org.apache.spark.sql.SparkSession spark,
String tablePath,
String tableName)
Delete the data of the first partition.
|
static void |
incrementalQuery(org.apache.spark.sql.SparkSession spark,
String tablePath,
String tableName)
Hudi also provides capability to obtain a stream of records that changed since given commit timestamp.
|
static org.apache.spark.sql.Dataset<org.apache.spark.sql.Row> |
insertData(org.apache.spark.sql.SparkSession spark,
org.apache.spark.api.java.JavaSparkContext jsc,
String tablePath,
String tableName,
HoodieExampleDataGenerator<HoodieAvroPayload> dataGen)
Generate some new trips, load them into a DataFrame and write the DataFrame into the Hudi dataset as below.
|
static org.apache.spark.sql.Dataset<org.apache.spark.sql.Row> |
insertOverwriteData(org.apache.spark.sql.SparkSession spark,
org.apache.spark.api.java.JavaSparkContext jsc,
String tablePath,
String tableName,
HoodieExampleDataGenerator<HoodieAvroPayload> dataGen)
Generate new records, load them into a
Dataset and insert-overwrite it into the Hudi dataset |
static void |
main(String[] args) |
static void |
pointInTimeQuery(org.apache.spark.sql.SparkSession spark,
String tablePath,
String tableName)
Lets look at how to query data as of a specific time.
|
static void |
queryData(org.apache.spark.sql.SparkSession spark,
org.apache.spark.api.java.JavaSparkContext jsc,
String tablePath,
String tableName,
HoodieExampleDataGenerator<HoodieAvroPayload> dataGen)
Load the data files into a DataFrame.
|
static void |
runQuickstart(org.apache.spark.api.java.JavaSparkContext jsc,
org.apache.spark.sql.SparkSession spark,
String tableName,
String tablePath)
Visible for testing
|
static org.apache.spark.sql.Dataset<org.apache.spark.sql.Row> |
updateData(org.apache.spark.sql.SparkSession spark,
org.apache.spark.api.java.JavaSparkContext jsc,
String tablePath,
String tableName,
HoodieExampleDataGenerator<HoodieAvroPayload> dataGen)
This is similar to inserting new data.
|
public static void main(String[] args)
public static void runQuickstart(org.apache.spark.api.java.JavaSparkContext jsc,
org.apache.spark.sql.SparkSession spark,
String tableName,
String tablePath)
public static org.apache.spark.sql.Dataset<org.apache.spark.sql.Row> insertData(org.apache.spark.sql.SparkSession spark,
org.apache.spark.api.java.JavaSparkContext jsc,
String tablePath,
String tableName,
HoodieExampleDataGenerator<HoodieAvroPayload> dataGen)
public static org.apache.spark.sql.Dataset<org.apache.spark.sql.Row> insertOverwriteData(org.apache.spark.sql.SparkSession spark,
org.apache.spark.api.java.JavaSparkContext jsc,
String tablePath,
String tableName,
HoodieExampleDataGenerator<HoodieAvroPayload> dataGen)
Dataset and insert-overwrite it into the Hudi datasetpublic static void queryData(org.apache.spark.sql.SparkSession spark,
org.apache.spark.api.java.JavaSparkContext jsc,
String tablePath,
String tableName,
HoodieExampleDataGenerator<HoodieAvroPayload> dataGen)
public static org.apache.spark.sql.Dataset<org.apache.spark.sql.Row> updateData(org.apache.spark.sql.SparkSession spark,
org.apache.spark.api.java.JavaSparkContext jsc,
String tablePath,
String tableName,
HoodieExampleDataGenerator<HoodieAvroPayload> dataGen)
public static org.apache.spark.sql.Dataset<org.apache.spark.sql.Row> delete(org.apache.spark.sql.SparkSession spark,
String tablePath,
String tableName)
public static void deleteByPartition(org.apache.spark.sql.SparkSession spark,
String tablePath,
String tableName)
public static void incrementalQuery(org.apache.spark.sql.SparkSession spark,
String tablePath,
String tableName)
public static void pointInTimeQuery(org.apache.spark.sql.SparkSession spark,
String tablePath,
String tableName)
Copyright © 2025 The Apache Software Foundation. All rights reserved.