Skip to main content

External Table

Component that constructs and updates an External Table. Currently supported for Snowflake only.

Examples

component:
external_table:
location: "@my_namespace.my_ext_stage/path"
file_format: "my_file_format"
pattern: ".*[.]csv"
auto_refresh: true
partitions:
- name: partition_column
data_type: STRING
expression: EXTRACT(YEAR FROM my_date_column)
aws_sns_topic: "arn:aws:sns:us-west-2:123456789012:my_sns_topic"
integration: "my_integration"

ExternalTableComponent

info

ExternalTableComponent is defined beneath the following ancestor nodes in the YAML structure:

Below are the properties for the ExternalTableComponent. Each property links to the specific details section further down in this page.

PropertyDefaultTypeRequiredDescription
data_plane  One of:
    SnowflakeDataPlane
    BigQueryDataPlane
    DuckdbDataPlane
    DatabricksDataPlane
NoData Plane-specific configuration options for Components.
skipboolean
NoBoolean flag indicating whether to skip processing for the Component or not.
retry_strategyNoRetry strategy configuration options for the Component if any exceptions are encountered.
data_maintenanceNoThe data maintenance configuration options for the Component.
descriptionstring
NoBrief description of what the model does.
metadataNoMeta information of a resource. In most cases it doesn't affect the system behavior but may be helpful to analyze project resources.
namestringYesThe name of the model
flow_namestring
NoName of the Flow that the Component belongs to.
testsNoDefines tests to run on this Component's data.
external_tableAny of:
YesConfiguration options for the External Table Component.

Property Details

Component

A Component is a fundamental building block of a data Flow. Supported Component types include: Read, Transform, Task, Test, and more.

PropertyDefaultTypeRequiredDescription
componentOne of:
  CustomPythonReadComponent
  ApplicationComponent
  AliasedTableComponent
  ExternalTableComponent
YesComponent configuration options.

BigQueryExternalTableOptions

Configuration options for the an External Table Component in BigQuery. Currently not implemented, just stubbed for future reference.

PropertyDefaultTypeRequiredDescription
dependenciesarray[None]
NoList of dependencies that must complete before this Component runs.
event_timestring
NoTimestamp column in the Component output used to represent Event time.
locationstring
No

SnowflakeExternalTableOptions

Configuration options for the an External Table Component in Snowflake.

PropertyDefaultTypeRequiredDescription
dependenciesarray[None]
NoList of dependencies that must complete before this Component runs.
event_timestring
NoTimestamp column in the Component output used to represent Event time.
locationstring
NoSnowflake Stage containing the data files, in the format @[namespace.]ext_stage_name[/path].
file_formatstring
NoThe file format configuration. See the Snowflake documentation at https://docs.snowflake.com/en/sql-reference/sql/create-external-table#required-parameters for more information.
patternstring
NoRegEx pattern to match files in the stage.
auto_refreshboolean
NoDetermines whether Snowflake should auto-refresh the table.
partitionsarray[None]
NoList of virtual columns to compute and partition the table by.
columnsarray[None]
NoList of virtual columns to compute.
aws_sns_topicstring
NoAmazon Resource Name (ARN) for the SNS topic for your S3 bucket (required for auto-refreshing tables from S3 using SNS).
integrationstring
NoName of the notification integration (required for auto-refreshing tables from GCS or Azure Blob Store).

SnowflakeVirtualColumnSpec

Configuration options for the Custom Python Read Component.

PropertyDefaultTypeRequiredDescription
namestringYesName of the virtual column.
data_typestringYesData type of the virtual column.
descriptionstring
NoDescription of the virtual column.
expressionstring
NoSQL expression that computes the value for the virtual column.

BigQueryDataPlane

PropertyDefaultTypeRequiredDescription
bigqueryYesBigQuery configuration options.

BigQueryDataPlaneOptions

PropertyDefaultTypeRequiredDescription
partition_byAny of:
NoPartition By clause for the table.
cluster_byarray[string]
NoClustering keys to be added to the table.

BigQueryRangePartitioning

PropertyDefaultTypeRequiredDescription
fieldstringYesField to partition by.
rangeYesRange partitioning options.

BigQueryTimePartitioning

PropertyDefaultTypeRequiredDescription
fieldstringYesField to partition by.
granularitystring ("DAY", "HOUR", "MONTH", "YEAR")YesGranularity of the time partitioning.

ComponentTestOptions

Options for component tests, including data quality tests and schema checks.

PropertyDefaultTypeRequiredDescription
columnsobject with property values of type array[One of: (Any of: (not_null, NotNullTest), Any of: (not_empty, NotEmptyTest), Any of: (unique, UniqueTest), CombinationUniqueTest, InRangeTest, DateInRangeTest, InSetTest, SubstringMatchTest, CountDistinctEqualTest, CountGreaterThanOrEqualTest, CountGreaterThanTest, CountLessThanOrEqualTest, CountLessThanTest, CountEqualTest, GreaterThanTest, LessThanTest, GreaterThanOrEqualTest, LessThanOrEqualTest, MeanInRangeTest, StddevInRangeTest, ColumnTestSql, ColumnTestPython)]
NoList of column-level data quality tests for a Component.
componentarray[One of: (Any of: (not_null, NotNullTest), Any of: (not_empty, NotEmptyTest), Any of: (unique, UniqueTest), CombinationUniqueTest, InRangeTest, DateInRangeTest, InSetTest, SubstringMatchTest, CountDistinctEqualTest, CountGreaterThanOrEqualTest, CountGreaterThanTest, CountLessThanOrEqualTest, CountLessThanTest, CountEqualTest, GreaterThanTest, LessThanTest, GreaterThanOrEqualTest, LessThanOrEqualTest, MeanInRangeTest, StddevInRangeTest, ColumnTestSql, ColumnTestPython)]
NoList of Component-level tests.
schemaNoList of schema checks for a Component.

ColumnTestPython

Test to validate data using a Python function for a single column.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
namestringYes
pythonYesConfiguration options for the Python column test.

ColumnTestPythonOptions

PropertyDefaultTypeRequiredDescription
entrypointstringYesEntry point for the Python Transform function.
sourcestringYesSource file for the Python Transform function.
paramsobject with property values of type None
NoParameters for the Python test function.
is_asset_testboolean
No

ColumnTestSql

Test to validate data using an SQL query for a single column.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
namestringYes
sqlstring
NoSQL query that tests data for conditions.

CombinationUniqueTest

Test to check if a value is unique.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
combination_uniqueYesTest to check if a value is unique.

CombinationUniqueTestOptions

Configuration options for the unique test.

PropertyDefaultTypeRequiredDescription
columnsarray[string]YesCombination of columns to check for uniqueness.

ComponentSchemaTest

Test to validate that component columns match expected types.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
matchexactstring ("exact", "ignore_missing")NoThe type of schema matching to perform. 'exact' requires all columns to be present, 'ignore_missing' allows for missing columns.
columnsobject with property values of type string
NoMapping of column names to their expected types.

CountDistinctEqualTest

Test to check if the number of distinct values is equal to a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
count_distinct_equalYes

CountDistinctEqualTestOptions

Configuration options for the count_distinct_equal test.

PropertyDefaultTypeRequiredDescription
countintegerYesNumber of distinct values to expect.
group_by_columnsarray[string]
NoColumns to group by.

CountEqualTest

Test to check if the number of rows is equal to a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
count_equalYesConfiguration options for the the count_equal test.

CountEqualTestOptions

Configuration options for the count_equal test.

PropertyDefaultTypeRequiredDescription
countintegerYesNumber of rows to expect.

CountGreaterThanOrEqualTest

Test to check if the number of rows is greater than or equal to a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
count_greater_than_or_equalYes

CountGreaterThanOrEqualTestOptions

Configuration options for the count_greater_than_or_equal test.

PropertyDefaultTypeRequiredDescription
countintegerYesValue to compare against.
group_by_columnsarray[string]
NoColumns to group by.

CountGreaterThanTest

Test to check if the number of rows is greater than a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
count_greater_thanYes

CountGreaterThanTestOptions

Configuration options for the count_greater_than test.

PropertyDefaultTypeRequiredDescription
countintegerYesValue to compare against.
group_by_columnsarray[string]
NoColumns to group by.

CountLessThanOrEqualTest

Test to check if the number of rows is greater than or equal to a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
count_less_than_or_equalYes

CountLessThanOrEqualTestOptions

Configuration options for the count_less_than_or_equal test.

PropertyDefaultTypeRequiredDescription
countintegerYesValue to compare against.
group_by_columnsarray[string]
NoColumns to group by.

CountLessThanTest

Test to check if the number of rows is less than a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
count_less_thanYes

CountLessThanTestOptions

Configuration options for the count_less_than test.

PropertyDefaultTypeRequiredDescription
countintegerYesValue to compare against.
group_by_columnsarray[string]
NoColumns to group by.

DatabricksDataPlane

PropertyDefaultTypeRequiredDescription
databrickscluster_by: null
pyspark_job_cluster_id: null
table_properties: null
NoDatabricks configuration options.

DatabricksDataPlaneOptions

PropertyDefaultTypeRequiredDescription
table_propertiesobject with property values of type string
NoTable properties to include when creating the data table. This setting is equivalent to the CREATE TABLE ... TBLPROPERTIES clause. Please refer to the Databricks documentation at https://docs.databricks.com/aws/en/delta/table-properties for available properties depending on your Data Plane.
pyspark_job_cluster_idstring
NoID of the compute cluster to use for PySpark jobs.
cluster_byarray[string]
NoClustering keys to be added to the table.

DateInRangeTest

Test to check if a date is within a certain range.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
date_in_rangeYes

DateInRangeTestOptions

Configuration options for the date_in_range test.

PropertyDefaultTypeRequiredDescription
minstringYesMinimum value to expect.
maxstringYesMaximum value to expect.

DuckdbDataPlane

PropertyDefaultTypeRequiredDescription
duckdbducklake_data_table_compaction:
  small_file_count_threshold: 50
  small_file_ratio_threshold: 0.25
  small_file_record_count_limit: 100000
ducklake_metadata_table_compaction:
  small_file_count_threshold: 10
  small_file_ratio_threshold: null
  small_file_record_count_limit: 10
NoDuckDB configuration options.

DuckDbDataPlaneOptions

PropertyDefaultTypeRequiredDescription
ducklake_metadata_table_compactionsmall_file_count_threshold: 10
small_file_ratio_threshold: null
small_file_record_count_limit: 10
DuckLakeTableCompactionSettingsNoSettings for compacting metadata tables. If present, metadata table compaction is enabled.
ducklake_data_table_compactionsmall_file_count_threshold: 50
small_file_ratio_threshold: 0.25
small_file_record_count_limit: 100000
DuckLakeTableCompactionSettingsNoSettings for compacting data tables. If present, data table compaction is enabled.

DuckLakeTableCompactionSettings

Settings for DuckLake table compaction.

PropertyDefaultTypeRequiredDescription
small_file_record_count_limit10integerNoFiles with fewer records than this limit are considered 'small'.
small_file_count_threshold10integerNoRun manual table compaction if the number of files with fewer than small_file_record_count_limit records exceeds this threshold.
small_file_ratio_thresholdnumber
NoPercentage (0.0-1.0) of small files relative to total files. If set, both absolute count AND ratio must pass for compaction to be triggered. If None, only absolute count check is performed.

FabricDataPlane

PropertyDefaultTypeRequiredDescription
fabricspark_session_config: null
NoFabric configuration options.

FabricDataPlaneOptions

PropertyDefaultTypeRequiredDescription
spark_session_configNoSpark session configuration.

GreaterThanOrEqualTest

Test to check if a value is greater than or equal to a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
greater_than_or_equalYes

GreaterThanOrEqualTestOptions

Configuration options for the greater_than_or_equal test.

PropertyDefaultTypeRequiredDescription
valueAny of:
  integer
  number
  string
YesValue to compare against.

GreaterThanTest

Test to check if a value is greater than a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
greater_thanYes

GreaterThanTestOptions

Configuration options for the greater_than test.

PropertyDefaultTypeRequiredDescription
valueAny of:
  integer
  number
  string
YesValue to compare against.

InRangeTest

Test to check if a value is within a certain range.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
in_rangeYes

InRangeTestOptions

Configuration options for the in_range test.

PropertyDefaultTypeRequiredDescription
minAny of:
  integer
  number
  string
YesMinimum value to expect.
maxAny of:
  integer
  number
  string
YesMaximum value to expect.

InSetTest

Test to check if a value is in a set of values.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
in_setYes

InSetTestOptions

Configuration options for the in_set test.

PropertyDefaultTypeRequiredDescription
valuesarray[Any of: (integer, number, string)]YesSet of values to expect.

LessThanOrEqualTest

Test to check if a value is less than or equal to a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
less_than_or_equalYes

LessThanOrEqualTestOptions

Configuration options for the less_than_or_equal test.

PropertyDefaultTypeRequiredDescription
valueAny of:
  integer
  number
  string
YesValue to compare against.

LessThanTest

Test to check if a value is less than a certain number.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
less_thanYes

LessThanTestOptions

Configuration options for the less_than test.

PropertyDefaultTypeRequiredDescription
valueAny of:
  integer
  number
  string
YesValue to compare against.

MeanInRangeTest

Test to check if a value is within a certain mean range.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
mean_in_rangeYes

MeanInRangeTestOptions

Configuration options for the mean_in_range test.

PropertyDefaultTypeRequiredDescription
minAny of:
  integer
  number
  string
YesMinimum value to expect.
maxAny of:
  integer
  number
  string
YesMaximum value to expect.

NotEmptyTest

Test to check if a value is not empty.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
not_emptyNoTest to check if a value is not empty.

NotNullTest

Test to check if a value is not null.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
not_nullNoTest to check if a value is not null.

RangeOptions

PropertyDefaultTypeRequiredDescription
startintegerYesStart of the range partitioning.
endintegerYesEnd of the range partitioning.
intervalintegerYesInterval of the range partitioning.

SnowflakeDataPlane

PropertyDefaultTypeRequiredDescription
snowflakeYesSnowflake configuration options.

SnowflakeDataPlaneOptions

PropertyDefaultTypeRequiredDescription
cluster_byarray[string]
NoClustering keys to be added to the table.

StddevInRangeTest

Test to check if a value is within a certain standard deviation range.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
stddev_in_rangeYes

StddevInRangeTestOptions

Configuration options for the stddev_in_range test.

PropertyDefaultTypeRequiredDescription
minAny of:
  integer
  number
  string
YesMinimum value to expect.
maxAny of:
  integer
  number
  string
YesMaximum value to expect.

SubstringMatchTest

Test to check if a value contains a substring.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
substring_matchYes

SubstringMatchTestOptions

Configuration options for the substring_match test.

PropertyDefaultTypeRequiredDescription
substringstringYesSubstring to search for.

SynapseDataPlane

PropertyDefaultTypeRequiredDescription
synapsespark_session_config: null
NoSynapse configuration options.

SynapseDataPlaneOptions

PropertyDefaultTypeRequiredDescription
spark_session_configNoSpark session configuration.

LivySparkSessionConfig

PropertyDefaultTypeRequiredDescription
poolstring
NoPool to use for the Spark session.
driver_memorystring
NoMemory to use for the Spark driver.
driver_coresinteger
NoNumber of cores to use for the Spark driver.
executor_memorystring
NoMemory to use for the Spark executor.
executor_coresinteger
NoNumber of cores to use for each executor.
num_executorsinteger
NoNumber of executors to use for the Spark session.
session_key_overridestring
NoKey to use for the Spark session.
max_concurrent_sessionsinteger
NoMaximum number of concurrent sessions allowed for this configuration.

UniqueTest

Test to check if a value is unique.

PropertyDefaultTypeRequiredDescription
severityerrorstring ("error", "warn")NoThe severity level for issues raised by the test. Default is 'error'. Use 'error' for critical issues that should interrupt flow processing. Use 'warn' for warnings/minor issues that should not interrupt flow processing.
uniqueNoTest to check if a value is unique.

NoTestOptions

Configuration options for tests that have no test body definition (not_null, unique, etc.).

No properties defined.

DataMaintenance

Data maintenance configuration options for Components.

PropertyDefaultTypeRequiredDescription
enabledboolean
NoBoolean flag indicating whether data maintenance is enabled for the Component.
manual_table_compactionboolean
NoBoolean flag indicating whether manual table compaction is enabled for the Component. This is currently only relevant for DuckLake Data Planes.
manual_table_compaction_record_count_threshold10integer
NoConsider files with fewer than this number of records in determining whether to perform manual table compaction. This is currently only relevant for DuckLake Data Planes.
manual_table_compaction_file_count_threshold10integer
NoRun manual table compaction if the number of files with fewer than manual_table_compaction_record_count_threshold records exceeds this threshold. This is currently only relevant for DuckLake Data Planes.

ResourceMetadata

Meta information of a resource. In most cases, it doesn't affect the system behavior but may be helpful to analyze Project resources.

PropertyDefaultTypeRequiredDescription
sourceNoThe origin or source information for the resource.
source_event_uuidstring
NoEvent UUID associated with creation of this resource.

ResourceLocation

The origin or source information for the resource.

PropertyDefaultTypeRequiredDescription
pathstringYesPath within repository files where the resource is defined.
first_line_numberinteger
NoFirst line number within path file where the resource is defined.

RetryStrategy

Retry strategy configuration for Component operations. This configuration leverages the tenacity library to implement robust retry mechanisms. The configuration options directly map to tenacity's retry parameters. Details on the tenacity library can be found here: https://tenacity.readthedocs.io/en/latest/api.html#retry-main-api Current implementation includes: - stop_after_attempt: Maximum number of retry attempts - stop_after_delay: Give up on retries one attempt before you would exceed the delay. Will need to supply at least one of the two parameters. Additional retry parameters will be added as needed to support more complex use cases.

PropertyDefaultTypeRequiredDescription
stop_after_attemptinteger
NoNumber of retry attempts before giving up. If set to None, it will not stop after any number of attempts.
stop_after_delayinteger
NoMaximum time (in seconds) to spend on retries before giving up. If set to None, it will not stop after any time delay.

InputComponent

Specification for input Components defining how partitioning behaviors should be handled. This metadata is required when a Component serves as an input to other Components within a Flow. The reshape parameter controls how input data is partitioned and processed. It accepts either full for full reduction operations or map for partition-wise operations.

PropertyDefaultTypeRequiredDescription
flowstringYesName of the parent Flow that the input Component belongs to.
namestringYesName of the input Component.
aliasstring
NoAlias to use for the input Component.
partition_specAny of:
  string ("full_reduction", "map")
NoInternal specification for how Component input data should be partitioned before processing. This field is populated based on the user-facing reshape parameter in ref() calls, which accepts full (for full reduction operations) or map (for partition-wise operations). Input partitioning is applied before the Component's logic is executed.
wherestring
NoOptional filter condition to apply to the input Component's data.
partition_bindingAny of:
  string
NoOptional partition binding specification to apply to the Component on a per-output-partition basis against other inputs' partitions.

PartitionBinding

PropertyDefaultTypeRequiredDescription
logical_operatorlogical_operatorstring ("AND", "OR")NoTLogical operator to use to combine the partition binding predicates provided
predicatespredicatesarray[string]NoList of partition binding predicates to apply to the input Component's data

RepartitionSpec

Specification for repartitioning operations on input Component's data

PropertyDefaultTypeRequiredDescription
repartitionNoOptions for repartitioning the input Component's data.

RepartitionOptions

Options for repartitioning the input Component's data.

PropertyDefaultTypeRequiredDescription
partition_bystringYesColumn to partition by.
granularitystringYesGranularity to use for the partitioning.