Skip to main content

Streams And Tasks

Overview

Snowflake streams and tasks out of the box.

Installation

  • Copy the Package ID @coalesce/streams-and-tasks.
  • In Coalesce, open the Workspace where you wish to install the package.
  • Go to the Build Setting of the Workspace, tab Packages, and click the Install button on the top right of the page.
  • Paste the Package ID, and proceed with the installation process.

Description

Streams and Task Nodes

The Coalesce Stream and Task Node Types Package includes:

Work With Task

The Coalesce Work with Task UDN is a work node that wraps the standard Coalesce Work node with a Task.

Tasks can be combined with Coalesce Stream node (table streams) for continuous ELT workflows to process recently changed table rows. Streams ensure exactly once semantics for new or changed data in a table.

Tasks can also be used independently to generate periodic reports by inserting or merging rows into a report table or performing other periodic work.

More information about Tasks can be found in Snowflake's Introduction to tasks.

Work With Task Node Configuration

The Work with Task node has two or three configuration groups depending on config options selected.

With the option Development Mode set to true there are three config groups:

Work_with_dev_true

With the Development Mode set to false there are four config groups:

Work_with_dev_false

Work With Task Node Properties

  • Storage Location: Storage Location where the Stream will be created.
  • Node Type: Name of template used to create node objects.
  • Description: A description of the node's purpose.
  • Deploy Enabled:
    • If TRUE the node will be deployed / redeployed when changes are detected.
    • If FALSE the node will not be deployed or will be dropped during redeployment.

Work With Task Options

Worktask-opt

  • Development Mode: True / False toggle that determines whether a task will be created or if the SQL to be used in the task will execute as DML as a Run action. Prior to creating a task, it is helpful to test the SQL the task will execute to make sure it runs without errors and returns the expected data.
    • True - A table will be created and SQL will execute as a Run action.
    • False - After sufficiently testing the SQL as a Run action setting Development Mode to false will wrap the SQL statement in a task with options specified in Scheduling Options.
  • Multi Source: True / False toggle that is Coalesce implementation of SQL UNIONs.
    • True - Multiple sources can be combined in a single node. The sources are combined using the option specified in the Multi Source Strategy.
      • UNION - Combines with duplicate elimination.
      • UNION ALL - Combines without duplicate elimination.
    • False - Single source node or multiple sources combined using a join.
  • Cluster key: True/False to determine whether Dimension is to be clustered or not.
    • True - Allows you to specify the column based on which clustering is to be done.
      • Allow Expressions Cluster Key: True/False. Add an expression to the specified cluster key.
    • False – No clustering done
  • Truncate Before: Specifies that the target table should be truncated before inserting the values into the table.

Work With Task General Options

Work with Task-GO

  • DISTINCT: True/False toggle that determines whether to add DISTINCT to SQL Query.
    • True - Group by All is invisible. DISTINCT data is chosen for processing.
    • False- Group by All is visible.
  • GROUP BY ALL: True/False toggle that determines whether to add GROUP BY ALL to SQL Query.
    • True - DISTINCT is invisible. Data is grouped by all columns for processing.
    • False- DISTINCT is visible.
  • Order By : True/False toggle that determines whether to add “ORDER BY” to SQL query along with the column and sort order.
    • True -Sort column and sort order drop down are visible and are required to form order by clause.
    • False-Sort column and sort order drop down are invisible.

Work With Task Scheduling Options

If development mode is set to false then Scheduling Options can be used to configure how and when the task will run.

  • Scheduling Mode: Specifies whether a warehouse or serverless compute is used to run the task
    • Warehouse Task - User managed warehouse will execute task.
    • Serverless Task - Utilize serverless compute to execute task.
  • When Source Stream has Data Flag: True / False toggle that checks whether source streams have data before executing a task.
    • True - Only run task if source stream has capture change data
    • False - Run task on schedule regardless of whether the source stream has data. If the source is not a stream should set this to false.
  • Multiple Stream has Data Logic: True / False toggle is enabled if Source Stream has Data Flag is set to true.
    • AND - If there are multiple streams task will run if all streams have data
    • OR - If there are multiple streams task will run if one or more streams has data
  • Select Warehouse on which to run task: Visible if Scheduling Mode is set to Warehouse Task.
    • Enter the name of the warehouse you want the task to run on without quotes
  • Select initial serverless Warehouse size: Visible when Scheduling Mode is set to Serverless Task.
    • Select the initial compute size on which to run the task. Snowflake will adjust size from there based on target schedule and task run times.
  • Task Schedule: Select how you want to schedule the task to run
    • Minutes - Allows you to specify a minute interval for running task
    • Cron - Allows you to specify a CRON schedule for running task
    • Predecessor - Allows you to specify a predecessor task to determine when a task should execute
  • Enter task schedule using minutes: Only visible when Task Schedule is set to Minutes. Enter a whole number from 1 to 11520 which represents the number of minutes between task runs.
  • Enter task schedule using Cron: Only visible when Task Schedule is set to Cron. Specifies a cron expression and time zone for periodically running the task. Supports a subset of standard cron utility syntax.
  • Enter predecessor tasks separated by a comma: Only visible when Task Schedule is set to Predecessor. One or more task names that precede the task being created in the current node. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor task separate the task names using a comma and no spaces.
  • Enter root task name: Name of the root task that controls scheduling for the DAG of tasks. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor task separate the task names using a comma and no spaces.

Example of Serverless Task with Multiple Predecessors and Root Task

Example_of_Serverless_Task

Example of Warehouse Task With 60 Minute Task Schedule

Example_of_Warehouse_Task

Example of Warehouse Task With Cron Schedule Not Using a Stream

Example_of_Warehouse_Task_with_Cron

Work With Task Deployment

Work With Task Deployment Parameters

The Stage with Task includes an environment parameter that allows you to specify a different warehouse used to run a task in different environments.

The parameter name is targetTaskWarehouse and the default value is DEV ENVIRONMENT.

When set to DEV ENVIRONMENT the value entered in the Scheduling Options config Select Warehouse on which to run the task will be used when creating the task.

{
"targetTaskWarehouse": "DEV ENVIRONMENT"
}

When set to any value other than DEV ENVIRONMENT the node will attempt to create the task using a Snowflake warehouse with the specified value.

For example, with the below setting for the parameter in a QA environment, the task will execute using a warehouse named compute_wh.

{
"targetTaskWarehouse": "compute_wh"
}

Work With Task Initial Deployment

When deployed for the first time into an environment the Work with Task node will execute three stages dependent on whether or not the task schedule relies on a predecessor task:

Work With Task No Predecessor Task Deployment

  • Create Target Table: This stage will create a table that will be loaded by the ask

  • Create Task: This stage will create a task that will load the target table on the schedule specified

  • Resume Task: After the task has been created it needs to be resume so that the task runs on the schedule

Work With Task Predecessor Task Deployment

  • Create Target Table: This stage will create a table that will be loaded by the ask

  • Suspend Root Task: To add a task into a DAG of task the root task needs to be put into a suspended state

  • Create Task: This stage will create a task that will load the target table on the schedule specified

If a task is part of a DAG of tasks, the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

The task node has no ALTER capabilities. All task enabled nodes are CREATE OR REPLACE only though this is subject to change.

Work With Task Redeployment

Work With Task Recreating the Task Redeployment

After the Task has deployed for the first time into a target environment, subsequent deployments with changes in task schedule, warehouse, or scheduling options will result in a CREATE TASK AND RESUME TASK statements being issued

The following stages are executed:

Work With Task No Predecessor Task Redeployment

  • Create Task: This stage will create a task that will load the target table on the schedule specified.

  • Resume Task: After the task has been created it needs to be resume so that the task runs on the schedule.

Work With Task Predecessor Task Redeployment

  • Suspend Root Task: To add a task into a DAG of task the root task needs to be put into a suspended state.

  • Create Task: This stage will create a task that will load the target table on the schedule specified

Work With Task Altering the Tables

After the Task has deployed for the first time into a target environment, subsequent deployments with changes in table like add/drop column and change in data type will result in an ALTER table statement followed by CREATE TASK AND RESUME TASK statements being issued.

The following stages are executed:

  • Change Column Attributes/Delete Column/Add Column/Change table description: Alter table statement is executed to perform the alter operation accordingly.
  • Create Task
  • Resume Task

Work With Task Undeployment

If a Work with Task node is deleted from a Workspace, that Workspace is committed to Git and that commit deployed to a higher level environment then all objects created by the node in the target environment will be dropped.

Work With Task No Predecessor Task Undeployment

  • Drop Table: This stage will drop the table originally created to be loaded by the task.

  • Drop Current Task: This stage will drop the task.

Work With Task Predecessor Task Undeployment

  • Drop Table: This stage will drop the table originally created to be loaded by the task

  • Suspend Root Task: To drop a task from a DAG of task the root task needs to be put into a suspended state.

  • Drop Task: This stage will drop the task.

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

Dimension With Task

The Coalesce Dimension with Task UDN is a node that wraps the standard Coalesce Dimension node with a Task.

Tasks can be combined with Coalesce Stream node (table streams) for continuous ELT workflows to process recently changed table rows. Streams ensure exactly once semantics for new or changed data in a table.

Tasks can also be used independently to generate periodic reports by inserting or merging rows into a report table or performing other periodic work.

More information about Tasks can be found in Snowflake's Introduction to tasks.

Dimension With Task Node Configuration

The Dimension with Task node has two or three configuration groups depending on config options selected.

With the option Development Mode set to true there are three config groups:

Work_with_dev_true

With the Development Mode set to false there are four config groups:

Work_with_dev_false

Dimension With Task Node Properties

  • Storage Location: Storage Location where the Stream will be created.
  • Node Type: Name of template used to create node objects.
  • Description: A description of the node's purpose.
  • Deploy Enabled:
    • If TRUE the node will be deployed / redeployed when changes are detected.
    • If FALSE the node will not be deployed or will be dropped during redeployment.

Dimension With Task Options

Dimension_task_opt

  • Development Mode: True / False toggle that determines whether a task will be created or if the SQL to be used in the task will execute as DML as a Run action. Prior to creating a task, it is helpful to test the SQL the task will execute to make sure it runs without errors and returns the expected data.
    • True - A table will be created and SQL will execute as a Run action.
    • False - After sufficiently testing the SQL as a Run action, setting Development Mode to false will wrap the SQL statement in a task with options specified in Scheduling Options.
  • Multi Source: True / False toggle that is Coalesce implementation of SQL UNIONs.
    • True - Multiple sources can be combined in a single node. The sources are combined using the option specified in the Multi Source Strategy.
      • UNION - Combines with duplicate elimination.
      • UNION ALL - Combines without duplicate elimination.
    • False - Single source node or multiple sources combined using a join.
  • Business key: It is a required column for both Type 1 and Type 2 Dimensions.
  • Change tracking: It is a required column for Type 2 Dimension.
  • Cluster key: True/False to determine whether Dimension is to be clustered or not.
    • True - Allows you to specify the column based on which clustering is to be done.
      • Allow Expressions Cluster Key: True/False. Add an expression to the specified cluster key.
    • False – No clustering done

Dimension With Task General Options

Work with Task-GO

  • DISTINCT: True/False toggle that determines whether to add DISTINCT to SQL Query.
    • True - Group by All is invisible. DISTINCT data is chosen for processing.
    • False- Group by All is visible.
  • GROUP BY ALL: True/False toggle that determines whether to add GROUP BY ALL to SQL Query.
    • True - DISTINCT is invisible. Data is grouped by all columns for processing.
    • False- DISTINCT is visible.
  • Order By : True/False toggle that determines whether to add “ORDER BY” to SQL query along with the column and sort order.
    • True -Sort column and sort order drop down are visible and are required to form order by clause.
    • False-Sort column and sort order drop down are invisible.

Dimension With Task Scheduling Options

If development mode is set to false then Scheduling Options can be used to configure how and when the task will run.

  • Scheduling Mode: Specifies whether a warehouse or serverless compute is used to run the task
    • Warehouse Task - User managed warehouse will execute task.
    • Serverless Task - Utilize serverless compute to execute task.
  • When Source Stream has Data Flag: True / False toggle that checks whether source streams have data before executing a task.
    • True - Only run task if source stream has capture change data
    • False - Run task on schedule regardless of whether the source stream has data. If the source is not a stream should set this to false.
  • Multiple Stream has Data Logic: True / False toggle is enabled if Source Stream has Data Flag is set to true.
    • AND - If there are multiple streams task will run if all streams have data
    • OR - If there are multiple streams task will run if one or more streams has data
  • Select Warehouse on which to run task: Visible if Scheduling Mode is set to Warehouse Task.
    • Enter the name of the warehouse you want the task to run on without quotes
  • Select initial serverless Warehouse size: Visible when Scheduling Mode is set to Serverless Task.
    • Select the initial compute size on which to run the task. Snowflake will adjust size from there based on target schedule and task run times.
  • Task Schedule: Select how you want to schedule the task to run
    • Minutes - Allows you to specify a minute interval for running task
    • Cron - Allows you to specify a CRON schedule for running task
    • Predecessor - Allows you to specify a predecessor task to determine when a task should execute
  • Enter task schedule using minutes: Only visible when Task Schedule is set to Minutes. Enter a whole number from 1 to 11520 which represents the number of minutes between task runs.
  • Enter task schedule using Cron: Only visible when Task Schedule is set to Cron. Specifies a cron expression and time zone for periodically running the task. Supports a subset of standard cron utility syntax.
  • Enter predecessor tasks separated by a comma: Only visible when Task Schedule is set to Predecessor. One or more task names that precede the task being created in the current node. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor task separate the task names using a comma and no spaces.
  • Enter root task name: Name of the root task that controls scheduling for the DAG of tasks. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor task separate the task names using a comma and no spaces.

Dimension With Task Deployment

Dimension With Task Deployment Parameters

The Dimension with Task includes an environment parameter that allows you to specify a different warehouse used to run a task in different environments.

The parameter name is targetTaskWarehouse and the default value is DEV ENVIRONMENT.

When set to DEV ENVIRONMENT the value entered in the Scheduling Options config Select Warehouse on which to run task will be used when creating the task.

{
"targetTaskWarehouse": "DEV ENVIRONMENT"
}

When set to any value other than DEV ENVIRONMENT the node will attempt to create the task using a Snowflake warehouse with the specified value.

For example, with the below setting for the parameter in a QA environment, the task will execute using a warehouse named compute_wh.

{
"targetTaskWarehouse": "compute_wh"
}

Dimension With Task Initial Deployment

When deployed for the first time into an environment the Dimension with Task node will execute three stages dependent on whether or not the task schedule relies on a predecessor task:

Dimension With Task No Predecessor Task Deployment

  • Create Target Table: This stage will create a table that will be loaded by the task.
  • Create Task: This stage will create a task that will load the target table on the schedule specified.
  • Resume Task: After the task has been created it needs to be resume so that the task runs on the schedule.

Dimension With Task Predecessor Task Deployment

  • Create Target Table: This stage will create a table that will be loaded by the task.
  • Suspend Root Task: To add a task into a DAG of task the root task needs to be put into a suspended state.
  • Create Task: This stage will create a task that will load the target table on the schedule specified.

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

The task node has no ALTER capabilities. All task enabled nodes are CREATE OR REPLACE only though this is subject to change.

Dimension With Task Redeployment

Dimension With Task Recreating the Task Redeployment

After the Task has deployed for the first time into a target environment, subsequent deployments with changes in task schedule,warehouse or scheduling options will result in a CREATE TASK AND RESUME TASK statements being issued.

Dimension With Task No Predecessor Task Redeployment

  • Create Task: This stage will create a task that will load the target table on the schedule specified.
  • Resume Task: After the task has been created it needs to be resume so that the task runs on the schedule.

Dimension With Task Predecessor Task Redeployment

  • Suspend Root Task: To add a task into a DAG of task the root task needs to be put into a suspended state.
  • Create Task: This stage will create a task that will load the target table on the schedule specified.

Dimension With Task Altering the Tables Redeployment

After the Task has deployed for the first time into a target environment, subsequent deployments with changes in table like add/drop column,change in data type will result in an ALTER table statement followed by CREATE TASK AND RESUME TASK statements being issued.

The following stages are executed:

  • Change Column Attributes/Delete Column/Add Column/Change table description: Alter table statement is executed to perform the alter operation accordingly.
  • Create Task
  • Resume Task

Dimension With Task Undeployment

If a Dimension with Task node is deleted from a Workspace, that Workspace is committed to Git and that commit deployed to a higher-level environment then all objects created by the node in the target environment will be dropped.

Dimension With Task No Predecessor Task Undeployment

  • Drop Table: This stage will drop the table originally created to be loaded by the task.
  • Drop Current Task: This stage will drop the task.

Dimension With Task Predecessor Task Undeployment

  • Drop Table: This stage will drop the table originally created to be loaded by the task.
  • Suspend Root Task: To drop a task from a DAG of task the root task needs to be put into a suspended state.
  • Drop Task: This stage will drop the task.

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

Fact With Task

The Coalesce Fact with Task UDN is a node that wraps the standard Coalesce Fact node with a Task.

Tasks can be combined with Coalesce Stream node (table streams) for continuous ELT workflows to process recently changed table rows. Streams ensure exactly once semantics for new or changed data in a table.

Tasks can also be used independently to generate periodic reports by inserting or merging rows into a report table or performing other periodic work.

More information about Tasks can be found in Snowflake's Introduction to tasks.

Fact With Task Node Configuration

The Work with Task node has two or three configuration groups depending on config options selected.

With the option Development Mode set to true there are three config groups:

Work_with_dev_true

With the Development Mode set to false Scheduling Options is added.

Work_with_dev_false

Fact With Task Node Properties

  • Storage Location: Storage Location where the Stream will be created.
  • Node Type: Name of template used to create node objects.
  • Description: A description of the node's purpose.
  • Deploy Enabled:
    • If TRUE the node will be deployed / redeployed when changes are detected.
    • If FALSE the node will not be deployed or will be dropped during redeployment.

Fact With Task Options

Worktask-opt

  • Development Mode: True / False toggle that determines whether a task will be created or if the SQL to be used in the task will execute as DML as a Run action. Prior to creating a task, it is helpful to test the SQL the task will execute to make sure it runs without errors and returns the expected data.
    • True - A table will be created and SQL will execute as a Run action.
    • False - After sufficiently testing the SQL as a Run action setting Development Mode to false will wrap the SQL statement in a task with options specified in Scheduling Options.
  • Multi Source: True / False toggle that is Coalesce implementation of SQL UNIONs.
    • True - Multiple sources can be combined in a single node. The sources are combined using the option specified in the Multi Source Strategy.
      • UNION - Combines with duplicate elimination.
      • UNION ALL - Combines without duplicate elimination.
    • False - Single source node or multiple sources combined using a join.
  • Cluster key: True/False to determine whether Dimension is to be clustered or not.
    • True - Allows you to specify the column based on which clustering is to be done.
      • Allow Expressions Cluster Key: True/False. Add an expression to the specified cluster key.
    • False – No clustering done
  • Truncate Before: Specifies that the target table should be truncated before inserting the values into the table.

Fact With Task General Options

Work with Task-GO

  • DISTINCT: True/False toggle that determines whether to add DISTINCT to SQL Query.
    • True - Group by All is invisible. DISTINCT data is chosen for processing.
    • False- Group by All is visible.
  • GROUP BY ALL: True/False toggle that determines whether to add GROUP BY ALL to SQL Query.
    • True - DISTINCT is invisible. Data is grouped by all columns for processing.
    • False- DISTINCT is visible.
  • Order By : True/False toggle that determines whether to add “ORDER BY” to SQL query along with the column and sort order.
    • True -Sort column and sort order drop down are visible and are required to form order by clause.
    • False-Sort column and sort order drop down are invisible.

Fact With Task Scheduling Options

If development mode is set to false then Scheduling Options can be used to configure how and when the task will run.

  • Scheduling Mode: Specifies whether a warehouse or serverless compute is used to run the task
    • Warehouse Task - User managed warehouse will execute task.
    • Serverless Task - Utilize serverless compute to execute task.
  • When Source Stream has Data Flag: True / False toggle that checks whether source streams have data before executing a task.
    • True - Only run task if source stream has capture change data
    • False - Run task on schedule regardless of whether the source stream has data. If the source is not a stream should set this to false.
  • Multiple Stream has Data Logic: True / False toggle is enabled if Source Stream has Data Flag is set to true.
    • AND - If there are multiple streams task will run if all streams have data
    • OR - If there are multiple streams task will run if one or more streams has data
  • Select Warehouse on which to run task: Visible if Scheduling Mode is set to Warehouse Task.
    • Enter the name of the warehouse you want the task to run on without quotes
  • Select initial serverless Warehouse size: Visible when Scheduling Mode is set to Serverless Task.
    • Select the initial compute size on which to run the task. Snowflake will adjust size from there based on target schedule and task run times.
  • Task Schedule: Select how you want to schedule the task to run
    • Minutes - Allows you to specify a minute interval for running task
    • Cron - Allows you to specify a CRON schedule for running task
    • Predecessor - Allows you to specify a predecessor task to determine when a task should execute
  • Enter task schedule using minutes: Only visible when Task Schedule is set to Minutes. Enter a whole number from 1 to 11520 which represents the number of minutes between task runs.
  • Enter task schedule using Cron: Only visible when Task Schedule is set to Cron. Specifies a cron expression and time zone for periodically running the task. Supports a subset of standard cron utility syntax.
  • Enter predecessor tasks separated by a comma: Only visible when Task Schedule is set to Predecessor. One or more task names that precede the task being created in the current node. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor task separate the task names using a comma and no spaces.
  • Enter root task name: Name of the root task that controls scheduling for the DAG of tasks. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor task separate the task names using a comma and no spaces.

Fact With Task Deployment

Fact With Task Deployment Parameters

The Fact with Task includes an environment parameter that allows you to specify a different warehouse used to run a task in different environments.

The parameter name is targetTaskWarehouse and the default value is DEV ENVIRONMENT.

When set to DEV ENVIRONMENT the value entered in the Scheduling Options config Select Warehouse on which to run the task will be used when creating the task.

{
"targetTaskWarehouse": "DEV ENVIRONMENT"
}

When set to any value other than DEV ENVIRONMENT the node will attempt to create the task using a Snowflake warehouse with the specified value.

For example, with the below setting for the parameter in a QA environment, the task will execute using a warehouse named compute_wh.

{
"targetTaskWarehouse": "compute_wh"
}

Fact With Task Initial Deployment

When deployed for the first time into an environment the Fact with Task node will execute three stages dependent on whether or not the task schedule relies on a predecessor task:

Fact With Task No Predecessor Task Initial Deployment

  • Create Target Table: This stage will create a table that will be loaded by the task.
  • Create Task: This stage will create a task that will load the target table on the schedule specified.
  • Resume Task: After the task has been created it needs to be resume so that the task runs on the schedule.

Fact With Task Predecessor Task Initial Deployment

  • Create Target Table: This stage will create a table that will be loaded by the ask
  • Suspend Root Task: To add a task into a DAG of task the root task needs to be put into a suspended state.
  • Create Task: This stage will create a task that will load the target table on the schedule specified.

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

The task node has no ALTER capabilities. All task enabled nodes are CREATE OR REPLACE only though this is subject to change.

Fact With Task Redeployment

After the Task has deployed for the first time into a target environment, subsequent deployments with changes in task schedule,warehouse or scheduling options will result in a CREATE TASK AND RESUME TASK statements being issued

The following stages are executed:

Fact With Task Redeployment No Predecessor Task

  • Create Task: This stage will create a task that will load the target table on the schedule specified.
  • Resume Task: After the task has been created it needs to be resume so that the task runs on the schedule.

Fact With Task Redeployment Predecessor Task

  • Suspend Root Task: To add a task into a DAG of task the root task needs to be put into a suspended state.
  • Create Task: This stage will create a task that will load the target table on the schedule specified.

Fact With Task Altering the tables

After the Task has deployed for the first time into a target environment, subsequent deployments with changes in table like add/drop column,change in data type will result in an ALTER table statement followed by CREATE TASK AND RESUME TASK statements being issued.

The following stages are executed:

  • Change Column Attributes/Delete Column/Add Column/Change table description: Alter table statement is executed to perform the alter operation accordingly.
  • Create Task
  • Resume Task

Fact With Task Undeployment

If a Fact with Task node is deleted from a Workspace, that Workspace is committed to Git and that commit deployed to a higher level environment then all objects created by the node in the target environment will be dropped.

Fact With Task Undeployment No Predecessor Task

  • Drop Table: This stage will drop the table originally created to be loaded by the task/
  • Drop Current Task: This stage will drop the task.

Fact With Task Undeployment Predecessor Task

  • Drop Table: This stage will drop the table originally created to be loaded by the task.
  • Suspend Root Task: To drop a task from a DAG of task the root task needs to be put into a suspended state.
  • Drop Task: This stage will drop the task.

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

Task DAG Create Root

The Coalesce Task DAG Create Root UDN is a node that helps to create a standalone root task.

The root task should have a defined schedule that initiates a run of the DAG. Each of the other tasks has at least one defined predecessor to link the tasks in the DAG. A child task runs only after all of its predecessor tasks run successfully to completion.

Tasks can also be used independently to generate periodic reports by inserting or merging rows into a report table or perform other periodic work.

More information about Tasks can be found in Snowflake's Introduction to tasks.

Task DAG Create Root Node Configuration

Task_dag_create_root_config

Task DAG Create Root Node Properties

  • Storage Location: Storage Location where the Stream will be created.
  • Node Type: Name of template used to create node objects.
  • Description: A description of the node's purpose.
  • Deploy Enabled:
    • If TRUE the node will be deployed / redeployed when changes are detected.
    • If FALSE the node will not be deployed or will be dropped during redeployment.

Task DAG Create Root Scheduling Options

The Scheduling Options can be used to configure how and when the task will run.

There are 4 to 5 options that can be selected depending on combinations of configs that are selected:

Task_dag_create_root_scheduling_options

  • Scheduling Options: Specifies whether a warehouse or serverless compute is used to run the task
    • Warehouse Task - User managed warehouse will execute task.
    • Serverless Task - Utilize serverless compute to execute task.
  • Select Warehouse on which to run task: Visible if Scheduling Mode is set to Warehouse Task
    • Enter the name of the warehouse you want the task to run on without quotes
  • Select initial serverless Warehouse size: Visible when Scheduling Mode is set to Serverless Task
    • Select the initial compute size on which to run the task. Snowflake will adjust size from there based on target schedule and task run times.
  • Task Schedule: Select how you want to schedule the task to run
    • Minutes - Allows you to specify a minute interval for running task
    • Cron - Allows you to specify a CRON schedule for running task
  • Enter root task SQL: The SQL statement associated to be run when a standalone root task to be executed.

Task DAG Create Root Deployment

Task DAG Create Root Deployment Parameters

The Dimension with Task includes an environment parameter that allows you to specify a different warehouse used to run a task in different environments.

The parameter name is targetTaskWarehouse and the default value is DEV ENVIRONMENT.

When set toDEV ENVIRONMENT the value entered in the Scheduling Options config Select Warehouse on which to run the task will be used when creating the task.

{
"targetTaskWarehouse": "DEV ENVIRONMENT"
}

When set to any value other than DEV ENVIRONMENT the node will attempt to create the task using a Snowflake warehouse with the specified value.

For example, with the below setting for the parameter in a QA environment, the task will execute using a warehouse named compute_wh.

{
"targetTaskWarehouse": "compute_wh"
}

Task DAG Create Root Initial Deployment

When deployed for the first time into an environment the Task DAG Create Root node will execute two stages.

Task DAG Create Root Task

  • Suspend Root Task: To create a root task, the root task needs to be put into a suspended state
  • Create Root Task: This stage will create a task that will execute the Root SQL on the schedule specified

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

Task DAG Create Root Redeployment

After the Task has deployed for the first time into a target environment, subsequent deployments will execute two stages Suspend Root Task and Create Root task

Task DAG Create Root Undeployment

If a Dimension with Task node is deleted from a Workspace, that Workspace is committed to Git and that commit deployed to a higher-level environment then two stages are executed

  • Suspend Root task
  • Drop current task

Task DAG Resume Root

The Coalesce Task DAG Resume Root UDN is a node type that helps to resume the root task and its dependents or child tasks. Recursively resumes all dependent tasks tied to a specified root task in a DAG using the root task name specified.

The root task should have a defined schedule that initiates a run of the DAG. Each of the other tasks has at least one defined predecessor to link the tasks in the DAG. A child task runs only after all of its predecessor tasks run successfully to completion.

Tasks can also be used independently to generate periodic reports by inserting or merging rows into a report table or perform other periodic work.

More information about Tasks can be found in the official Snowflake documentation here - (https://docs.snowflake.com/en/user-guide/tasks-intro).

Task DAG Resume Root Node Configuration

Task_dag_create_root_config

Task DAG Resume Root Node Node Properties

  • Storage Location: Storage Location where the Stream will be created.
  • Node Type: Name of template used to create node objects.
  • Description: A description of the node's purpose.
  • Deploy Enabled:
    • If TRUE the node will be deployed / redeployed when changes are detected.
    • If FALSE the node will not be deployed or will be dropped during redeployment.

Task DAG Resume Root Scheduling Options

The Scheduling Options prompts to mention the root task which needs to be resumed. It recursively resumes all dependent tasks tied to a specified root task

Task_dag_resume_root_scheduling_options

Task DAG Resume Root Deployment

Task DAG Resume RootInitial Deployment

When deployed for the first time into an environment the Dimension with Task node will execute:

Try Enable Root Task: to resume root and all its dependents.

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

Task DAG Resume Root Undeployment

If a Dimension with Task node is deleted from a Workspace, that Workspace is committed to Git and that commit deployed to a higher-level environment then two stages are executed

Stream

The Coalesce Stream UDN is a node that allows you to develop and deploy a stream on top of a table, view or external table.

A stream logically takes an initial snapshot of every row in the source object (e.g. table, external table, or the underlying tables for a view) by initializing a point in time (called an offset) as the current transactional version of the object. The change tracking system utilized by the stream then records information about the DML changes after this snapshot was taken. Change records provide the state of a row before and after the change. Change information mirrors the column structure of the tracked source object and includes additional metadata columns that describe each change event.

More information about Streams can be found in the official Snowflake's Introduction to Streams.

Stream Node Configuration

The Stream has two configuration groups:

Str-opt

Stream Node Properties

  • Storage Location: Storage Location where the Stream will be created.
  • Node Type: Name of template used to create node objects.
  • Description: A description of the node's purpose.
  • Deploy Enabled:
    • If TRUE the node will be deployed / redeployed when changes are detected.
    • If FALSE the node will not be deployed or will be dropped during redeployment.

Stream Options

  • Source Object: The type of object the stream will be created on. One of three options is required to be selected and this selection drives what other configs are available:
    • Table
      • Append Only Stream: True / False Toggle to set what type of stream is created.
        • True - Specifies an append-only stream.
        • False - Specifies a standard stream.
      • Show Initial Rows: True / False Toggle to specify the records to return the first time the stream is consumed.
        • True - The stream returns only the rows that existed in the source object at the moment when the stream was created. Subsequently, the stream returns any DML changes to the source object since the most recent offset; that is, the normal stream behavior.
        • False - The stream returns any DML changes to the source object since the most recent offset.
      • Redeployment Behavior: Options to determine redeployment behavior.
    • View:
      • Append Only Stream: True / False Toggle to set what type of stream is created.
        • True - Specifies an append-only stream.
        • False - Specifies a standard stream.
      • Show Initial Rows: True / False Toggle to specify the records to return the first time the stream is consumed.
        • True - The stream returns only the rows that existed in the source object at the moment when the stream was created. Subsequently, the stream returns any DML changes to the source object since the most recent offset; that is, the normal stream behavior.
        • False - The stream returns any DML changes to the source object since the most recent offset.
      • Redeployment Behavior: Options to determine redeployment behavior.
    • External table:
      • Redeployment Behavior: Options to determine redeployment behavior.
    • External iceberg table:
      • Redeployment Behavior: Options to determine redeployment behavior.

Stream System Columns

A Stream UDN adds three system columns to the output of the node.

  • METADATA$ACTION
  • METADATA$ISUPDATE
  • METADATA$ROW_ID

These columns can be used together to track INSERT, UPDATE and DELETE operations against a source object.

METADATA$ACTION

Indicates the DML operation (INSERT, DELETE) recorded.

METADATA$ISUPDATE

Indicates whether the operation was part of an UPDATE statement. Updates to rows in the source object are represented as a pair of DELETE and INSERT records in the stream with a metadata column METADATA$ISUPDATE values set to TRUE.

METADATA$ROW_ID

Specifies the unique and immutable ID for the row, which can be used to track changes to specific rows over time.

Stream Deployment

Stream Deployment Parameters

No deployment parameters are required.

Stream Initial Deployment

When deployed for the first time into an environment the Stream node will execute stage:

Create Stream This stage will execute a CREATE OR REPLACE statement and create a Stream in the target environment.

Stream Redeployment

After the Stream has deployed for the first time into a target environment, subsequent deployments will result in a new stream creation based on redeployment behavior chosen.

Redeployment BehaviorStage Executed
Create Stream if not existsRe-Create Stream at existing offset
Create or ReplaceCreate Stream
Create at existing streamRe-Create Stream at existing offset

Stream Undeployment

If a Stream Node is deleted from a Workspace, that Workspace is committed to Git and that commit deployed to a higher level environment then the Stream in the target environment will be dropped.

This is executed as a single stage:

  • Drop Stream

Stream-Insert or Merge

The Coalesce Streams and Insert or Merge UDN is a node that allows you to develop and deploy a stream on top of a table, view or external table.Also,provides option to create a target table to insert or merge data from source with a task on top of it.

A stream logically takes an initial snapshot of every row in the source object (e.g. table, external table, or the underlying tables for a view) by initializing a point in time (called an offset) as the current transactional version of the object. The change tracking system utilized by the stream then records information about the DML changes after this snapshot was taken. Change records provide the state of a row before and after the change. Change information mirrors the column structure of the tracked source object and includes additional metadata columns that describe each change event.

More information about Streams can be found in the official Snowflake's Introduction to Streams.

Stream-Insert or Merge Node Configuration

SIM_Opt

Stream-Insert or Merge Node Properties

  • Storage Location: Storage Location where the Stream will be created.
  • Node Type: Name of template used to create node objects.
  • Description: A description of the node's purpose.
  • Deploy Enabled:
    • If TRUE the node will be deployed / redeployed when changes are detected.
    • If FALSE the node will not be deployed or will be dropped during redeployment.

Stream-Insert or Merge General Options

SIM-GO

  • Development Mode: True / False toggle that determines whether a task will be created or if the SQL to be used in the task will execute as DML as a Run action. Prior to creating a task, it is helpful to test the SQL the task will execute to make sure it runs without errors and returns the expected data.
    • True - A table will be created and SQL will execute as a Run action.
    • False - After sufficiently testing the SQL as a Run action setting Development Mode to false will wrap the SQL statement in a task with options specified in Scheduling Options.
  • CREATE AS: In this dropdown you can select if you want to create target object as 'Table' or 'Transient Table' .
    • Table - A permanent table in Snowflake with data retention and fail-safe features for long-term storage and data protection.
    • Transient Table- A temporary table in Snowflake without data retention and fail-safe features, suitable for intermediate data processing.
  • DISTINCT: True/False toggle that determines whether to add DISTINCT to SQL Query.
    • True - Group by All is invisible. DISTINCT data is chosen for processing
    • False- Group by All is visible
  • GROUP BY ALL: True/False toggle that determines whether to add GROUP BY ALL to SQL Query.
    • True - DISTINCT is invisible. Data is grouped by all columns for processing
    • False- DISTINCT is visible

Stream-Insert or Merge Stream Options

  • Source Object: The type of object the stream will be created on. One of three options is required to be selected and this selection drives what other configs are available:
    • Table
      • Append Only Stream: True / False Toggle to set what type of stream is created.
        • True - Specifies an append-only stream.
        • False - Specifies a standard stream.
      • Show Initial Rows: True / False Toggle to specify the records to return the first time the stream is consumed.
        • True - The stream returns only the rows that existed in the source object at the moment when the stream was created. Subsequently, the stream returns any DML changes to the source object since the most recent offset; that is, the normal stream behavior.
        • False - The stream returns any DML changes to the source object since the most recent offset.
      • Redeployment Behavior: Options to determine redeployment behavior.
    • View:
      • Append Only Stream: True / False Toggle to set what type of stream is created.
        • True - Specifies an append-only stream.
        • False - Specifies a standard stream.
      • Show Initial Rows: True / False Toggle to specify the records to return the first time the stream is consumed.
        • True - The stream returns only the rows that existed in the source object at the moment when the stream was created. Subsequently, the stream returns any DML changes to the source object since the most recent offset; that is, the normal stream behavior.
        • False - The stream returns any DML changes to the source object since the most recent offset.
      • Redeployment Behavior: Options to determine redeployment behavior.

Stream-Insert or Merge Target Loading Options

SIM_TL

  • Load Type:
    • Insert:The data is inserted into the target table from source.
    • Merge:The latest record changes in the source are merged into target.
  • Table key(s): Enabled if target load type is Merge.The business key(s) column(s) based on which the data is merged into Target table.
  • Record Date/Timestamp:Enabled if target load type is Merge.Date/TImestamp column based on which latest record is merged into Target table.
  • Cluster key: True/False to determine whether Dimension is to be clustered or not.
    • True - Allows you to specify the column based on which clustering is to be done.
      • Allow Expressions Cluster Key: True/False. Add an expression to the specified cluster key.
    • False – No clustering done

Stream-Insert or Merge Scheduling Options

If development mode is set to false then Scheduling Options can be used to configure how and when the task will run.

SIM_scheduling

  • Scheduling Mode: Specifies whether a warehouse or serverless compute is used to run the task
    • Warehouse Task - User managed warehouse will execute tasks.
    • Serverless Task - Utilize serverless compute to execute tasks.
  • When Source Stream has Data Flag: True / False toggle that checks whether source streams have data before executing a task.
    • True - Only run task if source stream has capture change data
    • False - Run tasks on schedule regardless of whether the source stream has data. If the source is not a stream should set this to false.
  • Select Warehouse on which to run task: Visible if Scheduling Mode is set to Warehouse Task
    • Enter the name of the warehouse you want the task to run on without quotes
  • Select initial serverless Warehouse size: Visible when Scheduling Mode is set to Serverless Task
    • Select the initial compute size on which to run the task. Snowflake will adjust size from there based on target schedule and task run times.
  • Task Schedule: Select how you want to schedule the task to run
    • Minutes - Allows you to specify a minute interval for running task
    • Cron - Allows you to specify a CRON schedule for running task
    • Predecessor - Allows you to specify a predecessor task to determine when a task should execute
  • Enter task schedule using minutes: Only visible when Task Schedule is set to Minutes. Enter a whole number from 1 to 11520 which represents the number of minutes between task runs.
  • Enter task schedule using Cron: Only visible when Task Schedule is set to Cron. Specifies a cron expression and time zone for periodically running the task. Supports a subset of standard cron utility syntax.
  • Enter predecessor task(s) separated by a comma: Only visible when Task Schedule is set to Predecessor. One or more task names that precede the task being created in the current node. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor tasks separate the task names using a comma and no spaces.
  • Enter root task name: Name of the root task that controls scheduling for the DAG of tasks. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor tasks separate the task names using a comma and no spaces.

Stream-Insert or Merge System Columns

A Stream UDN adds three system columns to the output of the node. These columns can be used together to track INSERT, UPDATE and DELETE operations against a source object.

  • METADATA$ACTION: Indicates the DML operation (INSERT, DELETE) recorded.
  • METADATA$ISUPDATE: Indicates whether the operation was part of an UPDATE statement. Updates to rows in the source object are represented as a pair of DELETE and INSERT records in the stream with a metadata column METADATA$ISUPDATE values set to TRUE.
  • METADATA$ROW_ID: Specifies the unique and immutable ID for the row, which can be used to track changes to specific rows over time.

Stream-Insert or Merge Deployment

Stream-Insert or Merge Deployment Parameters

The Dimension with Task includes an environment parameter that allows you to specify a different warehouse used to run a task in different environments.

The parameter name is targetTaskWarehouse and the default value is DEV ENVIRONMENT.

When set to DEV ENVIRONMENT the value entered in the Scheduling Options config Select Warehouse on which to run task will be used when creating the task.

{
"targetTaskWarehouse": "DEV ENVIRONMENT"
}

When set to any value other than DEV ENVIRONMENT the node will attempt to create the task using a Snowflake warehouse with the specified value.

For example, with the below setting for the parameter in a QA environment, the task will execute using a warehouse named compute_wh.

{
"targetTaskWarehouse": "compute_wh"
}

Stream-Insert or Merge Initial Deployment

When deployed for the first time into an environment the Stream and Insert or Merge node will execute four to six stages dependent on whether or not the task schedule relies on a predecessor task:

Stream-Insert or Merge Deployment No Predecessor Task

Create Stream This stage will execute a CREATE OR REPLACE statement and create a Stream in the target environment.

Create Work Table/Transient Table This stage will create a table that will be loaded by the task.

Target Table Initial Load Loads initial data into table

Create Task This stage will create a task that will load the target table on the schedule specified.

Resume Task After the task has been created it needs to be resume so that the task runs on the schedule.

Stream-Insert or Merge Deployment Predecessor Task

Create Stream This stage will execute a CREATE OR REPLACE statement and create a Stream in the target environment.

Create Work Table/Transient Table This stage will create a table that will be loaded by the ask.

Target Table Initial Load Loads initial data into table

Suspend Root Task To add a task into a DAG of task the root task needs to be put into a suspended state.

Create Task This stage will create a task that will load the target table on the schedule specified.

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

The task node has no ALTER capabilities. All task enabled nodes are CREATE OR REPLACE only though this is subject to change.

Stream-Insert or Merge Redeployment

Redeployment BehaviorStage Executed
Create Stream if not existsRe-Create Stream at existing offset
Create or ReplaceCreate Stream
Create at existing streamRe-Create Stream at existing offset

Table

There are few column or table changes like Change in table name,Dropping existing column, Alter Column data type,Adding a new column if made in isolation or all-together will result in an ALTER statement to modify the Work Table in the target environment. The following stages are executed:

  • Rename Table| Alter Column | Delete Column | Add Column | Edit table description: Alter table statement is executed to perform the alter operation.
  • Target Inital Load :If the initial load toggle is enabled and the redeployment behavior of the stream is "Create or Replace," it loads the table with "INSERT OVERWRITE INTO." For all other scenarios, it uses "INSERT INTO."

If the materialization type is changed from transient table to table or vie versa,the following stages are executed

  • Drop table/transient table
  • Create Work table/transient table
  • Target Inital Load :If the initial load toggle is enabled and the redeployment behavior of the stream is "Create or Replace," it loads the table with "INSERT OVERWRITE INTO." For all other scenarios, it uses "INSERT INTO."

Task

Redeployment with changes in the stream or table will result in the creation and resumption of the task.

Stream-Insert or Merge Undeployment

If a Stream Node is deleted from a Workspace, that Workspace is committed to Git and that commit deployed to a higher level environment then the Stream in the target environment will be dropped.

This is executed in following stages:

  • Drop Stream
  • Drop Table
  • Drop Current Task

Iceberg table with Task

Iceberg tables with task node is an external iceberg table node wrapped with a task functionality.

An Iceberg table uses the Apache Iceberg open table format specification, which provides an abstraction layer on data files stored in open formats.Iceberg tables for Snowflake combine the performance and query semantics of regular Snowflake tables with external cloud storage that you manage. They are ideal for existing data lakes that you cannot, or choose not to, store in Snowflake.

An Iceberg table that uses an external catalog provides limited Snowflake platform support with read-only access. With this table type, Snowflake uses a catalog integration to retrieve information about your Iceberg metadata and schema.

You can use this option to create an Iceberg table registered in the AWS Glue Data Catalog or to create a table from Iceberg metadata files in object storage.

Iceberg table with task Configuration

The Snowflake Iceberg table has three configuration groups:

Go to the node and select the Config tab to see the Node Properties,Iceberg Options and Scheduling Options.

Iceberg table with task node properties

There are four configs within the Node Properties group.

  • Storage Location(required): Storage Location where the Dynamic Table will be created.
  • Node Type(required): Name of template used to create node objects.
  • Description: A description of the node's purpose.
  • Deploy Enabled(required):
    • If TRUE the node will be deployed or redeployed when changes are detected.
    • If FALSE the node will not be deployed or the node will be dropped during redeployment.

Iceberg Options

  • Type of Catalog:Specify the type of catalog *AWS Glue *Object Storage
  • Snowflake EXTERNAL VOLUME name: Specifies the identifier (name) for the external volume where the Iceberg table stores its metadata files and data in Parquet format.External volume needs to be created in snowflake as a prerequisite.
  • Catalog integration: Specifies the identifier (name) of the catalog integration for this table.
  • Catalog namespace:Optionally specifies the namespace (for example, my_glue_database) for the AWS Glue Data Catalog source.Option available if AWS Glue catalog is chosen.
  • Catalog table name:Name of the catalog table.Option available if AWS Glue catalog is chosen.
  • Metadata filepath:Specifies the relative path of the Iceberg metadata file to use for column definitions.Option available if Object Storage Catalog is chosen.
  • Schedule refresh: True / False toggle that determines whether a task will be created or if the SQL to be used in the task will execute as DML as a Run action. Prior to creating a task, it is helpful to test the SQL the task will execute to make sure it runs without errors and returns the expected data.
    • False - A table will be created and SQL will execute as a Run action.
    • True - After sufficiently testing the SQL as a Run action, setting Schedule refresh Mode to true will wrap the SQL statement in a task with options specified in Scheduling Options.

Scheduling Options

If schedule refresh mode is set to true then Scheduling Options can be used to configure how and when the task will run.

  • Scheduling Mode: Specifies whether a warehouse or serverless compute is used to run the task
    • Warehouse Task - User managed warehouse will execute tasks.
    • Serverless Task - Utilize serverless compute to execute tasks.
  • Select Warehouse on which to run task: Visible if Scheduling Mode is set to Warehouse Task
    • Enter the name of the warehouse you want the task to run on without quotes
  • Select initial serverless Warehouse size: Visible when Scheduling Mode is set to Serverless Task
    • Select the initial compute size on which to run the task. Snowflake will adjust size from there based on target schedule and task run times.
  • Task Schedule: Select how you want to schedule the task to run
    • Minutes - Allows you to specify a minute interval for running task
    • Cron - Allows you to specify a CRON schedule for running task
    • Predecessor - Allows you to specify a predecessor task to determine when a task should execute
  • Enter task schedule using minutes: Only visible when Task Schedule is set to Minutes. Enter a whole number from 1 to 11520 which represents the number of minutes between task runs.
  • Enter task schedule using Cron: Only visible when Task Schedule is set to Cron. Specifies a cron expression and time zone for periodically running the task. Supports a subset of standard cron utility syntax.
  • Enter predecessor task(s) separated by a comma: Only visible when Task Schedule is set to Predecessor. One or more task names that precede the task being created in the current node. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor tasks separate the task names using a comma and no spaces.
  • Enter root task name: Name of the root task that controls scheduling for the DAG of tasks. Task names are case sensitive, should not be quoted and must exist in the same schema in which the current task is being created. If there are multiple predecessor tasks separate the task names using a comma and no spaces.

Prerequisites

  • The Role we mention in the Workspace and Environment properties of Coalesce should be 'ACCOUNTADMIN' inorder to successfully create an iceberg table.You can also grant SYSADMIN roles to EXTERNAL VOLUME,CATALOG INTEGRATION created.
  • An EXTERNAL VOLUME,CATALOG INTEGRATION is expected to be created in Snowflake at the Storage Location chosen in the Node properties.

Iceberg tables with Task - System Columns

  • DATA - Column added for deployment but column is not added to iceberg table as columns specifications are not required for External Iceberg tables.

Iceberg tables With Task Deployment Parameters

The Iceberg tables With Task includes an environment parameter that allows you to specify a different warehouse used to run a task in different environments.

The parameter name is targetTaskWarehouse and the default value is DEV ENVIRONMENT.

When set to DEV ENVIRONMENT the value entered in the Scheduling Options config Select Warehouse on which to run the task will be used when creating the task.

{
"targetTaskWarehouse": "DEV ENVIRONMENT"
}

When set to any value other than DEV ENVIRONMENT the node will attempt to create the task using a Snowflake warehouse with the specified value.

For example, with the below setting for the parameter in a QA environment, the task will execute using a warehouse named compute_wh.

{
"targetTaskWarehouse": "compute_wh"
}

Iceberg tables with task Initial Deployment

When deployed for the first time into an environment the Work with Task node will execute three stages dependent on whether or not the task schedule relies on a predecessor task:

Iceberg tables with task No Predecessor Task Deployment

  • Create Iceberg Table: This stage will execute a CREATE OR REPLACE statement and create a Iceberg Table in the target environment.
  • Create Task: This stage will create a task that will load the target table on the schedule specified
  • Resume Task: After the task has been created it needs to be resume so that the task runs on the schedule

Iceberg tables with task Predecessor Task Deployment

  • Create Iceberg Table: This stage will execute a CREATE OR REPLACE statement and create a Iceberg Table in the target environment.
  • Suspend Root Task: To add a task into a DAG of task the root task needs to be put into a suspended state
  • Create Task: This stage will create a task that will load the target table on the schedule specified

If a task is part of a DAG of tasks, the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

The task node has no ALTER capabilities. All task enabled nodes are CREATE OR REPLACE only though this is subject to change.

Iceberg tables with Task Redeployment

Recreating a Snowflake Iceberg table

If any changes in config options like external volume,baselocation ,node properties ,column results in recreating iceberg table during redeployment

  • Create Iceberg Table: This stage will execute a CREATE OR REPLACE statement and create a Iceberg Table in the target environment.

Iceberg tables With Task Recreating the Task Redeployment

After the Task has deployed for the first time into a target environment, subsequent deployments with changes in task schedule, warehouse, or scheduling options will result in a CREATE TASK AND RESUME TASK statements being issued

The following stages are executed:

Iceberg tables With Task No Predecessor Task Redeployment

  • Create Task: This stage will create a task that will load the target table on the schedule specified.
  • Resume Task: After the task has been created it needs to be resume so that the task runs on the schedule.

Iceberg tables With Task Predecessor Task Redeployment

  • Suspend Root Task: To add a task into a DAG of task the root task needs to be put into a suspended state.
  • Create Task: This stage will create a task that will load the target table on the schedule specified

External Iceberg table Undeployment

If a snowflake iceberg table with task is dropped from the workspace and commited to GIT results in table and task dropped from target environment. This is executed in the below stages:

  • Drop Iceberg Table

Iceberg tables With Task No Predecessor Task Undeployment

  • Drop Iceberg Table
  • Drop Current Task: This stage will drop the task.

Iceberg tables With Task Predecessor Task Undeployment

  • Drop Iceberg Table
  • Suspend Root Task: To drop a task from a DAG of task the root task needs to be put into a suspended state.
  • Drop Task: This stage will drop the task.

If a task is part of a DAG of tasks the DAG needs to include a node type called Task Dag Resume Root. This node will resume the root node once all the dependent tasks have been created as part of a deployment.

Code

Code

Work with Task

Dimension with Task

Fact with Task

Task DAG Create Root

Task DAG Resume Root

Stream

Stream and Insert or Merge

Macros

Versions

Available versions of the package.

Version #Release DateNotes
1.1.4July 31, 2024
 New node type Iceberg table with Task added 
1.1.3July 29, 2024
 Fix for empty current storage location 
1.1.2July 08, 2024
 Support for external iceberg tables in Stream node type 
1.1.1June 28, 2024
 Stream Insert or Merge materialization type added 
1.1.0June 13, 2024
 Default options modified 
1.0.0May 15, 2024
 Initial Version 

Support

If you need help, please see our support section or contact us.