Databricks - Run Job

Overview

You can use this Snap to automate the execution of a set of tasks or processes within a Databricks workspace. It triggers the task and periodically checks its progress. The Snap stops after the job is complete, but if you cancel the pipeline before the task finishes, the Snap requests to terminate the task.



Prerequisites

  • Valid client ID.
  • A valid account with the required permissions.

Snap views

View Description Examples of upstream and downstream Snaps
Input Requires a valid task name, notebook path, and cluster-info.
Output Executes the selected notebook.
Error

Error handling is a generic way to handle errors without losing data or failing the Snap execution. You can handle the errors that the Snap might encounter when running the pipeline by choosing one of the following options from the When errors occur list under the Views tab. The available options are:

  • Stop Pipeline Execution Stops the current pipeline execution when an error occurs.
  • Discard Error Data and Continue Ignores the error, discards that record, and continues with the remaining records.
  • Route Error Data to Error View Routes the error data to an error view without stopping the Snap execution.

Learn more about Error handling in Pipelines.

Snap settings

Legend:
  • Expression icon (): Allows using pipeline parameters to set field values dynamically (if enabled). SnapLogic Expressions are not supported. If disabled, you can provide a static value.
  • SnapGPT (): Generates SnapLogic Expressions based on natural language using SnapGPT. Learn more.
  • Suggestion icon (): Populates a list of values dynamically based on your Snap configuration. You can select only one attribute at a time using the icon. Type into the field if it supports a comma-separated list of values.
  • Upload : Uploads files. Learn more.
Learn more about the icons in the Snap settings dialog.
Field / Field set Type Description
Label String

Required. Specify a unique name for the Snap. Modify this to be more appropriate, especially if more than one of the same Snaps is in the pipeline.

Default value: Databricks - Run Job

Example: Run Job
Task name String/Expression Required. Specify the name of the task to perform the job.

Default value: N/A

Example: Test username and password

Notebook path String/Expression/ Suggestion Required. Specify the path of the saved notebook that will run in this job. Notebook is a web-based interface that allows you to create, edit, and execute data science and data engineering workflows.

Learn more about Databricks notebooks.

Default value: N/A

Example: /Users/[email protected]/notebook

Cluster String/Expression/ Suggestion Required. Specify the cluster to run the job within its environment.

Default value: N/A

Example: Code Ammonite - Shared Compute Cluster - V2

Parameters Use this field set to specify the parameters to run the job.
Key String/Expression Required. Specify the parameter key.

Default value: N/A

Example: Age

Value String/Expression Required. Specify the parameter value.

Default value: N/A

Example: 35

Interval check (seconds) Integer/Expression Required. Specify the number of seconds to wait before checking the status of the task.

Default value: 10

Example: 15

Snap execution Dropdown list
Choose one of the three modes in which the Snap executes. Available options are:
  • Validate & Execute: Performs limited execution of the Snap and generates a data preview during pipeline validation. Subsequently, performs full execution of the Snap (unlimited records) during pipeline runtime.
  • Execute only: Performs full execution of the Snap during pipeline execution without generating preview data.
  • Disabled: Disables the Snap and all Snaps that are downstream from it.

Default value: Execute only

Example: Validate & Execute

Examples