Jobs (Databricks)
A Databricks job can contain single or multiple tasks (a workflow of tasks) with complex dependencies. Tasks include Notebook job, Spark-submit job, Spark SQL job, and so on. Each task can run on a different cluster.
The Workflows > Jobs tab lists all the job workflows.
The Job workflow list provides the following details:
Columns | Description |
---|---|
Last Run Status | Status of the last run job. |
Job Name/ID | Name and unique ID of the job. After you click the Job Name/ID link, the selected job name is filtered on the Job Runs page. You can view the runs for the selected job. For information, see Job Runs. |
Workspace | Name of the workspace where the job is run. |
User | Name of the user who submitted the job. |
Start Time | Start time and day of the job. |
Run Count | The count of jobs runs. |
Duration | Total time taken by the app to complete the job.
|
Read | Total data read by the app.
|
Write | Total data written by the app.
|
Cost | An estimated cost incurred in running the app. This cost is calculated based on VM and DBU prices. |