Skip to content

Efrat19/databricks-runs-exporter

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

44 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

databricks-runs-exporter

Prometheus Exporter for databricks runs.

Installation

# Create the credentials secret 
kubectl create secret generic databricks-creds -n monitoring \ 
--from-literal=host=<databricks_host>  --from-literal=token=<databricks_token>

# Apply the exporter
kubectl apply -f https://raw.githubusercontent.com/Efrat19/databricks-runs-exporter/main/k8s.yaml

Metrics

runs_total

  • Type: Count
  • Namespace: databricks_exporter
  • Description: runs info from databricks
  • Labels: job_id, run_id, number_in_job, original_attempt_run_id, life_cycle_state, result_state, user_cancelled_or_timedout, start_time, setup_duration, execution_duration, cleanup_duration, end_time, trigger, creator_user_name, run_name, run_type, attempt_number, format

Env Variables

Variable Default Description
DATABRICKS_HOST "" Databricks host URL, required
DATABRICKS_TOKEN "" Databricks token with permissions to list jobs, required
RUNS_SCRAPE_LIMIT 50 Max runs to scrape on each interval. Increase if you have more then $RUNS_SCRAPE_LIMIT runs created within a given $MAX_RUN_DURATION_SECONDS
RUNS_SCRAPE_TIMESPAN_SECONDS 10 Match this value to your scrape interval (which is usually 10s)
MAX_RUN_DURATION_SECONDS 3600 The duration of your slowest job, so the exporter knows how long ago to look for it. defaults to 1h

Grafana Dahsboard

dashboard.json

Credits

Awesome exporter template from go-gywn/query-exporter-simple