Databricks Pipeline Engineer
Listed on 2026-01-13
-
IT/Tech
Data Engineer, Data Science Manager, Cloud Computing
I am recruiting for a Databricks Pipeline Engineer to work on a remote basis.
The role falls inside IR35 so you will be required to work through an umbrella company for the duration of the contract.
Strong expertise in Databricks Pipelines.
You must have experience of designing, optimising and maintaining scalable Databricks pipelines (ETL, streaming, ML workflows).
You will be able to perform cluster and job performance tuning: optimise cluster sizing, caching, partitioning, and shuffle management.
You will monitor Spark job metrics, analyse logs, and identify bottlenecks in data throughput or latency.
You will also be able to implement cost-optimisation strategies for Databricks jobs and clusters using autoscaling and job consolidation.
You must have familiarity with orchestration tools:
Databricks Workflows, Airflow, or Azure Data Factory.
Proficiency in Python or Scala for data engineering and pipeline development is essential.
Hands-on experience with Azure, AWS, or multi-cloud Databricks deployments.
Knowledge of data storage layers (Azure Data Lake Storage, AWS S3) and performance trade-offs.
Experience of version control (Git, Git Hub Actions, Dev Ops pipelines) and CI/CD practices for Databricks.
Please apply ASAP to find out more.
To Search, View & Apply for jobs on this site that accept applications from your location or country, tap here to make a Search: