Job Description

Position Description:


We are seeking a skilled Azure Databricks Developer to design, develop, and optimize big data pipelines using Databricks on Azure. The ideal candidate will have strong expertise in PySpark, Azure Data Lake, and data engineering best practices in a cloud environment.

Your future duties and responsibilities:

Key Responsibilities:
Design and implement ETL/ELT pipelines using Azure Databricks and PySpark.
Work with structured and unstructured data from diverse sources (e.g., ADLS Gen2, SQL DBs, APIs).
Optimize Spark jobs for performance and cost-efficiency.
Collaborate with data analysts, architects, and business stakeholders to understand data needs.
Develop reusable code components and automate workflows using Azure Data Factory (ADF).
Implement data quality checks, logging, and monitoring.
Participate in code reviews and adhere to software engineering best practices.
Required Skills & Qualifications:
5+ years of experience in Apache Spark / PySpark.
5+ years working with Azure Databricks and Azure Data Services (ADLS Gen2, ADF, Synapse).
Strong understanding of data warehousing, ETL, and data lake architectures.
Proficiency in Python and SQL.
Experience with Git, CI/CD tools, and version control practices.

Required qualifications to be successful in this role:

Required Qualifications
Hands-on experience with Databricks and Apache Spark
Strong proficiency in Python and SQL
Experience with Delta Lake, Parquet, and data modeling
Familiarity with cloud platforms (AWS, Azure, or GCP)
Experience with workflow orchestration tools (Airflow, Databricks Jobs, etc.)

Skills:

  • ETL
  • SQL
  • Apply for this Position

    Ready to join ? Click the button below to submit your application.

    Submit Application