Job Description

Role Overview :

We are seeking a Senior Data Engineer with strong developer DNA — someone who codes daily in PySpark, SQL, and Python, builds end-to-end data solutions, and leads design of robust, production-grade data platforms on Microsoft Fabric, Azure Databricks, and ADF. This is a hands-on leadership role for engineers who can think architecturally and code efficiently.


Key Responsibilities:

• Design, develop, and optimize complex data pipelines and lakehouse architectures.

• Lead development of batch and streaming ingestion frameworks using ADF, Databricks, and Event Hubs.

• Write optimized PySpark/SQL transformations for performance, scalability, and cost efficiency.

• Build reusable data frameworks and automation scripts for continuous integration and deployment.

• Establish unit testing, code review, and versioning standards across engineering teams.

• Integrate with DevOps CI/CD pipelines for automated releases and environment management.

• Mentor junior engineers on coding practices and architecture design.

• Collaborate with data architects on Fabric-based Lakehouse implementations.


Skills & Qualifications:

• 6–12 years of strong hands-on experience in data engineering and development.

• Expertise in:

o PySpark, SQL, Python, and performance tuning of large-scale transformations.

o Databricks, ADF, ADLS, Fabric Data Pipelines, and Delta Lake.

• Deep understanding of data architecture, schema design, job orchestration, and data governance. • Strong knowledge of CI/CD (Azure DevOps) and infrastructure-as-code for data pipelines.

• Optional: Exposure to GCP BigQuery, Dataproc, or Dataflow as secondary platforms.


Preferred Certifications: • Databricks Certified Data Engineer Professional • Microsoft Certified: Fabric Analytics Engineer Associate (DP-600) • Microsoft Certified: Azure Solutions Architect Expert (AZ-305) • Optional: Google Cloud Certified – Professional Data Engineer

Apply for this Position

Ready to join ? Click the button below to submit your application.

Submit Application