Job Description
Role: Senior Data Engineer - Azure
Experience Level: 4 to 6 Years
Work location: Mumbai, Bangalore, Trivandrum (Hybrid)
Notice Perid: 0-30 days
Key Responsibilities:
● Design and build scalable ETL/ELT pipelines using Azure Data Factory (ADF), Azure Databricks (Spark), and Azure Synapse Analytics.
● Mandatory hands-on experience with Microsoft Azure and Databricks.
● Understanding of the Medallion Architecture (specifically the Bronze/Raw layer) and data modeling concepts for data lakes.
● Experience building robust Orchestration pipelines (scheduling, error handling, logging) and managing full/incremental data loads.
● Knowledge of Delta format best practices, including partitioning, Z-ordering, vacuuming, and storage tiering for cost efficiency.
● Familiarity with CI/CD pipelines, automated deployment scripts, version control (Git), and Unit/Integration testing frameworks.
● Experience implementing Data Quality rules, PII handling/masking, and metadata management.
● Develop and optimize PySpark/Spark SQL jobs for large-scale batch and streaming data transformations.
● Ingest data from various sources including Apache Kafka, REST APIs, and RDBMS, ensuring real-time or near-real-time availability.
● Implement data modeling strategies (star schema, snowflake schema) for analytics consumption layers in Synapse or ADLS.
● Collaborate with DevOps teams to automate deployment using CI/CD pipelines (Azure DevOps, GitHub Actions, etc.).
● Monitor, troubleshoot, and optimize data workflows for performance, cost-efficiency, and reliability.
● Follow coding standards, participate in peer reviews, and maintain version-controlled code in Git repositories.
● Support data quality checks, logging, alerting, and observability mechanisms for production workloads.
● Participate in sprint ceremonies and contribute to task estimation and delivery planning.
Must have skills:
● 3+ years of experience in data engineering roles.
● Hands-on experience with:
○ Azure Data Factory (ADF) – building pipelines, triggers, linked services.
○ Azure Databricks – building and managing Spark jobs in PySpark.
○ Azure Synapse Analytics – data warehousing, SQL queries, workspace orchestration.
○ Apache Kafka – consuming and processing real-time data streams.
● Strong in SQL, Python, and Spark for data manipulation and transformation.
● Exposure to CI/CD practices (Azure DevOps, Git workflows, build/release pipelines).
● Understanding of data lake architecture and modern data warehousing principles.
Apply for this Position
Ready to join ? Click the button below to submit your application.
Submit Application