Job Description
Senior Data Engineer, Assurant, GCC-India
Reports To: Director of Product Engineering & Integrations.
Position Summary
We are seeking a highly skilled Senior Data Engineer to design, develop, and optimize data pipelines and cloud-based data solutions. The ideal candidate will have strong expertise in Azure Data Factory, Snowflake, and modern ETL/ELT practices, enabling scalable, secure, and high-performance data workflows. This role will collaborate closely with analytics and BI teams to deliver reliable data infrastructure supporting enterprise reporting and advanced analytics.
This position will be in Bangalore/Chennai/Hyderabad at our India location.
Work Time: 3:30 PM IST ~ 12:30 AM IST
What will be my duties and responsibilities in this job?
Data Engineering & ETL Development
Design, develop, and maintain ETL/ELT pipelines using Azure Data Factory, Snowflake Tasks, and Snowpipe for real-time and batch ingestion. Implement best practices for data modeling, transformation, and performance tuning within Snowflake. Build and manage pipelines connecting multiple structured and unstructured data sources across cloud and on-prem environments. Automate data quality checks, data lineage tracking, and error handling within ETL workflows. Snowflake Development & Optimization
Develop and maintain Snowflake schemas, views, stored procedures, and materialized views. Configure and optimize Snowpipe for continuous data loading. Utilize Snowsight for monitoring query performance, cost optimization, and workload analysis. Implement role-based access control and ensure data security in Snowflake. Azure & Cloud Integration
Integrate Azure Data Factory with other Azure services (Blob Storage, Synapse, Key Vault). Design scalable cloud architectures and orchestrate pipelines across hybrid environments. Implement CI/CD pipelines for data workflows using GitHub Actions. Analytics & Reporting Enablement
Collaborate with business analysts and BI teams to enable Power BI dashboards backed by optimized Snowflake data models. Create semantic models and data marts to support self-service analytics and reporting. Scripting & Automation
Develop Python scripts for automation, data processing, and custom integrations. Leverage Python-based frameworks (Pandas, PySpark, Airflow) to enhance pipeline capabilities. What are the requirements needed for this position?
Bachelor’s or Master’s degree in Computer Science, Data Engineering, or related field. 8+ years of experience in data engineering, ETL development, and cloud data platforms. Strong proficiency in Snowflake, Azure Data Factory, and Python. Experience with CI/CD, data security, and performance optimization. Familiarity with BI tools (Power BI, Looker, etc.) and data modeling best practices. Excellent problem-solving skills and ability to work in a fast-paced environment. What are the Preferred requirements needed for this position?
Knowledge of Airflow, PySpark, and data orchestration frameworks. Experience with real-time data ingestion and streaming architectures. Understanding of cost optimization in cloud environments.
Apply for this Position
Ready to join ? Click the button below to submit your application.
Submit Application