Job Description

About The Position

Department: Sales and Delivery Team - Empower

Industry: Information Technology & Services, Computer Software, Management Consulting

Location: WFH/ India Remote

Experience Range : 6 - 10 years

Basic Qualification: Bachelor of Engineering or Equivalent

Travel Requirements: Not required

Website :

Exusia, a cutting-edge digital transformation consultancy, and is looking for top talent in DWH & Data Engineering space with specific skills in Snowflake/Python/DBT to join our global delivery team's Industry Analytics practice in India.

What’s the Role?

· Full-time job to work with Exusia's clients in the United States to deliver on bleeding-edge data-driven solutions

· Developing and managing large scale data pipelines and data repositories

· Collaborate with Product Owners, Solution Architects to develop optimized data engineering solutions.

Criteria for the Role!

· Minimum 6 years of experience working as a Data engineer

· Min 2 years exp in Snowflake and DBT

· Master of Science (preferably in Computer and Information Sciences or Business Information Technology) or an Engineering degree in the above areas.

· Excellent communication skills and should be able to work directly with business stakeholders, creative problem solver, flexible, proactive, attitude to learn newer tools/technologies

Responsibilities

· Implementing end to end data pipelines to move data from source systems into data lake or data warehouse

· Build pipeline automation and orchestration process

· Develop Snowflake data models (, star schema, snowflake schema) for optimized query performance.

· Working with Data Analysts to ensure the pipelines are tested and optimised to provide accurate and timely data

· Working in an agile software delivery model and manage changing requirements and priorities during the SDLC 

Requirements

Mandatory Skills

· Develop and maintain Snowflake data models (, star schema, snowflake schema) for optimized query performance.

· Create and maintain Snowpipe & SnowSQL scripts for data loading, data transformations, and data retrieval.

· Proficiency in SQL for data manipulation, transformation, and processing.

· Expertise in DBT to develop modular, scalable, and well-documented data pipelines.

· Strong python programming experience to support data processing and automation.

· Hands-on experience with Airflow for orchestrating data pipelines

· Knowledge of cloud platforms specifically storage and databases to source or stage data for Snowflake

· Problem-solving skills with the ability to work with large datasets and debug data pipeline issues.

Nice to have skills

· Understanding of Data Modeling & legacy ETL technologies

· Prior migration experience - On Prem legacy databases to snowflake

· Knowledge of Spark / Spark based data processing tools

· Exposure to one or more Cloud platforms – Amazon Web Services (AWS), Microsoft Azure, Google Cloud Platform

· Exposure to Data Governance aspects like Metadata Management, Data Dictionary, Data Glossary, Data Lineage

Apply for this Position

Ready to join ? Click the button below to submit your application.

Submit Application