Job Description
Job Title: AWS Data Engineer
Experience: 6+ Years
Positions: multiple
Engagement Type: Contract
Location: Remote
Role Overview
We are looking for two highly skilled Data Engineers to join our team on a contract basis. The ideal candidates will have strong experience in designing, building, and maintaining scalable data pipelines and working with modern cloud-based data platforms. This role requires hands-on expertise in AWS, Apache Airflow, Snowflake, and GitHub.
Key Responsibilities
- Design, develop, and maintain robust and scalable data pipelines.
- Build and orchestrate ETL/ELT workflows using Apache Airflow.
- Ingest, process, and manage data using AWS services.
- Develop and optimize data models and transformations in Snowflake.
- Ensure data quality, reliability, and performance across pipelines.
- Collaborate with analytics, product, and business teams.
- Implement best practices for version control and CI/CD using GitHub.
- Monitor, troubleshoot, and resolve data pipeline issues.
- Document data workflows, architecture, and technical processes.
Required Skills & Experience
- 6+ years of experience in Data Engineering or similar roles.
- Strong hands-on experience with AWS (S3, EC2, Lambda, Redshift, etc.).
- Solid experience with Apache Airflow for workflow orchestration.
- Strong expertise in Snowflake for cloud data warehousing.
- Proficiency in GitHub for version control and collaboration.
- Advanced SQL skills and experience working with large datasets.
- Experience with ETL/ELT design patterns and data modeling.
- Good understanding of data quality, monitoring, and performance tuning.
Good to Have
- Experience with Python or PySpark.
- Exposure to streaming tools like Kafka/Kinesis.
- Experience with CI/CD and DevOps practices.
- Knowledge of data governance and security best practices.
Apply for this Position
Ready to join ? Click the button below to submit your application.
Submit Application