Job Description

Job Responsibilities

  • Collaborate with project stakeholders (client) to identify product and technical requirements.
  • Develop, implement, and tune large-scale distributed systems and pipelines that process large volume of data
  • Write clean, maintainable, and testable code for data workflows.
  • Troubleshoot data issues and perform root cause analysis

Must have:


  • 3+ years of Hands-on coding experience in PySpark & SQL.
  • Excellent verbal and business communication skills.
  • Worked on complex SQL Query and has work on query performance optimization



Good to have:

  • Experience working on large-scale data warehouse projects;
    Teradata experience is a plus.
  • Experience with ETL tools.
  • Experience working with workflow scheduler tools;
    experience with Apache Airflow is a plus.
  • Working experience with Kubernetes, Unix, github

Apply for this Position

Ready to join ? Click the button below to submit your application.

Submit Application