Job Description
Job Responsibilities
- Collaborate with project stakeholders (client) to identify product and technical requirements.
- Develop, implement, and tune large-scale distributed systems and pipelines that process large volume of data
- Write clean, maintainable, and testable code for data workflows.
- Troubleshoot data issues and perform root cause analysis
Must have:
- 3+ years of Hands-on coding experience in PySpark & SQL.
- Excellent verbal and business communication skills.
- Worked on complex SQL Query and has work on query performance optimization
Good to have:
- Experience working on large-scale data warehouse projects;
Teradata experience is a plus. - Experience with ETL tools.
- Experience working with workflow scheduler tools;
experience with Apache Airflow is a plus. - Working experience with Kubernetes, Unix, github
Apply for this Position
Ready to join ? Click the button below to submit your application.
Submit Application