Job Description
Job Responsibilities
- Collaborate with project stakeholders (client) to identify product and technical requirements.
- Develop, implement, and tune large-scale distributed systems and pipelines that process large volume of data
- Write clean, maintainable, and testable code for data workflows.
- Troubleshoot data issues and perform root cause analysis
Must have:
- 3+ years of Hands-on coding experience in PySpark & SQL .
- Excellent verbal and business communication skills.
- Worked on complex SQL Query and has work on query performance optimization
Good to have:
- Experience working on large-scale data warehouse projects; Teradata experience is a plus.
- Experience with ETL tools.
- Experience working with workflow scheduler tools; experience with Apache Airflow is a plus.
Apply for this Position
Ready to join EXL? Click the button below to submit your application.
Submit Application