Job Description

Job Description:

Key responsibilities include the following:

  • Develop and maintain scalable data pipelines using Pyspark and proven experience as developer with expertise in PySpark.
  • Good to have knowledge on Ab Initio.
  • Experience with distributed computing and parallel processing.
  • Proficiency in SQL and experience with database systems.
  • Collaborate with data engineers and data scientists to understand and fulfil data processing needs.
  • Optimize and troubleshoot existing PySpark applications for performance improvements.
  • Write clean, efficient, and well-documented code following best practices.
  • Participate in design and code reviews.
  • Develop and implement ETL processes to extract, transform, and load data.
  • Ensure data integrity and quality throughout the data lifecycle.
  • Stay current with the latest industr...

Apply for this Position

Ready to join Virtusa? Click the button below to submit your application.

Submit Application