Job Description

Job Role: -AWS Data Engineer


Job Location: -Pune/Hyderabad/Chennai/Mysuru/Bhubaneswar/Mangalore/Trivandrum/Chandigarh/Jaipur/Nagpur/Indore/Gurgaon


Experience: - 7+ Years


Job Roles & Responsibilities: -


  • Design, develop, and maintain data pipelines and assets on AWS.
  • Optimize and refactor legacy PySpark / Spark SQL code for performance and maintainability.
  • Implement unit testing / TDD to ensure robust, bug-free code.
  • Debug and resolve complex performance, concurrency, and logic issues.
  • Manage code versioning and repositories (Git, JFrog Artifactory).
  • Leverage AWS services (S3, EC2, Lambda, Redshift, CloudFormation) for scalable data solutions.


Job Skills & Requirements: -


  • 7+ years hands-on experience in Python, PySpark, Boto3, and related frameworks/libraries.
  • Proven expertise in Spark SQL & PySpark optimization.
  • Strong knowledge of AWS architecture (S3, EC2, Lambda, Redshift, CloudFormation).
  • Experience in code refactorization for clean, maintainable solutions.
  • Familiarity with Git, JFrog Artifactory, and modern CI/CD practices.
  • Strong debugging and problem-solving skills.
  • Solid understanding of unit testing and TDD methodologies.


Apply for this Position

Ready to join ? Click the button below to submit your application.

Submit Application