Job Description

Purpose of the Position:

You will be a critical member of the InfoCepts Cloud Data Architect Team. This position requires a sound knowledge on Big Data using Cloud technologies such as such as Databricks, EMR, Athena, PySpark, S3, AWS Lambda etc. Strong foundation on database concepts and SQL is also required.

/ Bangalore/Chennai

Type of Employment: Full time

Key Result Areas and Activities:

Technology Assessment and Design

  • Study existing technology landscape and understand current data integration frameworks and do impact assessment for the requirements.
  • Should be able to design complex Big Data use cases using AWS Services and Databricks, under guidance of Architect.
  • Ensure optimal balance between cost and performance.

Documentation and Stakeholder Communication

  • Project documentation, adheres to quality guidelines & schedules.
  • Works hand in hand with Architect & PM for successful delivery of project and provide estimation, scoping, scheduling assistance.
  • Articulate design decisions very clearly with stakeholders.
  • Perform Proof-of-Concepts and document all observations before proposing a new solution.
  • Conduct design review sessions with other teams and identify scope of improvement.

Process Improvement and Automation

  • Suggests automation to improve existing processes.
  • Assist junior Data Engineers by providing expert advice or troubleshooting steps whenever required.
  • Create technology-focused training plans whenever required.
  • Deliver technology-focused training sessions with team members whenever required.
  • Conduct Expert Knowledge Sharing sessions with Client Stakeholders whenever required.
  • Assist in designing case study documents whenever required.

Roles and Responsibility

Work and Technical Experience:

Must-Have:

  • In-depth knowledge of the following AWS services: S3, EC2, EMR, Athena, AWS Glue, Lambda
  • Experience with at least one MPP database: AWS Redshift, Snowflake, SingleStore
  • Proficiency in Big Data technologies: Apache Spark, Databricks
  • Must have strong programming skills in Python
  • Responsible for building data pipelines in AWS And Databricks
  • Experience with Big Data table formats, such as Delta Lake (open source)
  • Must have very strong SQL skills
  • Experience with orchestration tools like Apache Airflow
  • Expertise in developing ETL workflows with complex transformations such as SCD, deduplications, aggregations, etc.

Good to Have:

  • Cloud Databases - Snowflake, AWS Aurora
  • Big Data - Hadoop, Hive
  • Cloud Databases- AWS Aurora
  • Associate Level or Professional Level AWS Certification or Databricks Certification

Qualifications:

  • Overall 7+ years of IT Exp
  • 5+ years of relevant experience in AWS related project
  • Bachelor's degree in computer science, engineering, or related field (master's degree is a plus)

Qualities:

  • Strong hold technical knowledge and experience.
  • Should have the capability to deep dive and research in various technical related fields
  • Self-motivated and focused on delivering outcomes for a fast-growing team and firm.
  • Prior experience of working in a large media company would be added advantage

Apply for this Position

Ready to join ? Click the button below to submit your application.

Submit Application