Job Description
Position Requirement:
• 4+ years of experience using Data Integration Tools - Pentaho Or any other ETL/ELT tools.
• 4+ years of experience using traditional databases like Postgres, MSSQL, Oracle
• 1+ years of experience using Columnar databases like Vertica, Google BigQuery, Amazon Redshift
• 1+ years of experience in Scheduler/Orchestration Tools Like Control-M, Autosys, Airflow, JAMS
• Good conceptual knowledge on ETL/ELT Strategies.
• Good conceptual knowledge in any Code Versioning Tools
• Good collaboration, communication and documentation skills.
• Experience of working in Agile Delivery Model.
• Requires minimal or no direct supervision
• Good knowledge in Data Visualization Tools like Tableau, Pentaho BA Tools.
• Digital Marketing/Web analytics or Business Intelligence a plus.
• Knowledge of scripting languages such as Python.
• Experience in the Linux environment is preferred but not mandatory.
Roles & Responsibilities:
• Develop & Support multiple Data Engineering projects with heterogeneous data sources,
produce/consume data to/from messaging queues like Kafka, push/pull data to/from REST API’s.
• Support in-house build Data Integration Framework, Data Replication Framework, Data Profiling &
Reconciliation Framework.
• Develop Data Pipelines with good coding standards, unit testing with detailed test cases.
• Willingness to learn new technologies.
Qualification: B.E. Computer Science/IT degree (or any other engineering discipline)
Work Timings: 2 PM to 11 PM IST
Apply for this Position
Ready to join ? Click the button below to submit your application.
Submit Application