Job Description
Job Description– Big Data Engineer
Experience: 8+ Years
Location: Chennai & Gurgaon
Mode: Hybrid
We are looking for a Big Data Engineer with strong experience in SQL, Hive, ETL pipelines, Py Spark, and GCP to design and build scalable data solutions for large, complex datasets.
Key Responsibilities
- Develop and optimize Big Data pipelines using SQL, Hive, Py Spark, and ETL frameworks.
- Build and maintain scalable data solutions on GCP (Big Query, Big Table, Dataflow, Dataproc, etc.).
- Design and implement data models for analytical and operational systems.
- Work with diverse storage systems — relational, No SQL, document, column-family, and graph databases.
- Ensure high performance, reliability, data quality, and secure data management.
- Optimize SQL queries and improve performance across distributed systems.
- Collaborate with cross-functional teams following Agile/Scrum methodologies.
Required Skills
- Strong hands-on experience with SQL, Hive, Py Spark/Python, and Big Data ETL.
- Experience with GCP Big Data services (Big Query, Dataflow, Big Table, etc.).
- Knowledge of No SQL systems (HBase, Cassandra, Mongo DB, etc.).
- Understanding of distributed systems, data storage formats, and performance tuning.
- Experience with data modeling tools (Er Win, ER/Studio).
- Familiarity with data management concepts: replication, partitioning, encryption, high availability.
Nice to Have
- Experience with OLTP/OLAP systems.
- Understanding of infrastructure for performance tuning (e.g., Nutanix, network-attached storage).
- Exposure to Java or other programming languages.
Experience: 8+ Years
Location: Chennai & Gurgaon
Mode: Hybrid
We are looking for a Big Data Engineer with strong experience in SQL, Hive, ETL pipelines, Py Spark, and GCP to design and build scalable data solutions for large, complex datasets.
Key Responsibilities
- Develop and optimize Big Data pipelines using SQL, Hive, Py Spark, and ETL frameworks.
- Build and maintain scalable data solutions on GCP (Big Query, Big Table, Dataflow, Dataproc, etc.).
- Design and implement data models for analytical and operational systems.
- Work with diverse storage systems — relational, No SQL, document, column-family, and graph databases.
- Ensure high performance, reliability, data quality, and secure data management.
- Optimize SQL queries and improve performance across distributed systems.
- Collaborate with cross-functional teams following Agile/Scrum methodologies.
Required Skills
- Strong hands-on experience with SQL, Hive, Py Spark/Python, and Big Data ETL.
- Experience with GCP Big Data services (Big Query, Dataflow, Big Table, etc.).
- Knowledge of No SQL systems (HBase, Cassandra, Mongo DB, etc.).
- Understanding of distributed systems, data storage formats, and performance tuning.
- Experience with data modeling tools (Er Win, ER/Studio).
- Familiarity with data management concepts: replication, partitioning, encryption, high availability.
Nice to Have
- Experience with OLTP/OLAP systems.
- Understanding of infrastructure for performance tuning (e.g., Nutanix, network-attached storage).
- Exposure to Java or other programming languages.
Apply for this Position
Ready to join ? Click the button below to submit your application.
Submit Application