Job Description
Job Description– Big Data Engineer
Experience: 8+ Years
Location: Chennai & Gurgaon
Mode: Hybrid
We are looking for a Big Data Engineer with strong experience in SQL, Hive, ETL pipelines, PySpark, and GCP to design and build scalable data solutions for large, complex datasets.
Key Responsibilities
Develop and optimize Big Data pipelines using SQL, Hive, PySpark, and ETL frameworks.
Build and maintain scalable data solutions on GCP (BigQuery, BigTable, Dataflow, Dataproc, etc.).
Design and implement data models for analytical and operational systems.
Work with diverse storage systems — relational, NoSQL, document, column-family, and graph databases.
Ensure high performance, reliability, data quality, and secure data management.
Optimize SQL queries and improve performance across distributed systems.
Collaborate with cross-functional teams following Agile/Scrum methodologies.
Required Skills
Strong hands-on experience with SQL , Hive ...
Experience: 8+ Years
Location: Chennai & Gurgaon
Mode: Hybrid
We are looking for a Big Data Engineer with strong experience in SQL, Hive, ETL pipelines, PySpark, and GCP to design and build scalable data solutions for large, complex datasets.
Key Responsibilities
Develop and optimize Big Data pipelines using SQL, Hive, PySpark, and ETL frameworks.
Build and maintain scalable data solutions on GCP (BigQuery, BigTable, Dataflow, Dataproc, etc.).
Design and implement data models for analytical and operational systems.
Work with diverse storage systems — relational, NoSQL, document, column-family, and graph databases.
Ensure high performance, reliability, data quality, and secure data management.
Optimize SQL queries and improve performance across distributed systems.
Collaborate with cross-functional teams following Agile/Scrum methodologies.
Required Skills
Strong hands-on experience with SQL , Hive ...
Apply for this Position
Ready to join BCforward? Click the button below to submit your application.
Submit Application