Job Description
We are seeking a highly skilled Data Engineer with expertise in ETL, PySpark, AWS, and big data technologies. The ideal candidate will have in-depth knowledge of Apache Spark, Python, and Java programming (Java 8 and above, including Lambda, Streams, Exception Handling, Collections, etc.). This role involves designing and developing scalable data processing pipelines for batch and real-time analytics.
Key Responsibilities
- Develop data processing pipelines using PySpark.
- Create Spark jobs for data transformation and aggregation.
- Optimize query performance using file formats like ORC, Parquet, and AVRO.
- Design scalable pipelines for both batch and real-time analytics.
- Perform data enrichment and integrate with
Apply for this Position
Ready to join Evnek? Click the button below to submit your application.
Submit Application