Job Description

Job Title: Senior Data Engineer

Location: IN-Hyderabad


Overview:

We are seeking a talented Senior Data Engineer specializing in Starburst (Trino) and Dell Data Lakehouse to join our AI & Data team. You will be responsible for deploying, maintaining and optimizing Starburst installations & Dell Data Lakehouse, enabling our clients to seamlessly access their data across multiple platforms. The ideal candidate will have excellent communication skills, an advanced understanding of Starburst & Dell Data Lakehouse, and proficiency with troubleshooting and root cause analysis.


Responsibilities:

• Deploy and manage Starburst Enterprise/Galaxy and Dell Data Lakehouse installations, overseeing environment setup, configuration, maintenance, upgrades, and ensuring optimal performance.

• Configure various server and application settings and parameters.

• Integrate Starburst with various data sources to create a unified data platform.

• Design and tune the container solution for performance and scalability.

• Set up and configure data catalogs in various modes.

• Implement robust security controls for data access, ensure compliance with data regulations, and manage potential vulnerabilities.

• Coordinate with various support partners and vendor teams.

• Troubleshoot and investigate server related issues and provide root cause analysis for incidents.

• Perform daily server administration and monitoring, and leverage automation (such as Ansible) for efficient maintenance.

• Plan and execute disaster recovery testing.

• Create documentation and provide training on Starburst administration and best practices.


Qualifications:

• Required Skills & Experience:

• Bachelor’s degree in Computer Science, Information Systems, Data Science, Engineering or related field (or equivalent work experience).

• Proven experience with Trino/Starburst Enterprise/Galaxy administration / CLI.

• Implementation experience with container orchestration solutions (Kubernetes/OpenShift).

• Knowledge of Big Data (Hadoop/Hive/Spark) and Cloud technologies (AWS, Azure, GCP).

• Understanding of distributed system architecture, high availability, scalability, and fault tolerance.

• Familiarity with security authentication systems such as LDAP, Active Directory, OAuth2, Kerberos.

• Excellent Unix/Linux skills.

• Familiarity with JDBC / ODBC

• Preferred Skills:

• Certification: Starburst Certified Practitioner.

• Experience Python and/or Java programming.

• Proficient with infrastructure automation tools such as Ansible.

• Knowledge of data requirements for AI and machine learning workloads.

• Familiarity with Data Federation and Cached Services

• Familiarity with Data pipeline (Series of steps that move and transform data from one source to another for analyses and storage)

• Experience with Dell Data Lakehouse administration.

• Experience in Demand Driven Adaptive Enterprise (DDAE) administration


• Working Conditions

• This position may require evening and weekend work for time-sensitive project implementations.

Apply for this Position

Ready to join ? Click the button below to submit your application.

Submit Application