Job Description

What you ’ll do

- Design, build, and maintain scalable data pipelines and ETL processes using Python and distributed processing frameworks on Databricks and cloud platforms

- Ingest, normalize, and enrich large volumes of structured and unstructured data including market feeds, vendor datasets, documents, and alternative data

- Collaborate with AI engineers, data scientists, and software engineers to define data schemas, interfaces, and APIs that support model development and production services

- Implement automated data validation, testing, monitoring, and alerting to ensure data quality, freshness, and drift detection for offline and online workloads

- Optimize pipeline performance and cost through efficient use of distributed compute, storage patterns, and cloud resource management

- Enforce data governance, access controls, and privacy standards while maintaining metadata and end-to-end data lineage using data cataloging tools

- Support onboar...

Apply for this Position

Ready to join Point72? Click the button below to submit your application.

Submit Application