Job Description
- Design, build, and maintain scalable data pipelines and ETL processes using Python and distributed processing frameworks on Databricks and cloud platforms
- Ingest, normalize, and enrich large volumes of structured and unstructured data including market feeds, vendor datasets, documents, and alternative data
- Collaborate with AI engineers, data scientists, and software engineers to define data schemas, interfaces, and APIs that support model development and production services
- Implement automated data validation, testing, monitoring, and alerting to ensure data quality, freshness, and drift detection for offline and online workloads
- Optimize pipeline performance and cost through efficient use of distributed compute, storage patterns, and cloud resource management
- Enforce data governance, access controls, and privacy standards while maintaining metadata and end-to-end data lineage using data cataloging tools
- Support onboar...
Apply for this Position
Ready to join Point72? Click the button below to submit your application.
Submit Application