Job Description

Senior Data Scientist: AI Training Data (2-4 Months Contract)

Company: BespokeLabs (VC-backed; founded by IIT & Ivy League alumni)

Location: Remote

Role Type: Contract (2-4 Months)

Time Commitment: 40 hrs/week (Full-time availability required)

Compensation: Hyper-competitive hourly rate (matching top-tier Senior Data Scientist bands) Experience: 6+ years


About BespokeLabs

BespokeLabs is a premier, VC-backed AI Research lab with an exceptionally talent-dense team of IIT and Ivy League alumni. We don’t just build tooling around AI—we build the massive-scale data systems and reasoning architectures that directly power next-generation models. Our research shapes the frontier of AI: we’ve published breakthroughs like GEPA, driven foundational datasets like OpenThoughts, and shipped state-of-the-art models including Bespoke-MiniCheck and Bespoke-MiniChart. More on our website bespokelabs.ai :)

Role Overview

We are looking for a high-impact Senior Data Scientist for an intensive, 2-month sprint. You will leverage your deep expertise in production-grade machine learning and applied statistics to develop the algorithms and logic that curate and evaluate datasets for advanced AI model training.

This is not a traditional model-building or research role. We need a seasoned practitioner who has already owned the end-to-end DS lifecycle at scale. You will use your intuition for feature engineering, statistical validity, and large-scale data processing to programmatically generate, shape, and validate AI training data.


What You Will Do (The Contract)

  • Algorithm Design: Design and implement custom statistical models and programmatic logic (e.g., anomaly detection, active learning, similarity scoring) to evaluate data quality, complexity, and redundancy at scale.
  • Hands-on At-Scale Coding: Write scalable PySpark and Python (NumPy/Pandas) code to apply these algorithms across massive datasets, translating experimental logic into reliable, large-scale workflows.
  • Metric Formulation: Develop custom quantitative metrics and heuristic benchmarks to rigorously assess the fidelity and suitability of data subsets for specific AI training objectives.
  • Validation & Iteration: Run high-speed validation cycles, analyzing the output of data-curation algorithms to diagnose skew, bias, or noise, and iteratively refining the logic.
  • High-Level Curation: Apply Senior-level domain expertise in predictive modeling and feature engineering to ensure the final training inputs meet the strict standards required for state-of-the-art ML systems.


What You Bring to the Table (Your Past Experience)

To be successful in this contract, you must have a track record of:

  • The End-to-End DS Lifecycle: Framing problems, modeling, validation, production, and iteration.
  • Production Ownership: Building and deploying ML and statistical models on large-scale datasets.
  • Large-Scale Data Processing: Working with Apache Spark to develop scalable feature pipelines and offline training workflows.
  • Experimentation: Designing and analyzing rigorous experiments (A/B tests, causal inference).
  • Impact: Translating complex model outputs into clear product and business decisions.


Required Qualifications (Non-Negotiable)

  • Experience: 6+ years as a Data Scientist or Applied Scientist.
  • Production Background: Proven ownership of models running in production environments.
  • Applied Statistics: Strong background in applied statistics and experimentation frameworks.

Core Technical Skills

  • Languages: Python (NumPy, Pandas, Scikit-learn, PyTorch / TensorFlow) and Strong SQL.
  • Big Data: Apache Spark (PySpark or Spark SQL) for large-scale data processing.
  • Methodologies: Feature engineering, model evaluation, statistical modeling, and hypothesis testing.


Strong Signals (Highly Valued)

  • Scale: Models trained on TB-scale datasets.
  • Domain Specificity: Experience in high-complexity domains such as: Recommendations, Pricing, Fraud / risk, Search / ranking, or Growth & experimentation.
  • Collaboration: Experience deploying models alongside data engineering pipelines.


Out of Scope (Who Should Not Apply)

  • BI / reporting-only roles
  • SQL-only analysts
  • Research-only ML roles with no production ownership
  • Early-career profiles

Apply for this Position

Ready to join ? Click the button below to submit your application.

Submit Application