Crossing Hurdles is a referral partner. We connect & refer top candidates to Mercor that working with the world’s leading AI research labs to help build and train cutting-edge AI models.
Role Responsibilities (Training support will be provided)
Build robust data pipelines to ingest, transform, and consolidate data from diverse sources including MongoDB, Airtable, PostHog, and production databases
Design dbt models and transformations to unify disparate data tables into clean, production-ready schemas
Implement scalable, fault-tolerant workflows using Fivetran, dbt, SQL, and Python
Collaborate with engineers, data scientists, and business stakeholders to ensure data availability, accuracy, and usability
Own data quality and reliability across the entire data stack from ingestion to consumption
Continuously improve pipeline performance, monitoring, and scalability
Requirements
Proven experience in data engineering with strong proficiency in SQL and Python
Experience using modern data stack tools like Fivetran, dbt, Snowflake, or similar platforms
Skilled in building and maintaining large-scale ETL/ELT pipelines across heterogeneous data sources and platforms
Strong understanding of data modeling, schema design, and transformation best practices
Familiarity with data governance, monitoring, and quality assurance tools and processes
Ability to work effectively across engineering, product, and operations teams
Bonus: prior experience supporting machine learning workflows or analytics platforms
Access to a desktop or laptop computer (Chromebooks are not supported)
Application Process (15 min)
Upload resume
Participate in AI interview (10 min)
Submit form