5+ years of experience as a Data Engineer or similar role.
Strong proficiency in Python and SQL.
Hands-on experience building data pipelines using tools like Airflow, Spark, or similar.
Experience with data warehouses and databases (e.g., BigQuery, Snowflake, ClickHouse).
Solid understanding of ETL/ELT, data modeling, and schema design.
Experience working at fast growing b2c startup.
Experience working with high-scale, production data systems.
Willingness to work in a fast-paced startup environment.
Bonus points for experience with real-time data, ML pipelines, or GenAI systems.
This role is based in the San Francisco Bay Area.
Responsibilities
Own the Source of Truth by translating product- and finance-defined metrics into durable warehouse models ensuring consistent usage across the company.
Eliminate conflicting logic across dashboards and teams to ensure executive- and board-level reporting is powered by consistent, traceable warehouse logic.
Build Guardrails by maturing existing BigQuery-based stacks through centralizing business logic in the warehouse with enforced governance controls.
Implement automated validation and reconciliation between product systems, billing systems, and financial reporting to prevent silent metric drift and post-close surprises.
Ensure new products, models, and features are correctly integrated into reporting for accurate company metrics representation.
Enable the Company by partnering with various departments (Product, Engineering, ML, Growth, Finance) to deliver trusted data that supports decision-making processes.
Improve documentation, testing, and transparency across warehouse systems to reduce key-person risk and increase confidence in company metrics.