> job detail
D
⚙️Data Engineer
Founding Sr. Data Engineer
Diligent Robotics · Anywhere in the US
// classified as
Data Engineer (Pipelines, infra, ingestion, ETL.)
posted
2d ago
location
Anywhere in the US
languages
sql
tools
airbyte, bigquery, dbt
> stack
sqlairbytebigquerydbtfivetranredshiftsnowflakedbt
> education
master
> description
<p><strong>What we’re doing isn’t easy, but nothing worth doing ever is. </strong></p>
<p>At Diligent Robotics, we envision a future powered by robots that work seamlessly with human teams. We build artificial intelligence that enables service robots to collaborate with people and adapt to dynamic human environments. Our robots operate in complex environments such as hospitals, supporting staff so they can focus on higher-value work and patient care.</p>
<p>We're hiring our <strong>Founding Senior Data Engineer</strong> — the first dedicated data hire on a team that already has active analytics demand, working dashboards, and a clear plan for what comes next. You'll own the design and build of our BigQuery data warehouse, set the standards the rest of the data function will follow, and partner closely with our existing analyst team to build the analytics platform that will power the next stage of the business.</p>
<p>This is a hands-on senior IC role with significant architectural scope. You will set the long-term shape of our analytics platform and write much of the code that gets us there.</p>
<p><strong>Key Responsibilities</strong></p>
<ul>
<li><strong>Stand up the warehouse.</strong> Design and implement our BigQuery environment from the ground up — project structure, IAM, dataset organization, naming conventions, cost controls — so it scales cleanly as we add sources and users.</li>
<li><strong>Build the transformation layer.</strong> Set up dbt as the home for our transformation logic. Migrate the business logic from our existing analytics environment into version-controlled, tested dbt models.</li>
<li><strong>Model master data.</strong> Introduce dimensional models for our core business entities so data from different sources can be joined consistently.</li>
<li><strong>Own ingestion.</strong> Configure native Firestore-to-BigQuery integration, then stand up Airbyte as our managed connector layer for SaaS sources and log exports as we broaden the warehouse.</li>
<li><strong>Partner across the org.</strong> Work with source-system owners, the analyst team, and engineering teams that will eventually consume curated data from the warehouse. Help define the boundary between what the warehouse is and is not for.</li>
</ul>
<p><strong>Basic Qualifications</strong></p>
<ul>
<li>5+ years of hands-on data engineering experience, including significant work in a modern data stack (cloud warehouse + dbt + version-controlled transformations).</li>
<li>Deep, hands-on experience with <strong>BigQuery</strong> (or another columnar cloud warehouse — Snowflake, Redshift — with willingness to ramp on BigQuery quickly).</li>
<li>Deep, hands-on experience with <strong>dbt</strong>, including project structure, testing, and modeling patterns. You can debate the tradeoffs of staging vs. intermediate vs. marts layering and have opinions about when to use incremental models.</li>
<li>Strong dimensional modeling fundamentals (Kimball or equivalent). You understand conformed dimensions, slowly changing dimensions, and grain.</li>
<li>Strong SQL. Comfortable reviewing and refactoring complex SQL written by analysts.</li>
<li>Experience designing ingestion pipelines from multiple source types — operational databases, SaaS APIs, log-based sources.</li>
<li>Excellent technical communication. You can explain tradeoffs and modeling decisions to analysts, engineers, and executives.</li>
<li>Comfortable being the first dedicated data engineer in an organization — setting standards, making decisions without a peer to defer to, partnering with non-data folks.</li>
</ul>
<p><strong>Preferred Qualifications</strong></p>
<ul>
<li>Experience with <strong>Firestore</strong> and the native Firestore-to-BigQuery integration, particularly at scale (multi-project setups).</li>
<li>Experience with <strong>Airbyte</strong> (self-hosted or cloud) or comparable managed ingestion tools (Fivetran, Stitch).</li>
<li>Experience working with <strong>Elastic</strong> as a data source — extracting metrics or operational data out of Elastic indexes into downstream systems.</li>
<li>Experience designing a <strong>metrics catalog</strong> or semantic layer (dbt Semantic Layer, MetricFlow, LookML, or similar).</li>
</ul>