Job Title: Data Engineer Company Name: Junction Job Details: $160k-$200k/yrRemoteFull,Time Job Url: https://hiring.cafe/viewjob/qmvp28dypowm6j22 Job Description: Posted 2mo agoData Engineer@ JunctionView All JobsWebsiteUnited States or United Kingdom$160k-$200k/yrRemoteFull TimeResponsibilities:Design pipelines, Operate ELT/ETL workflows, Create analytics-ready schemasRequirements Summary:Data engineer with solid software fundamentals; build, own, and scale reliable data pipelines and warehouse infrastructure. Proficient in Python/SQL; experience with BigQuery and GCP; familiarity with orchestration tools (Temporal, Airflow, Dagster); startup or small-team experience; strong communication and ownership.Technical Tools Mentioned:Python, SQL, BigQuery, GCP, Temporal, Airflow, Dagster, PostgreSQL, BigTable Healthcare is in crisis and the people behind the results deserve better. With data exploding across wearables, lab tests, and patient–doctor interactions, we’re entering an era where data is abundant.Junction is building the infrastructure layer for diagnostic healthcare, making patient data accessible, actionable, and automated across labs and devices. Our mission is simple but ambitious: use health data to unlock unprecedented insight into human health and disease.If you're passionate about how technology can supercharge healthcare, you’ll fit right in.Backed by Creandum, Point Nine, 20VC, YC, and leading angels, we’re working to solve one of the biggest challenges of our time: making healthcare personalized, proactive, and affordable. We’re already connecting millions and scaling fast.Short on time? Who you are: A data engineer with solid software engineering fundamentals who can build, own, and scale reliable data pipelines and warehouse infrastructure.Ownership: You’ll shape our data foundation from ingestion through transformation — and make it analytics-ready at scale.Salary: $160K - $200k+ equityTime zone: Preferably NYC; EST required.Why we need youJunction powers modern diagnostics at scale and as we grow, our platform is becoming increasingly data-intensive. The way we move, structure, and surface data directly affects our ability to support customers, deliver real-time insights, and unlock the next generation of diagnostics products.We’re hiring our first Data Engineer to take ownership of that foundation.Build and run pipelines that turn raw, messy healthcare data into clean, trusted, usable informationPower customer products, internal analytics, and the AI models behind our next wave of diagnosticsDesign how data flows through an entire diagnostics ecosystem — not just maintain ETLsBuild scalable, cloud-native pipelines on GCP and eliminate bottlenecks as we scaleHunt down edge cases, build guardrails for quality, and ship systems other engineers rely on dailyIf you love untangling complexity and building data systems that truly make an impact, you’ll fit right in — and the systems you build will unlock new products and accelerate everything we ship.What you’ll be doing day to dayDesigning and operating ingestion, transformation, and replication pipelines on GCPManaging orchestration and streamlining ELT/ETL workflows (e.g., Temporal)Creating clean, scalable, analytics-ready schemas in BigQueryImplementing monitoring, alerting, testing, and observability across data flowsIntegrating data from APIs, operational databases, and unstructured sourcesCollaborating with product, engineering, analytics, and compliance on secure, high-quality data deliveryRequirementsSolid engineering fundamentals and experience building pipelines from scratchPython and SQL fluency; comfortable across relational + NoSQL systemsExperience with orchestrators like Temporal, Airflow, or DagsterHands-on with BigQuery, BigTable, and core GCP data toolingAbility to turn messy, ambiguous data problems into clear, scalable solutionsStartup or small-team experience; comfortable moving fast with ownershipCommunication skills, attention to detail, and a bias toward clarity and reliabilityYou don’t need to tick every box to fit in here. If the problems we’re solving genuinely interest you and you know you can contribute, we’d love to talk.Nice to haveExperience with HIPAA/PHI or regulated healthcare dataBackground with time-series data or event-driven architecturesFamiliarity with dbt or similar transformation frameworksExperience with healthcare, diagnostics, or ML/AI workloadsHow you'll be compensatedSalary: $160K - $200k + early stage optionsYour salary is dependant on your location and experience level, generated by our salary calculator. Read more in our handbook here.Generous early stage options (extended exercise post 2 years employment) - you will receive 3 offers based on how much equity you'd likeRegular in person offsites, last were in Morocco and TenerifeBi-weekly team happy hours & events remotelyMonthly learning budget of $300 for personal development/productivityFlexible, remote-first working - including $1K for home office equipment25 days off a year + national holidaysHealthcare cover depending on locationOh and before we forget: Backend Stack: Python (FastAPI), Go, PostgreSQL, Google Cloud Platform (Cloud Run, GKE, Cloud BigTable, etc), Temporal CloudFrontend Stack: TypeScript, Next.jsAPI docs are here: https://docs.junction.com/Company handbook is here with engineering values + principlesImportant details before applying:We only hire folks physically based in GMT and EST timezones - more information here.We do not sponsor visas right now given our stage