Job Title: Software Engineer, Data Company Name: LiveKit Job Url: https://jobright.ai/jobs/info/69bb928606c1ba00c54e2ef6 Job Description: LiveKit ยท 11 hours ago Software Engineer, Data United States Full-time Remote Mid, Senior Level $120K/yr - $250K/yr 59% FAIR MATCH 72% Exp. Level 32% Skill 82% Industry Exp. LiveKit is building the infrastructure layer for the voice-driven era of computing. They are seeking a Software Engineer, Data to help design and operate the metering and analytics infrastructure that powers their platform, processing massive volumes of usage data generated by developers and end-users. Artificial Intelligence (AI) Cloud Computing Information Technology Big Data SaaS Software Cloud Infrastructure Real Time H1B Sponsor Likely Insider Connection @LiveKit 2 email credits available today Discover valuable connections within the company who might provide insights and potential referrals. Get 3x more responses when you reach out via email instead of LinkedIn. Beyond Your Network Find More Connections From Your Previous Company Find More Connections From Your School Find More Connections Find Any Email Responsibilities Design and evolve metering and analytics infrastructure spanning real-time analytics, long-term analysis, data transfer, governance, and retention policies Collaborate with teams across the organization to ensure metering and analytics are correct and complete for their domains - Agents, Agent Insights, Cloud Dashboard, and customer-facing reporting Monitor and manage datasets with varying cardinality - both internally-defined datasets we control and customer-produced datasets where cardinality is unbounded and efficient querying is essential Ensure data reliability through delivery guarantees, dead letter queues, reconciliation, validation, alerting, and anomaly detection across our distributed service fleet Design and enforce schema evolution strategies (e.g., schema registries, backward/forward compatibility contracts) to evolve infrastructure without breaking downstream consumers Optimize ClickHouse and blob storage for query performance, cost efficiency, and reliability across global regions Reduce operational toil through automation, self-service tooling, and runbooks Qualification Represents the skills you have Find out how your skills align with this job's requirements. If anything seems off, you can easily click on the tags to select or unselect skills to reflect your actual expertise. Go Distributed systems architecture Data pipelines Columnar/analytical databases ClickHouse BigQuery Blob storage Data correctness Schema evolution Delivery semantics Idempotency Stream processing frameworks Kafka Pulsar Kubernetes OpenTelemetry Protobuf schema registries Avro schema registries Query federation engines Trino Presto Dremio Usage-based billing Metering systems Cross-team collaboration Required strong experience designing and operating data pipelines and distributed systems in production across dozens of global regions extensively worked with Go and contributes comfortably to a distributed systems architecture deep experience with columnar/analytical databases (ClickHouse, BigQuery, or similar) and blob storage for high-volume workloads think deeply about data correctness - delivery semantics, idempotency, schema compatibility, and the failure modes that cause silent data loss strong cross-team collaborator who translates domain requirements into practical infrastructure designs previous experience working on data-intensive SaaS applications with web-based dashboards in the analytics (reporting, observability or finance) space Preferred Experience with stream processing frameworks (Kafka, Pulsar) Kubernetes OpenTelemetry query federation engines (Trino, Presto, Dremio) protobuf/Avro schema registries usage-based billing/metering systems Benefits Health, dental, and vision benefits Flexible vacation policy