Back to Jobs

Data Engineer - AI (REMOTE)

Remote, USA Full-time Posted 2025-11-24
Description • Own the end-to-end data lifecycle for Upbound’s AI initiatives, from raw ingestion through model-ready datasets, powering the next generation of Crossplane and the Intelligent Control Plane. • Architect and maintain scalable, cloud-native data pipelines (batch + streaming) that collect, clean, and enrich telemetry from thousands of Kubernetes clusters, cloud APIs, and customer workloads worldwide. • Partner with ML engineers, product, and SRE teams to define data contracts, schema evolution strategies, and governance policies that keep petabyte-scale lakes reliable, secure, and compliant (SOC 2, GDPR, HIPAA). • Design real-time feature stores that feed both online inference services and offline training jobs, ensuring sub-second latency for critical control-plane decisions while guaranteeing reproducibility and version control. • Build self-service tooling (SDKs, notebooks, observability dashboards) that empowers analysts and data scientists to discover, profile, and experiment with datasets without bottlenecks. • Optimize compute and storage costs through intelligent partitioning, incremental processing, and auto-scaling clusters on AWS/GCP, cutting spend by double-digit percentages year-over-year. • Implement advanced data quality frameworks—unit tests, anomaly detection, lineage tracking—that surface issues before they reach production models or customer dashboards. • Contribute to open-source Crossplane providers and Upbound’s internal “Data as Infrastructure” codebase, turning repeatable patterns into reusable packages the community can adopt. • Champion a culture of documentation and knowledge sharing: run internal tech talks, write runbooks, and mentor junior engineers to raise the bar for data excellence across the company. • Stay ahead of the curve by evaluating emerging technologies (Iceberg, DuckDB, Flink, vector databases) and running proof-of-concepts that translate into competitive advantages for Upbound’s AI roadmap. Requirements • 5+ years building production-grade data pipelines in Python, SQL, and at least one JVM language (Scala/Java/Kotlin). • Deep expertise with cloud data stacks: S3/GCS, Redshift/BigQuery, EMR/Dataproc, Kinesis/PubSub, Airflow/Mage, dbt, Terraform. • Hands-on experience with Kubernetes, Docker, and infrastructure-as-code; familiarity with Crossplane is a strong plus. • Proven track record designing real-time streaming architectures (Kafka, Pulsar, Flink) and batch ETL at multi-terabyte scale. • Nice-to-have: contributions to open-source data projects, advanced SQL performance tuning, or prior work in ML feature engineering. ️ Benefits • Fully remote-first culture with quarterly off-sites in inspiring global locations. • Competitive salary + equity package that grows with the company’s valuation. • $3,000 annual learning stipend for conferences, courses, and certifications. • Flexible PTO policy and 16-week gender-neutral parental leave. • Home-office setup budget and monthly wellness stipend. Apply tot his job Apply To this Job

Similar Jobs