Senior Data Engineer
Rentable
Date: 16 hours ago
City: Madison, WI
Contract type: Full time
Remote

Your New Team
The ApartmentIQ team is revolutionizing how property managers optimize their operations. We provide a powerful platform that centralizes critical property data, offering actionable insights and intuitive tools to enhance efficiency and decision-making. Our product empowers users to streamline everything from leasing and maintenance to resident communication and financial reporting.
Behind the scenes, our data platform runs on a modern AWS-based tech stack designed to support big data architectures and machine learning models at scale. We believe in fostering an inclusive environment built on mutual trust, continuous learning, and a commitment to simplicity, where every engineer can contribute and grow.
The Role
As a Senior Data Engineer, you'll be the technical backbone of the data layer that powers Daylight — ApartmentIQ's revenue-management product that delivers real-time rent recommendations to property managers. You'll design, build, and own the ingestion framework that pulls operational data from a variety of property-management systems, transforms it into analytics-ready models, and serves it to the machine-learning workflows that forecast demand and optimize pricing.
Working hand-in-hand with data scientists, you'll ensure every byte flowing through Daylight is trustworthy, traceable, and available at the cadence our algorithms require. You'll architect cloud-native, Terraform-managed infrastructure; implement scalable batch and streaming ETL/ELT jobs in Python; and layer in observability, testing, and data-quality guards so teams can iterate on models with confidence. You'll also build and own core MLOps components that power model training, inference, and deployment — ensuring our ML systems are reliable, repeatable, and production-ready.
Beyond coding, you'll collaborate with product managers, backend engineers, and customer-facing teams to translate business requirements—like a new rent rule or occupancy forecast—into performant data solutions. If you thrive on end-to-end ownership, relish tough debugging sessions, and want to see your work directly influence rent recommendations across thousands of units, we'd love to meet you.
Responsibilities
The ApartmentIQ team is revolutionizing how property managers optimize their operations. We provide a powerful platform that centralizes critical property data, offering actionable insights and intuitive tools to enhance efficiency and decision-making. Our product empowers users to streamline everything from leasing and maintenance to resident communication and financial reporting.
Behind the scenes, our data platform runs on a modern AWS-based tech stack designed to support big data architectures and machine learning models at scale. We believe in fostering an inclusive environment built on mutual trust, continuous learning, and a commitment to simplicity, where every engineer can contribute and grow.
The Role
As a Senior Data Engineer, you'll be the technical backbone of the data layer that powers Daylight — ApartmentIQ's revenue-management product that delivers real-time rent recommendations to property managers. You'll design, build, and own the ingestion framework that pulls operational data from a variety of property-management systems, transforms it into analytics-ready models, and serves it to the machine-learning workflows that forecast demand and optimize pricing.
Working hand-in-hand with data scientists, you'll ensure every byte flowing through Daylight is trustworthy, traceable, and available at the cadence our algorithms require. You'll architect cloud-native, Terraform-managed infrastructure; implement scalable batch and streaming ETL/ELT jobs in Python; and layer in observability, testing, and data-quality guards so teams can iterate on models with confidence. You'll also build and own core MLOps components that power model training, inference, and deployment — ensuring our ML systems are reliable, repeatable, and production-ready.
Beyond coding, you'll collaborate with product managers, backend engineers, and customer-facing teams to translate business requirements—like a new rent rule or occupancy forecast—into performant data solutions. If you thrive on end-to-end ownership, relish tough debugging sessions, and want to see your work directly influence rent recommendations across thousands of units, we'd love to meet you.
Responsibilities
- Design, build, and maintain scalable MLOps infrastructure to support model training, deployment, monitoring, and continuous integration/continuous delivery (CI/CD) of ML models.
- Develop and manage robust data pipelines to extract, transform, and load (ETL/ELT) data from a variety of structured and unstructured sources.
- Collaborate with data scientists and ML engineers to understand model requirements and ensure production readiness of data and model workflows.
- Debug complex data issues and ML pipeline failures, collaborating closely with data scientists and ML engineers to diagnose root causes in data or algorithm behavior.
- Debug data / algorithmic related problems in production for user-facing applications
- Design and optimize data storage solutions using modern data warehousing, and relational database systems.
- Codify and manage cloud infrastructure using Infrastructure as Code tools, primarily Terraform, to ensure reproducibility, scalability, and auditability across environments.
- Implement observability, alerting, and data quality frameworks to ensure pipeline health and uphold data integrity.
- 5+ years of software engineering experience, including 3+ years working directly with data-intensive systems, pipelines, and infrastructure.
- Display a strong sense of ownership and delivering end-to-end systems — from architecture and implementation to CI/CD, observability, and infrastructure management.
- Runs toward problems: has zero tolerance for bugs/issues, leans into complex issues, and proactively investigates and resolves failures.
- Strong debugging capabilities; seeks root causes, not band-aids — whether it's a data anomaly, algorithmic quirk, or system failure.
- Strong Python experience — can write clear, idiomatic code and understands best practices.
- Comfortable writing SQL queries to analyze relational data.
- Experience with Terraform or other Infrastructure-as-Code tools for provisioning cloud-based infrastructure (e.g., AWS, GCP).
- Hands-on experience designing and implementing big data architectures, streaming or batch ETL pipelines, and understanding the trade-offs between complexity, performance, and cost.
- Experience with data lakes, data warehouses, relational databases, and document stores, and when to use each.
- Math or CS background preferred and/or experience working with algorithms.
- Uses LLMs and AI agents to enhance engineering productivity and explore solution spaces creatively.
- Operates effectively in fast-paced, startup environments; adapts quickly and communicates clearly.
- Strong collaborator and communicator, deeply integrated with the team, and proactively shares context and decisions.
- Ruby and/or Ruby on Rails framework.
- Writing performant code using methods like parallelism and concurrency.
- AWS services: SageMaker, Lambda, Redshift, OpenSearch, Kinesis.
- Experience with distributed systems.
- 100% remote across the U.S., with quarterly in-person gatherings for team offsites
- Competitive Compensation
- Flexible vacation and parental leave policies
- Medical, Dental, and Vision Insurance
- 100% paid Short-Term Disability, Long-Term Disability, and Life Insurance Program
- 401k Program
- A supportive, learning-first culture where you'll help shape the next generation of AI-driven marketing tools for the apartment rental industry
How to apply
To apply for this job you need to authorize on our website. If you don't have an account yet, please register.
Post a resume