Data Engineer



To our future Data Engineer


At Density, we build one of the most advanced people sensing systems in the world. The product and infrastructure is nuanced and one-of-a-kind. Building this product for scale has been an exercise in patience, creativity, remarkable engineering, laser physics, global logistics, and grit. The team is thoughtful, driven, and world-class.


Why this is an important role


Last week our deployed DPUs detected a million humans walking through doors. A number that increases every week.


As engineers, we think it's pretty cool to be capturing events at this volume. Especially when it's done anonymously, accurately, and in real-time. Our customers, however, are interested in what happens after these events enter our system.


Density is deployed globally, servicing a variety of use cases, by some of the largest companies in the world. Here are just a few examples. A leading cloud storage company is using Density to strengthen their physical security by detecting unauthorized access. A marquee hotel brand is using Density to measure lounge occupancy and dynamically deliver world-class service. An international telecom is using Density to better design and optimize their real estate portfolio.


These use cases may seem disparate, but a common thread holds them together. They all share the need for real-time and ongoing data analysis. And that's where you come in.


Our systems must efficiently and reliably:


- Update current counts and analytics whenever a relevant event occurs, for every impacted space,
- Publish the information to low-latency receivers via webhooks and websockets,
- Alert and notify as appropriate through SMS, email, and push notifications, and
- Aggregate events into analytics used for dashboards, forecasts, and fleet management.


Turning millions of events into actionable insight is a nuanced dance. Are you up for the job?


What you'll work on



  • Scale event ingestion pipelines requiring high availability and real-time data processing/delivery.

  • Store, optimize, and deliver analytical data via a RESTful API.

  • Backend development on internal and customer facing projects, utilizing Python, Django, and NodeJS.

  • Refine APIs and data delivery mechanisms for applications such as web dashboards, alerting & health systems, mobile applications, and third party integrations

  • Work closely with DevOps to monitor inefficiencies and improve infrastructure

  • React to customer needs and feedback through tight-looped, iterative development

  • Contribute to open source initiatives

  • Document and teach best practices across our stack


What we're looking for



  • 5+ years industry experience building and scaling web applications and APIs

  • Deep experience with stream processing systems (i.e. Kafka)

  • Experience writing ETL pipelines on a cloud infrastructure (AWS)

  • Deep experience with Python, Django / DRF, and Postgres

  • Experience building data-centric applications, including analytics pipelines, report generation systems, and alerting & health systems

  • An understanding and appreciation for application performance monitoring and profiling tools

  • A desire to define, document, and teach web engineering standards

  • Strong writing skills, especially with crafting clear and concise documentation

  • A motivation for constant learning


Icing on the cake



  • Experience with statistical analysis and trend data modeling

  • A deep appreciation for design

  • A strange obsession with counting people (or what you can do with the resultant data)


While we have offices in Syracuse (NY), San Francisco, and NYC, we embrace and have built a culture around remote work.


Apply Now

Back to jobs