the datalake for the home — unify smart sensors, cameras, energy meters, and everything else into a single queryable lake. no cloud. no subscriptions. just data.
everything you need to collect, store, query, and visualize your home data — without sending it to someone else's cloud.
real-time ingestion from MQTT, Zigbee, Z-Wave, HomeAssistant, and custom APIs. zero-config auto-discovery for 200+ device types.
data lands in Apache Parquet on local disk — compressed, partitioned, and ready for analytical queries. no external DB needed.
full SQL engine (DataFusion-powered) runs directly against your lake. joins across device types, time-range filters, aggregates — it just works.
zero outbound telemetry. your data never leaves your network. runs on a Pi, an N100 mini-PC, or that old laptop collecting dust.
point your browser at localhost:9900 and get live dashboards. no Grafana config hell. drag, drop, done.
write custom sources, transforms, and sinks in Python or Rust. hot-reload them without restarting your lake.
no infrastructure to provision. no schemas to define. pond figures out your data and makes it queryable.
run pond init to create a datalake on any directory. pond auto-detects device sources on your local network and sets up streaming connections.
pond ingest --all starts collecting from every discovered source. data streams into columnar Parquet files, partitioned by time and device type. compaction runs automatically.
pond query opens an interactive SQL shell. or hit the dashboard at :9900. cross-device joins, time windows, anomaly detection — all local, all fast.
pond isn't just a tool — it's a platform. embed it, script it, extend it.
one command to install. zero config to start. your home data has been waiting for this.