Dashdive has three key components: (1) event collection, (2) event ingestion, and (3) analytical queries on ingested events.
1. Event Collection
Each individual event must be tagged by relevant attributes, such as associated feature (
feature_id) and originating customer (
customer_id). We also support custom attributes; please contact us to enable this.
Extracting attributes. When using the agent, you can write small attribute extractor functions to pick out attributes from available context. For example, you might deduce the
customer_id from the API key on an inbound HTTP request. When sending events directly, you must extract attributes manually in your code.
Sending events. There are two options for collecting events:
- Collect your cloud usage events manually and send them to our ingestion endpoint directly, either via custom code (see the API reference) or via our SDKs.
- Install our agent to collect and tag events automatically in the background. To access builds of the agent, please reach out.
2. Event Ingestion
Our top priority is data completeness and correctness. This necessitates no dropped events (high availability) and no duplicates (exactly-once semantics).
- High availability. All components in our ingestion pipeline, including the API server nodes, Kafka brokers, and ClickHouse databases are replicated for high availability and redundancy.
- Exactly-once semantics. Exactly-once semantics are hard. Thankfully, we can use the exactly-one semantics already implemented in Kafka and the Kafka ClickHouse connector to ensure the same is true of our entire pipeline.
Additionally, while the schema for ingested events is strictly enforced to ensure well-structured data, malformed events aren’t simply discarded. The schema is permissive, meaning that extraneous fields are ignored rather than rejected. Also, any truly malformed events — with required fields missing or having incorrect data types — are recorded in raw form in the database. This way, although the events were rejected on first receipt, they can be manually reviewed and corrected later.
3. Analytical Queries on Ingested Events
You can use the graphical interface in the Dashdive web app to answer queries about the ingested events. Unlike other cloud products, since we collect individual events, you can break down by ultra-high-cardinality fields such as
The dashboard is intended to be real-time and interactive, in contrast to some data warehouses that take minutes or more to execute a single query. In most cases, Dashdive achieves sub-second query times by pre-aggregating raw events into optimized tables and choosing the right table intelligently per-query. The result is a query latency of <1 second even on >100 billion raw events, in many cases. You can see this in action with medium-scale data in our demo.