article thumbnail

Comparing ClickHouse vs Rockset for Event and CDC Streams

Rockset

Streaming data feeds many real-time analytics applications, from logistics tracking to real-time personalization. Event streams, such as clickstreams, IoT data and other time series data, are common sources of data into these apps. Flink, Kafka and MySQL. The software was subsequently open sourced in 2016.

MySQL 52
article thumbnail

Druid Deprecation and ClickHouse Adoption at Lyft

Lyft Engineering

Our initial use for Druid was for near real-time geospatial querying and high performance on high-cardinality data sets. It also allowed us to optimize for handling time-series data and event data at scale. Pre-aggregating data at ingestion time helped optimize our query performance and reduce our storage costs.

Kafka 104
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Internal services pipeline in Analytics Platform

Picnic Engineering

The data is loaded into Snowflake, Picnic’s single source of truth Data Warehouse (DWH). Almost all internal services emit events over RabbitMQ. Our pipeline captures these events and sends them to Confluent Cloud. We use the RabbitMQ Source connector for Apache Kafka Connect.

Kafka 52
article thumbnail

Building Real-time Machine Learning Foundations at Lyft

Lyft Engineering

The Event Driven Decisions capability in particular turned out to be general enough as to be applicable to a wide range of use cases. At the time of writing, a Mapping team is working to utilize theEvent Driven Decisions product to rebuild Lyft’s Traffic infrastructure by aggregating data per geohash and applying a model.

article thumbnail

How Rockset Enables SQL-Based Rollups for Streaming Data

Rockset

Apache Kafka has made acquiring real-time data more mainstream, but only a small sliver are turning batch analytics, run nightly, into real-time analytical dashboards with alerts and automatic anomaly detection. But until this release, all these data sources involved indexing the incoming raw data on a record by record basis.

SQL 52
article thumbnail

Machine Learning with Python, Jupyter, KSQL and TensorFlow

Confluent

The blog posts How to Build and Deploy Scalable Machine Learning in Production with Apache Kafka and Using Apache Kafka to Drive Cutting-Edge Machine Learning describe the benefits of leveraging the Apache Kafka ® ecosystem as a central, scalable and mission-critical nervous system. For now, we’ll focus on Kafka.

article thumbnail

Addressing the Challenges of Sample Ratio Mismatch in A/B Testing

DoorDash Engineering

Experiment exposures are one of our highest volume events. On a typical day, our platform produces between 80 billion and 110 billion exposure events. We stream these events to Kafka and then store them in Snowflake. Users can query this data to troubleshoot their experiments. For this we used Apache Pinot.