article thumbnail

How Snowflake Enhanced GTM Efficiency with Data Sharing and Outreach Customer Engagement Data

Snowflake

To improve go-to-market (GTM) efficiency, Snowflake created a bi-directional data share with Outreach that provides consistent access to the current version of all our customer engagement data. In this blog, we’ll take a look at how Snowflake is using data sharing to benefit our SDR teams and marketing data analysts.

BI 72
article thumbnail

Building Real-time Machine Learning Foundations at Lyft

Lyft Engineering

Our goal was to develop foundations that would enable the hundreds of ML developers at Lyft to efficiently develop new models and enhance existing models with streaming data. In this blog post, we will discuss what we built in support of that goal and some of the lessons we learned along the way.

Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

B2B Data Enrichment for Beginners

Precisely

That’s where data enrichment comes into the picture. In this blog post, we’ll explain what data enrichment is, why you need it, how it works, and how B2B companies can use enriched data to drive results. What is data enrichment? How does data enrichment work? That depends on your objectives.

article thumbnail

Druid Deprecation and ClickHouse Adoption at Lyft

Lyft Engineering

In this particular blog post, we explain how Druid has been used at Lyft and what led us to adopt ClickHouse for our sub-second analytic system. Druid at Lyft Apache Druid is an in-memory, columnar, distributed, open-source data store designed for sub-second queries on real-time and historical data.

Kafka 104
article thumbnail

Deployment of Exabyte-Backed Big Data Components

LinkedIn Engineering

Our RU framework ensures that our big data infrastructure, which consists of over 55,000 hosts and 20 clusters holding exabytes of data, is deployed and updated smoothly by minimizing downtime and avoiding performance degradation. We needed a deep understanding of system dependencies to ensure a smooth deployment process.

article thumbnail

Tips to Build a Robust Data Lake Infrastructure

DareData

In this blog post, we aim to share practical insights and techniques based on our real-world experience in developing data lake infrastructures for our clients - let's start! The Data Lake acts as the central repository for aggregating data from diverse sources in its raw format.

article thumbnail

Addressing the Challenges of Sample Ratio Mismatch in A/B Testing

DoorDash Engineering

Experiment exposures are one of our highest volume events. On a typical day, our platform produces between 80 billion and 110 billion exposure events. We stream these events to Kafka and then store them in Snowflake. Users can query this data to troubleshoot their experiments.