Remove flink-sql-introduction
article thumbnail

Data News — Week 24.08

Christophe Blefari

JVM vs. SQL data engineer — There's a big discussion in the community about what real data engineering is. Is it DataFrames or SQL? Still, I prefer SQL/Python data engineering, as you know me. I did not read the paper except the introduction and a the first schema, but it looks like awesome. PyIceberg 0.6.0:

Data Lake 130
article thumbnail

Data News — Week 23.16

Christophe Blefari

As introduction Tristan gives the original vision of dbt that became mainstream, today. A lot of data teams embraced dbt, or at least the SQL with engineering practices to transform data in cloud data warehouses. This is a preambule to cross-project dependencies I guess. Building a ChatGPT Plugin for Medium.

Raw Data 130
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Reducing The Barrier To Entry For Building Stream Processing Applications With Decodable

Data Engineering Podcast

RudderStack Profiles takes the SaaS guesswork and SQL grunt work out of building complete customer profiles so you can quickly ship actionable, enriched data to every downstream team. It’s the only true SQL streaming database built from the ground up to meet the needs of modern data products. With Materialize, you can!

Process 182
article thumbnail

Our First Netflix Data Engineering Summit

Netflix Tech

Streaming SQL on Data Mesh using Apache Flink Mark Cho, Guil Pires and Sujay Jain, Engineers from the Netflix Data Platform talk about how a managed Streaming SQL using Apache Flink can help unlock new Stream Processing use cases at Netflix.

article thumbnail

Your Guide to Flink SQL: An In-Depth Exploration

Confluent

Get an in-depth introduction to Flink SQL. Learn how it relates to other APIs, its built-in functions and operations, which queries to try first, and see syntax examples.

SQL 70
article thumbnail

Druid Deprecation and ClickHouse Adoption at Lyft

Lyft Engineering

Introduction At Lyft, we have used systems like Apache ClickHouse and Apache Druid for near real-time and sub-second analytics. Real-time Ingestion Events from our real-time analytics pipeline were configured to be sent into our internal Flink application, streamed to Kafka, and written into Druid. This was our main form of ingestion.

Kafka 104
article thumbnail

Enriching Streams with Hive tables via Flink SQL

Cloudera

Introduction. Flink SQL does this and directs the results of whatever functions you apply to the data into a sink. Therefore, there are two common use cases for Hive tables with Flink SQL: A lookup table for enriching the data stream. A sink for writing Flink results. Using Flink DDL with JDBC connector.

SQL 57