Remove transactions-apache-kafka
article thumbnail

Transactional Machine Learning at Scale with MAADS-VIPER and Apache Kafka

Confluent

This blog post shows how transactional machine learning (TML) integrates data streams with automated machine learning (AutoML), using Apache Kafka® as the data backbone, to create a frictionless machine learning […].

article thumbnail

Fraud Detection With Cloudera Stream Processing Part 2: Real-Time Streaming Analytics

Cloudera

In part 1 of this blog we discussed how Cloudera DataFlow for the Public Cloud (CDF-PC), the universal data distribution service powered by Apache NiFi, can make it easy to acquire data from wherever it originates and move it efficiently to make it available to other applications in a streaming fashion.

Process 85
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Fraud Detection with Cloudera Stream Processing Part 1

Cloudera

In a previous blog of this series, Turning Streams Into Data Products , we talked about the increased need for reducing the latency between data generation/ingestion and producing analytical results and insights from this data. This blog will be published in two parts. This is what we call the first-mile problem.

Process 80
article thumbnail

Big Data Technologies that Everyone Should Know in 2024

Knowledge Hut

In this blog post, we will discuss such technologies. There are a variety of big data processing technologies available, including Apache Hadoop, Apache Spark, and MongoDB. In general, Hadoop and Spark are good choices for batch processing, while Kafka and Storm are better suited for streaming applications.

article thumbnail

A Closer Look at The Next Phase of Cloudera’s Hybrid Data Lakehouse

Cloudera

Cloudera is now the only provider to offer an open data lakehouse with Apache Iceberg for cloud and on-premises. Apache Ozone As AI and other advanced analytics continue to grow in scale, performance and scalable data storage will need to expand right along with them. ZDU gives organizations a more convenient means of upgrading.

article thumbnail

Deploying Data Pipelines using the Saga pattern

Picnic Engineering

In our previous blog, Dima Kalashnikov explained how we configure our Internal services pipeline in the Analytics Platform. The steps of this story are actually local transactions distributed across the various components, all coordinated to achieve our larger goal. A true heroic story indeed! How does it work?

article thumbnail

What is Apache Kafka Used For?

ProjectPro

Did you know thousands of businesses, including over 80% of the Fortune 100, use Apache Kafka to modernize their data strategies? Apache Kafka is the most widely used open-source stream-processing solution for gathering, processing, storing, and analyzing large amounts of data. What is Apache Kafka Used For?

Kafka 52