article thumbnail

Mainframe Optimization: 5 Best Practices to Implement Now

Precisely

It frequently also means moving operational data from native mainframe databases to modern relational databases. Typically, a mainframe to cloud migration includes re-factoring code to a modern object-oriented language such as Java or C# and moving to a modern relational database. Best Practice 2. Best Practice 3.

article thumbnail

Reflections On Designing A Data Platform From Scratch

Data Engineering Podcast

If you’re a data engineering podcast listener, you get credits worth $3000 on an annual subscription TimescaleDB, from your friends at Timescale, is the leading open-source relational database with support for time-series data. Time-series data is relentless and requires a database like TimescaleDB with speed and petabyte-scale.

Designing 100
Insiders

Sign Up for our Newsletter

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.

article thumbnail

Iceberg Tables: Catalog Support Now Available

Snowflake

Iceberg supports many catalog implementations: Hive, AWS Glue, Hadoop, Nessie, Dell ECS, any relational database via JDBC, REST, and now Snowflake. After making an initial connection to Snowflake via the Iceberg Catalog SDK, Spark can read Iceberg metadata and Parquet files directly from the customer-managed storage account.

article thumbnail

AWS Glue-Unleashing the Power of Serverless ETL Effortlessly

ProjectPro

Users can schedule ETL jobs, and they can also choose the events that will trigger them. Then, Glue writes the job's metadata into the embedded AWS Glue Data Catalog. AWS Glue then creates data profiles in the catalog, a repository for all data assets' metadata, including table definitions, locations, and other features.

AWS 98
article thumbnail

Toward a Data Mesh (part 2) : Architecture & Technologies

François Nguyen

To illustrate that, let’s take Cloud SQL from the Google Cloud Platform that is a “Fully managed relational database service for MySQL, PostgreSQL, and SQL Server” It looks like this when you want to create an instance. ” He/She is managing triggers, he/she needs to check conditions (event type ?

article thumbnail

An Engineering Guide to Data Creation - A Data Contract perspective - Part 1

Data Engineering Weekly

Data engineering starts to add value to the business by capturing events at each step of the business process. The events are then further enriched and analyzed to bring visibility to business operations. Event Sourcing Change Data Capture [CDC] Outbox pattern 1. However, Event sourcing comes with a few major limitations.

article thumbnail

Turning Streams Into Data Products

Cloudera

For governance and security teams, the questions revolve around chain of custody, audit, metadata, access control, and lineage. Apache Flink is a distributed processing engine for stateful computations ideally suited for real-time, event-driven applications. Meet Laila, a very opinionated practitioner of Cloudera Stream Processing.

Kafka 86