January 31, 2024

Apache Hudi vs Delta Lake vs Apache Iceberg - Data Lakehouse Feature Comparison

Apache Hudi vs Delta Lake vs Apache Iceberg - Data Lakehouse Feature Comparison

Introduction

With the growing popularity of the data lakehouse there has been a rising interest in the analysis and comparison of the three open source projects which are at the core of this data architecture: Apache Hudi, Delta Lake, and Apache Iceberg.

Most comparison articles currently published seem to evaluate these projects merely as table/file formats for traditional append-only workloads, overlooking some qualities and features that are critical for modern data lake platforms that need to support update-heavy workloads with continuous table management. This article will dive into greater depth as to the feature comparisons and also comprehensively cover benchmarks and community statistics.

If, when analyzing the comparisons, you find it hard to choose which format you want to use, take a look at a brand new project: Onetable, which offers seamless interoperability between Hudi, Delta, and Iceberg. You no longer have to choose between formats or be locked into just one format. OneTable was co-launched into open source by Microsoft, Google, and Onehouse with VentureBeat covering the story. OneTable is now being donated to the Apache Software Foundation and at current time of writing is in the process of finalizing a name change.

Feature Comparisons

First, let's look at an overall feature comparison. As you read, notice how the Hudi community has invested heavily into comprehensive platform services on top of the lake storage format. While formats are critical for standardization and interoperability, table/platform services give you a powerful toolkit to easily develop and manage your data lake deployments.

Community Momentum

Equally as important as the features and capabilities of an open source project is the community. The community can make or break the development momentum, ecosystem adoption, or the usefulness of the platform. Below is a comparison of Hudi, Delta, and Iceberg when it comes to their communities. 

Github Stars:

Github stars is a vanity metric that represents popularity more than contribution. Delta Lake leads the pack in awareness and popularity.

Github Watchers and Forks

A closer indication of engagement/usage of the project is the number of watchers a project has on Github and the number of times it has been forked.

Github Contributors

In December 2022, Apache Hudi had almost 90 unique authors contribute to the project - more than double the number for Iceberg and triple the number for Delta Lake.

Github PRs and Issues

In December 2022 Hudi and Iceberg merged about the same # of pull requests (PRs), while the number of PRs opened was double in Hudi.

Contribution Diversity

Apache Hudi and Apache Iceberg have strong diversity as to the breadth of the community that contributes to the project.

Apache Hudi

Contributors to Apache Hudi

Apache Iceberg

Contributors to Apache Iceberg

Delta Lake

Contributors to Delta Lake

TPC-DS Performance Benchmarks

Standard performance benchmarks rarely are representative of real-life workloads, and we strongly encourage the community to run their own analysis against their own data. Nonetheless, these benchmarks can serve as an interesting data point while you start your own research into choosing a lakehouse platform. Below are references to relevant benchmarks.

Databeans and Onehouse

Databeans worked with Databricks to publish a benchmark used in their Data+AI Summit Keynote in June 2022, but they misconfigured an obvious out-of-box setting. Onehouse corrected the benchmark here:
https://www.onehouse.ai/blog/apache-hudi-vs-delta-lake-transparent-tpc-ds-lakehouse-performance-benchmarks
 

Brooklyn Data and Onehouse

Databricks asked Brooklyn Data to publish a benchmark of Delta vs Iceberg in November 2022:
https://brooklyndata.co/blog/benchmarking-open-table-formats 

Onehouse added Apache Hudi and published the code in the Brooklyn Github repo:
https://github.com/brooklyn-data/delta/pull/2
 

A clear pattern emerges from these benchmarks: Delta and Hudi are comparable, while Apache Iceberg consistently trails behind as the slowest of the projects. Performance isn’t the only factor you should consider, but performance does translate into cost savings that add up throughout your pipelines.

A note on running TPC-DS benchmarks:

One key thing to remember when running TPC-DS benchmarks comparing Hudi, Delta, and Iceberg is that, by default, Delta and Iceberg are optimized for append-only workloads. Hudi default settings, by contrast, are optimized for mutable workloads. By default, Hudi uses an `upsert` write mode which naturally has a write overhead compared to inserts. Without this knowledge you may be comparing apples to oranges. Change this one out-of-the-box configuration to `bulk-insert` for a fair assessment:
https://hudi.apache.org/docs/write_operations/  

Feature Highlights

Building a data lakehouse platform takes more than just reviewing checkboxes listing feature availability. Let’s pick a few of the differentiating features above and dive into the use cases and real benefits in plain English.

Incremental Pipelines

The majority of data engineers today feel like they have to choose between streaming data and old-school batch ETL pipelines. Apache Hudi has pioneered a new paradigm called incremental pipelines. Out of the box, Hudi tracks all changes (appends, updates, deletes) and exposes them as change streams. With record-level indexes you can more efficiently leverage these change streams to avoid recomputing data and just process changes using incremental updates. While other data lake platforms may enable a way to consume changes incrementally, Hudi is designed from the ground up to enable incrementalization efficiently, which results in cos- efficient ETL pipelines at lower latencies.

Databricks recently developed a similar feature they call change data feed, which they held as proprietary until it was finally released to open source in Delta Lake 2.0. Iceberg has an incremental read, but it only allows you to read incremental appends, not updates and deletes - which are essential for true change data capture (CDC) and transactional data within a data lakehouse.

Concurrency Control

ACID transactions and concurrency control are key characteristics of a data lakehouse, but how do current designs actually stack up compared to real-world workloads? Hudi, Delta, and Iceberg all support optimistic concurrency control (OCC). In optimistic concurrency control, writers check if they have overlapping files; if a conflict exists, they fail the operations and retry. For Delta Lake, as an example, this was just a Java virtual machine (JVM)-level lock held on a single Apache Spark driver node - which meant you had no OCC outside of a single cluster, until recently.

While this may work fine for append-only immutable datasets, optimistic concurrency control struggles with real-world scenarios, which introduce the need for frequent updates and deletes - either because of the data loading pattern or due to the need to reorganize the data for better query performance. 

Oftentimes, it’s not practical to take writers offline for table management to ensure the table is healthy and performant. Apache Hudi concurrency control is more granular than other data lakehouse platforms (file-level), and with a design optimized for multiple small updates/deletes, the conflict possibility can be largely reduced to negligible in most real-world use cases. 

You can read more details in this blog as to how you can operate with asynchronous table services even in multi-writer scenarios, without the need to pause writers. This achieves a level of concurrency very close to the level supported by standard databases.

Merge On Read

Any good database system supports different trade-offs between write performance and query performance. The Hudi community has made some seminal contributions in terms of defining these concepts for data lake storage across the industry. Hudi, Delta, and Iceberg all write and store data in Apache Parquet files. When updates occur, these Parquet files are versioned and rewritten. This write mode pattern is what the industry now calls copy on write (CoW). This model works well for optimizing query performance, but can be limiting for write performance and data freshness. 

In addition to CoW, Apache Hudi supports another table storage layout called merge on read (MoR). Merge on read stores data using a combination of columnar Parquet files and row-based Apache Avro log files. Updates can be batched up in log files that can later be compacted into new Parquet files, synchronously or asynchronously, to balance maximum query performance and lower write amplification.

Thus, for a near real-time streaming workload, Hudi could use the more efficient row-oriented formats, while for batch workloads, Hudi uses the vectorizable, column-oriented format, with seamless merging of the two formats when required. Many users turn to Apache Hudi since it is the only project with this capability, which allows them to achieve unmatched write performance and end-to-end data pipeline latencies.

Partition Evolution

One feature often highlighted for the Apache Iceberg data lakehouse project is hidden partitioning, which unlocks what is called partition evolution. The basic idea is that when your data starts to evolve, or when you just aren’t getting the performance you need out of your current partitioning scheme, partition evolution allows you to update your partitions for new data without the need to rewrite your data. 

When you evolve your partitions, old data is left in the old partitioning scheme and only new data is partitioned with your evolution. However, a table partitioned multiple ways pushes complexity to the user and cannot guarantee consistent performance if the user is unaware of, or simply fails to account for, the evolution history.

Apache Hudi takes a different approach to address the problem of adjusting data layout as your data evolves with clustering. You can choose a coarse-grained partition strategy, or even leave your data unpartitioned, and use a more fine-grained clustering strategy without partitions. With Apache Hudi, clustering can be run synchronously or asynchronously and can be evolved without rewriting any data. This approach is comparable to the micro-partitioning and clustering strategy used by Snowflake.

Multi-Modal Indexing

Indexing is an integral capability for databases and data warehouses, yet is largely absent in data lakes. In recent releases, Apache Hudi created a first-of-its-kind high-performance indexing subsystem for the data lakehouse that we call the Hudi multi-modal index. Apache Hudi offers an asynchronous indexing mechanism that allows you to build and change indexes without impacting write latency. This indexing mechanism is extensible and scalable to support any popular index techniques, such as Bloom, hash, bitmap, R-tree, etc.

These indexes are stored in the Hudi metadata table, which is stored in cloud storage next to your data. In this new release the metadata is written in optimized, indexed file formats which results in 10-100x performance improvements for point lookups versus Delta or Iceberg generic file formats. When testing real-world workloads, this new indexing subsystem results in 10-30x overall query performance.

Ingestion Tools 

What sets a data platform apart from data formats are the operational services available. A differentiator for Apache Hudi is the powerful data ingestion utility called DeltaStreamer. (“Delta” refers to changes to data, not to a particular data lakehouse project.) DeltaStreamer is battle-tested and used in production to build some of the largest data lakes on the planet today. DeltaStreamer is a standalone utility which allows you to incrementally ingest upstream changes from a wide variety of sources such as DFS, Kafka, database changelogs, S3 events, JDBC, and more.

Iceberg has no solution for a managed ingestion utility, and Delta Autoloader remains a Databricks-proprietary feature that only supports cloud storage sources such as S3.

Use Cases - Examples from the community

Feature comparisons and benchmarks can help newcomers orient themselves as to what technology choices are available, but more important is sizing up your own use cases and workloads to find the right fit for your data architecture. All three of these technologies, Hudi, Delta, and Iceberg, have different origin stories and advantages for certain use cases. Hudi, the original data lakehouse project, was born at Uber to power petabyte-scale data lakes in near real-time, with painless table management. Iceberg was born at Netflix and was designed to overcome cloud storage scale problems such as file listings. Delta was born at Databricks and it has deep integrations and accelerations when using the Databricks Spark runtime. 

From years of engaging in real-world comparison evaluations in the community, Apache Hudi routinely has a technical advantage when you have mature workloads that grow beyond simple append-only inserts. Once you start processing many updates, start adding real concurrency, or attempt to reduce the end-to-end latency of your pipelines, Apache Hudi stands out as the industry leader in performance and feature set.

Following are a few examples and stories from members of the community who independently evaluated and decided to use Apache Hudi.

Amazon package delivery system

This story describes how Amazon Transportation Services (ATS) implemented an Apache Hudi-based data lakhouse to handle massive data ingestion challenges and highly mutable workloads at scale. 

“One of the biggest challenges ATS faced was handling data at petabyte scale with the need for constant inserts, updates, and deletes with minimal time delay, which reflects real business scenarios and package movement to downstream data consumers.”

“In this post, we show how we ingest data in real time in the order of hundreds of GBs per hour and run inserts, updates, and deletes on a petabyte-scale data lake using Apache Hudi tables loaded using AWS Glue Spark jobs and other AWS serverless services including AWS Lambda, Amazon Kinesis Data Firehose, and Amazon DynamoDB.”

ByteDance/Tiktok  

This ByteDance/Tiktok scenario involves even larger datasets and shows Hudi being chosen after careful consideration of all three data lakehouse projects. 

“In our scenario, the performance challenges are huge. The maximum data volume of a single table reaches 400PB+, the daily volume increase is PB level, and the total data volume reaches EB level.”

“The throughput is relatively large. The throughput of a single table exceeds 100 GB/s, and the single table needs PB-level storage. The data schema is complex. The data is highly dimensional and sparse. The number of table columns ranges from 1,000 to 10,000+. And there are a lot of complex data types.”

“When making the decision on the engine, we examine three of the most popular data lake engines, Hudi, Iceberg, and DeltaLake. These three have their own advantages and disadvantages in our scenarios. Finally, Hudi is selected as the storage engine based on Hudi's openness to the upstream and downstream ecosystems, support for the global index, and customized development interfaces for certain storage logic.”

Walmart

Walmart, with roughly 11,000 stores worldwide, selling more than $1M on average per store each week, deals with data at massive scale and criticality. 

From video transcription:

“Okay so what is it that enables us for us and why do we really like the Hudi features that have unlocked this in other use cases? We like the optimistic concurrency or MVCC controls that are available to us. We've done a lot of work around asynchronous compaction. We're in the process of looking at doing asynchronous compaction rather than inline compaction on our merge on read tables.

We also want to reduce latency and so we leverage merge on read tables, significantly because that enables us to append data much faster. We also love native support for deletion. It's something we had custom frameworks built for things like CCPA and GDPR, where somebody would put in a service desk ticket and we'd have to build an automation flow to remove records from HDFS; now this comes out of the box for us.

Row versioning is really critical; obviously, a lot of our pipelines have out-of-order data and we need the latest records to show up, so we provide version keys as part of our framework for all upserts into the Hudi tables.

The fact that customers can pick and choose how many versions of a row to keep be able to provide snapshot queries and get incremental updates, like what's been updated in the last five hours, is really powerful for a lot of users.”

Robinhood

Investment site Robinhood makes extensive use of change data capture (CDC) with Kafka streaming to maximize data freshness within a data lakehouse. 

“Robinhood has a genuine need to keep data freshness low for the Data Lake. Many of the batch processing pipelines that used to run on daily cadence after or before market hours had to be run at hourly or higher frequency to support evolving use cases. It was clear we needed a faster ingestion pipeline to replicate online databases to the data lake.”

“We are using Apache Hudi to incrementally ingest changelogs from Kafka to create data-lake tables. Apache Hudi is a unified Data Lake platform for performing both batch and stream processing over Data Lakes. Apache Hudi comes with a full-featured out-of-box Spark-based ingestion system called Deltastreamer with first-class Kafka integration and exactly-once writes. Unlike immutable data, our CDC data has a fairly significant proportion of updates and deletes. Hudi Deltastreamer takes advantage of its pluggable, record-level indexes to perform fast and efficient upserts on the data lake table."

Zendesk

Cloud-based customer service provider Zendesk also uses CDC extensively with their Hudi lakehouse on Amazon. 

“The data lake pipelines consolidate the data from Zendesk’s highly distributed databases into a data lake for analysis.

Zendesk uses Amazon Database Migration Service (AWS DMS) for change data capture (CDC) from over 1,800 Amazon Aurora MySQL databases in eight AWS Regions. It detects transaction changes and applies them to the data lake using Amazon EMR and Hudi.

Zendesk ticket data consists of over 10 billion events and petabytes of data. The data lake files in Amazon S3 are transformed and stored in Apache Hudi format and registered on the AWS Glue catalog to be available as data lake tables for analytics querying and consumption via Amazon Athena.”

GE Aviation

GE Aviation also uses Apache Hudi to manage CDC pipelines, enabling rapid increases in scale. 

“The introduction of a more seamless Apache Hudi experience within AWS has been a big win for our team. We’ve been busy incorporating Hudi into our CDC transaction pipelines and are thrilled with the results. We’re able to spend less time writing code for managing the storage of our data and more time focusing on the reliability of our system. This has been critical in our ability to scale. Our development pipeline has grown beyond 10,000 tables and more than 150 source systems as we approach another major production cutover.”

A Community that Innovates

Finally, given how quickly data lakehouse technologies are evolving, it's important to consider where open source innovation in this space has come from. Below are a few foundational ideas and features that originated in Hudi and that are now being adopted into the other projects. 

  Hudi OSS Community Innovation    Equivalent Feature 
  Transactional updates (March 2017)   Delta OSS  (April 2019)
  Merge On Read (Oct 2017)   Iceberg (Aug 2021, v2 format approval)
  Incremental Queries (March 2017)   Delta Change Feed OSS 2.x (June 2022)
  Z-order/Hilbert Space Curves (Dec 2021)   Delta OSS 2.x (June 2022)

In fact, outside of the table metadata (file listings, column stats) support, the Hudi community has pioneered most of the critical features that make up today’s data lakehouses. The community has supported over 1500 user issues and 5500+ Slack support threads over the last four years and is rapidly growing stronger, with an ambitious vision ahead. Users can consider this track record of innovation as a leading indicator for the future.

Conclusion

When choosing the technology for your data lakehouse it is important to perform an evaluation for your own use cases. Feature comparison spreadsheets and benchmarks should not be the deciding factor, so we simply hope that this blog post provides a starting point and reference for you in your decision-making process. Apache Hudi is innovative, battle-hardened and here to stay. Join us on Hudi Slack where you can ask questions and collaborate with the vibrant Apache Hudi community from around the globe.

If you would like one-to-one consultation to dive deep into your use cases and architecture, feel free to reach out at info@onehouse.ai. At Onehouse we have decades of experience designing, building, and operating some of the largest distributed data systems in the world. We recognize that these technologies are complex and rapidly evolving. Also, it’s possible that we missed a feature or could have read the documentation more closely on some of the above comparisons. Please drop a note to info@onehouse.ai if you see that any of the comparisons above are in need of correction, so we can keep the facts in this article accurate.

Update Notes

8/11/22 - Original publish date
1/11/23 - Refresh feature comparisons, added community stats + benchmarks
1/12/23 - Databricks contributed few minor corrections
10/31/23 - Minor edits
1/31/24 - Minor update about current state of OneTable

Read More:

Subscribe to the Blog

Be the first to read new posts

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.