Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Xet is on the Hub blog post #2744

Merged
merged 8 commits into from
Mar 18, 2025
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Next Next commit
initial commit for xet-on-the-hub
jsulz committed Mar 14, 2025
commit 61eb6c8162b9818c9847abe91bd06670ff117cbf
14 changes: 13 additions & 1 deletion _blog.yml
Original file line number Diff line number Diff line change
@@ -5683,4 +5683,16 @@
- llm
- vlm
- community
- research
- research

- local: xet-on-the-hub
title: "Xet is on the Hub"
author: jsulz
thumbnail: /blog/assets/xet-on-the-hub/thumbnail.png
date: March 17, 2025
tags:
- dedupe
- storage
- content defined chunking
- infrastructure
- content addressed store
Binary file added assets/xet-on-the-hub/thumbnail.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
194 changes: 194 additions & 0 deletions xet-on-the-hub.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,194 @@
---
title: "Xet is on the Hub"
thumbnail: /blog/assets/xet-on-the-hub/thumbnail.png
authors:
- user: assafvayner
org: xet-team
- user: brianronan
org: xet-team
- user: seanses
org: xet-team
- user: jgodlewski
org: xet-team
- user: sirahd
org: xet-team
- user: jsulz
org: xet-team
---

# Xet is on the Hub

<div
class="
p-6 mb-4 rounded-lg
border-2 border-gray-100
pt-6 sm:pt-9
bg-gradient-to-t
from-purple-500
dark:from-purple-500/20
"
>
Want to skip the details and get straight to faster uploads and downloads with bigger files than ever before?

[Click here](#ready-xet-go) to read about joining the Xet waitlist (or [head over to join immediately](https://huggingface.co/join/xet)).
</div>

Over the past few weeks, Hugging Face’s [Xet Team](https://huggingface.co/xet-team) took a major step forward by [migrating the first Model and Dataset repositories off LFS and to Xet storage](https://huggingface.co/posts/jsulz/911431940353906).

This marks one of many steps to [fulfill Hugging Face’s vision for the Hub](https://huggingface.co/blog/xethub-joins-hf) by empowering AI builders to build, iterate, and collaborate more effectively on massive models and datasets.

We've covered a lot of the background technology in a series of posts - check them out if you're interested in deeper dives:
* [From Files to Chunks: Improving Hugging Face Storage Efficiency](https://huggingface.co/blog/from-files-to-chunks)
* [Rearchitecting Hugging Face Uploads and Downloads](https://huggingface.co/blog/rearchitecting-uploads-and-downloads)
* [From Chunks to Blocks: Accelerating Uploads and Downloads on the Hub](https://huggingface.co/blog/from-chunks-to-blocks)

This migration shifted \~6% of the Hub’s download traffic onto Xet infrastructure, validating several integral components and testing integrations with the myriad of ways repositories are accessed (e.g., via local development environments, different libraries, CI systems, cloud platforms, etc.). With nearly two million developers working on over two million public repositories, it’s real-world usage that’s the ultimate proving ground.

Engineering a complex system like Xet storage is a balancing act. You plan for scale, performance, and reliability, but once bytes start moving, challenges emerge. The trick is knowing when to move from design and theory to practice. This is a behind-the-scenes view of the migration, taking you through our proof-of-concept to launching on the Hub.

## The Xet Difference

[LFS](https://huggingface.co/docs/hub/en/repositories-getting-started#requirements), the storage system behind repositories today, stores large files in a separate object storage outside of the repository. LFS deduplicates at the file level. Even tiny edits create a new revision to upload in full; painful for the multi-gigabyte files found in many Hub repositories.

In contrast, Xet storage uses [content-defined chunking (CDC)](https://huggingface.co/blog/from-files-to-chunks), to deduplicate at the level of bytes (\~64KB chunks of data). If you edit one row in a Parquet file or tweak a small piece of metadata in a GGUF model, only those changed chunks are sent over the wire.

This provides significant benefits to large file transfers. For example, the Xet team keeps an internal \~5 GB SQLite database that powers a metrics dashboard. With LFS, appending 1 MB (around the average update currently) requires re-uploading the entire 5 GB file. With Xet storage, we push only the new data.

<figure class="image text-center">
<img class="mx-auto" src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/sqlite-dedupe.png" alt="SQLite Dedupe from one version to another" width=30%>
<figcaption>Comparison of two recent runs with the <a href="https://github.com/huggingface/dedupe_estimator">dedupe_estimator</a> (a tool for estimating the chunk-level deduplication available between files). Red chunks indicate what needs to be uploaded when using Xet.</figcaption>
</figure>

This means uploads for the dashboard take a tenth of a second (at 50Mb/s) instead of the 13 minutes if the database were backed by LFS.

Making this work requires coordination between the following components:

1. A Xet-aware client
* Splits data into \~64 KB chunks.
* Deduplicates identical chunks locally and aggregates them into \~64 MB blocks before upload.
2. Hugging Face Hub
* Receives client requests and routes them to the correct repository.
* Provides authentication and security guarantees
3. The content addressed store (CAS)
* Enforces chunk-based deduplication for uploads and downloads.
* Includes an LFS Bridge for older, non-Xet clients, acting like a traditional LFS server and providing a single URL for downloads.
4. Amazon S3
* Stores file contents ([blocks](https://huggingface.co/blog/from-chunks-to-blocks#scaling-deduplication-with-aggregation)) and file reconstruction metadata ([shards](https://huggingface.co/blog/from-chunks-to-blocks#scaling-deduplication-with-aggregation)).
* Provides the final persisted layer of data referenced by the CAS

Before moving to production, the system was launched into an ephemeral environment, where we built a [steel thread of functionality](https://www.rubick.com/steel-threads/) to enable uploading and downloading via the `huggingface_hub`. The video below shows our progress one month in:

<video alt="Steel thread of Xet storage" controls>
<source src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/steel-thread.mp4" type="video/mp4">
</video>

After the heady highs of a quick proof-of-concept, the team settled into battling gnarly integration points (privacy, backward compatibility, fragmentation, etc) with the complex ecosystem that is the Hub. Eventually, the infrastructure moved into production for Hugging Face employees. With real usage now rolling in, we moved forward with the first large-scale migration.

## Migration Day

On February 20th at 10 a.m., it was all hands on deck. In the weeks prior the team had built the internal tooling to migrate files from LFS to Xet and, just as importantly, roll back a repository to LFS if necessary. Now it was time to put it to the test. Grafana and Kibana lit up with new logs and metrics, showing the load shift in real time.

<figure class="image text-center">
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/grafana-requests.png" alt="Grafana the day of the migration">
<figcaption>Grafana view of requests coming into Xet.</figcaption>
</figure>

The picture is the same in Kibana, only here you can tell when we took a lunch break:

<figure class="image text-center">
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/kibana-requests.png" alt="Grafana the day of the migration">
<figcaption>Kibana view of requests coming into Xet.</figcaption>
</figure>

By the end of the day, we had successfully migrated all target repositories for a total of 4.5 TB into Xet storage. Downloads from each repository were now served by our infrastructure.

## Post-Migration Challenges

No launch is without its lessons. The system handled the load smoothly and migrations went off without any major disruptions, but as the bytes flowed through various components we began to see some trends.

### Download Overhead from Block Format

Reviewing network throughput post-migration, we saw that CAS was downloading four times more data than it was returning to clients:

<figure class="image text-center">
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/pre-block-format-change.png" alt="Grafana throughput metrics before block format change">
<figcaption>Downloads (rx) in yellow, amount sent to clients (tx) in green.</figcaption>
</figure>

Upon further investigation we found the vast majority of requests to CAS were for 10MB ranges of data within files. These requests come from users that are using `hf_transfer` to accelerate their downloads of large files. The requested ranges do not align on block boundaries and the block format doesn’t provide uncompressed chunk lengths to facilitate skipping directly to the data within the block. Instead, CAS streams the block starting at the beginning and reads until it finds the requested data.

Given a block with an average size of 60MB and multiple partial requests to ranges within that block, the overhead accumulates. CAS reads 210MB for just 60MB delivered, a ratio of 3.5 bytes downloaded : 1 byte sent to the client. At scale across blocks of irregular sizes and ranges that can span several blocks, the 4 bytes downloaded : 1 byte sent ratio is matched.

As usual, an image is worth a thousand words:

<figure class="image text-center">
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/block-format-download-problem.png" alt="Graphical representation of block format issue">
<figcaption>Red blocks are unnecessarily streamed content while green blocks show the data CAS is actually requesting and ultimately sends.</figcaption>
</figure>

To fix this, we updated the block format to store chunk-length metadata, enabling CAS to download only what each request actually needs. This format change required coordinated updates across:

* CAS APIs
* Block metadata format
* All 65,000+ blocks in S3

All of which were undertaken without downtime. Not only did this result in a balanced ratio in CAS downloads to returned data, but also reduced GET latency by \~35% across the board.

<figure class="image text-center">
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/post-block-format-change.png" alt="Grafana throughput metrics after block format change">
<figcaption>Balanced view of downloaded and sent data with reduced throughput across the board.</figcaption>
</figure>

### Pod Load Imbalance

At the same time, we also saw an unexpected load imbalance across our CAS cluster. One pod would spike to hundreds of active uploads while other pods chugged along at single digits. Swapping the load balancer’s [routing algorithm](https://docs.aws.amazon.com/elasticloadbalancing/latest/userguide/how-elastic-load-balancing-works.html#routing-algorithm) (from [round-robin to least outstanding requests](https://medium.com/dazn-tech/aws-application-load-balancer-algorithms-765be2eca158)) and adding more nodes made little difference in the moment.

Digging deeper we realized that during uploads, CAS writes to a temporary file for validation before uploading without calling `fsync`, allowing the OS to buffer writes in the page cache. When a spike of uploads comes in, the unflushed page cache leads to increased memory pressure.

As the OS tries to flush the page cache to disk, the pod experiences throughput throttling from [block storage](https://aws.amazon.com/ebs/), causing increased latency on uploads. While uploads slowly process, new requests continue to come in, pushing more to the disk. This creates a vicious cycle, leading to a backlog of uploads and the imbalance that we saw during the migration:

<figure class="image text-center">
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/pod-load-imbalance.png" alt="Pod load imbalance">
<figcaption>Balanced view of downloaded and sent data with reduced throughput across the board.</figcaption>
</figure>

To keep a pod from getting in this state, we put a limit to the number of concurrent uploads each machine accepts before rejecting requests that, when reached, pushes requests on to other pods in the cluster. If all pods in the cluster get in this state, then our autoscaling policies kick in. Longer-term solutions might include forcing disk syncs, removing the on-disk buffering, or switching to ephemeral storage with better throughput.

## Takeaways

These two issues led to important architectural improvements, but they weren’t the only challenges the team tackled in the weeks following the migration:

* **Memory Leak** \- In the first few days after the migration, we noticed CAS memory usage increasing in a stepwise pattern, hinting at a leak which was fixed after investigation.
* **Slow LFS Bridge Downloads** \- While investigating prolonged download times (a reminder to always keep an eye on tail latencies), we uncovered particularly troublesome cases. Interestingly, fixing the memory leak also resolved these slow downloads.
* **Switch from ALB to NLB** \- We eventually migrated the LFS Bridge from an [Application Load Balancer](https://docs.aws.amazon.com/elasticloadbalancing/latest/application/introduction.html) (ALB) to a [Network Load Balancer](https://docs.aws.amazon.com/elasticloadbalancing/latest/network/introduction.html) (NLB), improving scalability during bursts of traffic. The DNS cutover to the NLB was performed with zero downtime.

Looking back, these issues underscore a set of critical lessons:

* No test environment can simulate user behavior at scale. Even after careful integration work and Hugging Face employees testing the infrastructure for months, corner cases only surfaced once we funneled real usage through the system.
* By staging migrations incrementally and uncovering these issues before more traffic was on the infrastructure, we avoided downtime and disruption. Managing a fraction of traffic and storage made it relatively easy in comparison to if all the Hub was on Xet from day one.
* The infrastructure and system design was hardened iteratively over weeks. Every future byte and network request on Xet will receive the benefits of these lessons.

In short, real-world load was essential for exposing these challenges, and incremental migrations let us tackle them safely. That’s how we’re ensuring Xet-backed repositories scale reliably for the entire Hugging Face community with faster uploads, fewer bottlenecks, and minimal disruption all the way.

### Ready. Xet. Go.

With the initial migration complete, Xet is now on the Hugging Face Hub. We want to open up access more broadly, so you can feel the ease and speed of Xet storage.

<p align="center">
<img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/blog/xet-on-the-hub/migration.gif" alt="Xet speed" width=40%>
</p>

We’re finalizing an official integration with the `huggingface_hub` that will mean you get the benefits of Xet without any significant changes to your current workflow. To get going:

1. **Join the Waitlist**
* [Sign up here](https://huggingface.co/join-xet) with your user or organization account (if you have admin/write permissions).
* Once accepted, your new repositories will automatically leverage Xet-backed storage.
* We’ll also migrate existing repositories, so you don’t miss out on faster transfers.
2. **Use Your Usual Tools**
* An `hf_xet` Python package is coming soon to `huggingface_hub`. If you use the `transformers` or `datasets` libraries, it's already using `huggingface_hub` so you can simply install `hf_xet` in the same environment. Installation steps and docs are on their way.
* Continue downloading and uploading large files with the same commands you already use.
3. **Enjoy the Benefits**
* Less waiting on uploads and downloads and faster iterations on big files.
* Everyone on your team should upgrade to `hf_xet` to get the full benefits, but legacy clients will stay compatible via the LFS Bridge.

If you have any questions, drop us a line in the comments 👇 or [open a discussion](https://huggingface.co/spaces/xet-team/README/discussions/new) on the [Xet team](https://huggingface.co/xet-team) page. And stay tuned as we roll out more pieces (faster web uploads, faster Git workflows), bringing faster big-file collaboration to every AI builder on Hugging Face\!