Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
LABS
Glossary

Data Availability Sampling (DAS)

Data Availability Sampling (DAS) is a cryptographic technique that allows network participants to verify the availability of all data in a block by probabilistically checking only small, random samples.
Chainscore © 2026
definition
BLOCKCHAIN SCALING

What is Data Availability Sampling (DAS)?

A cryptographic technique that allows light nodes to verify the availability of large data blocks without downloading them in full, a core innovation for scaling blockchain data layers.

Data Availability Sampling (DAS) is a cryptographic protocol that enables network participants, such as light clients, to probabilistically verify that all data for a block is published and accessible by downloading only a small, random subset of it. Instead of requiring every node to download an entire, massive data block—a major bottleneck for scalability—DAS uses erasure coding and random sampling to provide a high statistical guarantee of data availability. This is foundational for data availability layers like Celestia and Ethereum's danksharding roadmap, as it allows the network to securely scale block size while keeping verification lightweight.

The protocol relies on two key components: erasure coding and commitments. First, block data is expanded using an erasure code (e.g., Reed-Solomon), which redundantly encodes the data so that the original can be reconstructed even if a significant portion (e.g., 50%) is missing. This encoded data is then committed to using a Merkle root or a KZG polynomial commitment. Light clients then randomly select a small number of positions within this committed data and request the corresponding data chunks and proofs from the network. If all sampled chunks are available, the client can be statistically confident the entire block data is available.

For the system to be secure, DAS requires an honest minority assumption: a sufficient number of participants must perform sampling to ensure that if any data is withheld, it will be detected with overwhelming probability. The security guarantee is probabilistic; as a node conducts more random samples, its confidence that the data is fully available approaches certainty. This creates a scalable and secure bridge between full nodes, which store all data, and light clients, which perform minimal work. A failure in sampling—where requested data is unavailable—serves as a fraud proof that the block producer is malicious.

The primary application of DAS is in modular blockchain architectures, where the execution, consensus, and data availability functions are separated. In such designs, rollups or other execution layers post their transaction data to a dedicated data availability layer. Light nodes on the execution layer can use DAS to trustlessly verify that this critical data is available for anyone to download, which is necessary for fraud proofs or rebuilding state, without relying on a centralized data provider. This enables highly scalable execution layers without compromising on decentralization or security.

DAS is often contrasted with older data availability solutions like data availability committees (DACs) or simple full-node replication. While committees provide a weaker trust assumption based on a known group, DAS provides a cryptoeconomic guarantee that is secure as long as a sufficient number of independent, honest samplers exist in the permissionless network. Its development is central to solving the data availability problem, which asks: "How can we ensure that data needed to validate a block is actually published?" DAS provides a scalable answer, forming the bedrock for the next generation of high-throughput blockchains.

how-it-works
MECHANISM

How Data Availability Sampling Works

Data Availability Sampling (DAS) is a cryptographic technique that allows a network of light nodes to probabilistically verify that all data for a block is published and accessible without downloading the entire dataset.

Data Availability Sampling (DAS) is a cornerstone protocol for scaling blockchains through data availability layers and modular architectures. Its core function is to solve the data availability problem: ensuring that a block producer has made all transaction data public so that the network can verify state transitions and detect invalid transactions. Instead of requiring every node to download an entire block—which becomes impractical with large data blocks—DAS enables light clients or sampling nodes to perform random checks on small, randomly selected pieces of the data. If a sufficient number of these random samples can be successfully retrieved, the node gains high statistical confidence that the entire dataset is available.

The process relies on encoding the block data using an erasure coding scheme like Reed-Solomon. This transforms the original data into extended data with redundancy, so that the original can be reconstructed from any sufficient subset of the coded pieces. The block producer commits to this extended data using a Merkle root or a more advanced polynomial commitment like a KZG commitment. Samplers then randomly select indices and request the data chunk at that index along with a Merkle proof against the published commitment. A successful retrieval of the proof validates that specific piece; repeated failures to retrieve proofs for sampled indices signal that data is being withheld.

For the system to be secure, sampling must be unpredictable and conducted by a sufficiently large, decentralized set of nodes. Protocols like Ethereum's DankSharding envision a network where thousands of light nodes each perform a small, fixed amount of sampling work. The security property emerges collectively: if any meaningful portion of the data is hidden, a random sampler has a proportional chance of detecting it. After many independent samples, the probability of all samplers missing the missing data becomes astronomically low, providing probabilistic security that the data is available.

This mechanism is fundamental to validium and volition scaling solutions, which keep data off the main chain, and to celestia-style data availability layers. It also underpins Ethereum's proto-danksharding (EIP-4844) and full danksharding roadmap, where it allows the network to securely scale block data into the megabytes or gigabytes. By separating data availability verification from execution, DAS enables a more efficient division of labor in the blockchain stack, allowing for high throughput without requiring all participants to process all data.

key-features
CORE MECHANICS

Key Features of DAS

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to efficiently verify that all data for a block is published without downloading it entirely. This is achieved through probabilistic sampling of small, random chunks of data.

01

Probabilistic Verification

Instead of downloading the entire block, a light client randomly selects and downloads a small number of data chunks (e.g., 30 out of 4096). The probability of missing unavailable data decreases exponentially with each sample, making verification highly efficient and secure with minimal data transfer.

02

Erasure Coding

Block data is expanded using an erasure code (like Reed-Solomon) before being split into chunks. This creates redundancy, allowing the original data to be reconstructed even if up to 50% of the chunks are missing. This is the foundation that makes sampling possible and secure.

03

2D Reed-Solomon Encoding

A specific implementation used by systems like Celestia and Ethereum's Proto-Danksharding. Data is arranged in a matrix and encoded in both rows and columns. This structure allows for highly efficient sampling and robust recovery, even if large, contiguous sections of data are withheld.

04

KZG Commitments

A polynomial commitment scheme (Kate/KZG) is often used to create a compact cryptographic proof (a commitment) for the erasure-coded data. This allows samplers to verify that each downloaded chunk is consistent with the overall data promise without needing the full dataset.

05

Light Client Scalability

DAS enables light clients to participate in consensus-level security with minimal resources. They can independently verify data availability, breaking the reliance on full nodes for this trust assumption. This is critical for scaling blockchains while maintaining decentralization.

06

Enabling Secure Rollups

DAS is the security backbone for optimistic rollups and validiums. It guarantees that the transaction data necessary to reconstruct a rollup's state is published on-chain, allowing anyone to verify correctness or challenge fraud proofs, without forcing all data onto the base layer.

visual-explainer
MECHANISM OVERVIEW

Visualizing the Data Availability Sampling (DAS) Process

A step-by-step walkthrough of how Data Availability Sampling enables light clients and nodes to efficiently verify that block data is published without downloading it in full.

The Data Availability Sampling (DAS) process begins when a block producer publishes a new block. Instead of transmitting the entire, potentially massive data block to all participants, the producer encodes the data using an erasure coding scheme like Reed-Solomon. This transforms the original data into a larger set of data shares, where only a subset is needed for full reconstruction. The encoded shares are then arranged into a two-dimensional matrix, forming the foundation for the sampling process.

Light clients and full nodes performing DAS do not download this entire matrix. Instead, they randomly select and request a small number of distinct data shares from the network. This is the core "sampling" action. For each requested share, the node receives a cryptographic proof—typically a Merkle proof—that verifies the share is part of the correctly encoded data block. By successfully sampling a statistically significant number of random shares, a node gains high probabilistic certainty that the entire dataset is available.

The security model relies on the properties of erasure coding. If any part of the data is withheld, a significant portion of the randomly sampled shares will be unavailable or fail their proofs. After a sufficient number of successful samples (e.g., 30-50), the probability that a malicious block producer has hidden data becomes astronomically low. This allows resource-constrained devices to act as light clients that securely verify data availability, a critical innovation for scaling solutions like Ethereum danksharding and Celestia.

In practical implementations, nodes often coordinate sampling through a Distributed Hash Table (DHT) or a peer-to-peer network dedicated to serving data shares. Advanced systems may use KZG polynomial commitments to generate more efficient proofs for each share. The entire process is designed to be parallelizable and non-interactive, enabling thousands of nodes to perform sampling simultaneously with minimal communication overhead, thus securing the network's data layer without imposing full data download requirements.

ecosystem-usage
DATA AVAILABILITY SAMPLING

Ecosystem Usage & Implementations

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to verify that all data for a block is published without downloading it entirely. Its implementation is critical for scaling blockchains securely.

06

The Sampling Process & Security Guarantees

The core operational mechanism involves:

  1. Erasure Coding: Block data is expanded (e.g., from k to 2k chunks).
  2. Random Sampling: A light node randomly requests a small, fixed number of these chunks (e.g., 30).
  3. Statistical Security: The probability of missing unavailable data decreases exponentially with more samples. If all samples are returned, the node concludes with high probability that the full data is available.
security-considerations
DATA AVAILABILITY SAMPLING (DAS)

Security Considerations & Guarantees

Data Availability Sampling is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published and accessible, without downloading the entire dataset. This is a foundational security primitive for scaling solutions like validiums and rollups.

01

The Data Availability Problem

Block producers can withhold transaction data while still publishing a valid block header, creating a data withholding attack. This prevents nodes from verifying state transitions, allowing invalid transactions (e.g., double-spends) to be included. DAS directly solves this by ensuring data is available for verification.

02

How Sampling Provides Guarantees

Block data is erasure-coded and split into many small chunks. Light clients randomly sample a fixed number of these chunks (e.g., 30). Statistically, if the data is unavailable, a client will detect a missing chunk with near-certain probability (e.g., 99.9% confidence). This allows for scalable security without full data download.

03

Erasure Coding Requirement

DAS requires data to be expanded using Reed-Solomon or similar erasure codes. This creates redundancy: the original data can be reconstructed from any 50% of the coded chunks. This ensures the network can tolerate up to 50% of chunks being missing or withheld before reconstruction becomes impossible, defining the security threshold.

04

Honest Majority Assumption

DAS security relies on an honest majority of sampling nodes. If a malicious block producer withholds data, they must also control >50% of the sampling nodes to provide fake proofs for missing chunks. The security model shifts from requiring all nodes to download all data to requiring a majority of light clients to be honest.

05

Comparison to Data Availability Committees (DACs)

  • DAS: Trust-minimized, cryptographic. Security derived from probabilistic sampling by a decentralized set of nodes.
  • DAC: Trusted, multisignature-based. Relies on a known committee to sign off on data availability. DAS provides stronger cryptoeconomic security guarantees, while DACs offer simpler implementation with a weaker trust model.
DATA VERIFICATION METHODS

DAS vs. Full Node Download

Comparison of resource requirements and trust assumptions between Data Availability Sampling (DAS) and traditional full block download for verifying blockchain data.

Feature / MetricData Availability Sampling (DAS)Full Node Download

Primary Function

Probabilistic verification of data availability

Deterministic download and execution of all data

Minimum Hardware

Light client (mobile phone, browser)

High-performance server with >2 TB SSD

Network Bandwidth

< 100 KB/s (sampling queries)

50 MB/s (sustained block sync)

Storage Requirement

Minimal (stores only samples and headers)

Full blockchain state history (>1 TB)

Trust Assumption

Honest majority of sampling nodes

None (fully self-verified)

Time to Verify Block

< 1 second (for availability)

Varies by chain (seconds to minutes)

Suitable For

End-users, wallets, light applications

Validators, exchanges, indexers

FAQ

Common Misconceptions About Data Availability Sampling (DAS)

Data Availability Sampling (DAS) is a critical scaling technology, but its mechanics are often misunderstood. This section clarifies the most frequent points of confusion.

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published and available, without downloading the entire block. It works by having nodes randomly sample and download small, random chunks (or erasure-coded shares) of the block data. If a node can successfully retrieve all its requested samples, it gains high statistical confidence that the entire data set is available. This process is repeated by many nodes across the network, creating a robust, trust-minimized guarantee that data can be reconstructed if needed, which is foundational for fraud proofs and validity proofs in scaling solutions like Ethereum danksharding and Celestia.

DATA AVAILABILITY SAMPLING

Frequently Asked Questions (FAQ)

Common questions about Data Availability Sampling (DAS), a critical scaling technology that allows light nodes to securely verify that block data is available without downloading it entirely.

Data Availability Sampling (DAS) is a cryptographic technique that allows a node to verify with high statistical certainty that all data in a block is available for download, without having to download the entire block. It works by having the block producer encode the data using erasure coding (like Reed-Solomon), expanding it into data "shards." Light clients or validators then randomly sample a small number of these shards. If all sampled shards can be retrieved, the probability that a significant portion of the data is missing becomes astronomically low. This enables secure, trust-minimized scaling by separating data availability verification from full data processing.

ENQUIRY

Get In Touch
today.

Our experts will offer a free quote and a 30min call to discuss your project.

NDA Protected
24h Response
Directly to Engineering Team
10+
Protocols Shipped
$20M+
TVL Overall
NDA Protected Directly to Engineering Team