Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
LABS
Glossary

Data Availability Sampling (DAS)

Data Availability Sampling (DAS) is a cryptographic technique where light clients verify the full availability of a large data block by randomly sampling and checking small portions of it.
Chainscore © 2026
definition
BLOCKCHAIN SCALING MECHANISM

What is Data Availability Sampling (DAS)?

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to efficiently verify that all transaction data for a new block is published and available, without downloading the entire block.

Data Availability Sampling (DAS) is a core scaling mechanism for blockchain scalability solutions, particularly modular blockchains and layer-2 rollups. It solves the data availability problem, which is the challenge of ensuring that all data necessary to reconstruct a block's state is actually published to the network. Without this guarantee, a malicious block producer could withhold data, making it impossible for others to validate transactions or detect fraud. DAS enables light clients or validators to perform random checks on small pieces of the data, providing high statistical confidence that the entire dataset is available, a process far more efficient than downloading the full block.

The protocol works by having the block producer erasure-code the block data, expanding it with redundant pieces. This data is then arranged in an extension of a Merkle tree, often a 2D Reed-Solomon erasure coding scheme with a KZG polynomial commitment or similar. Light nodes then randomly select a small, fixed number of these data chunks and query the network for proofs that they were correctly included. By sampling multiple independent points, a node can achieve exponentially high certainty—for example, sampling 30 chunks can provide 99.9% confidence—that the entire data is available. This allows the network to securely increase block sizes without requiring every participant to process all the data.

DAS is a foundational technology for data availability layers like Celestia and EigenDA, and is integral to the security model of optimistic rollups and zk-rollups. For rollups, posting transaction data to a robust data availability layer via DAS is often more secure and decentralized than using a calldata on a monolithic chain like Ethereum. The implementation of DAS, alongside technologies like Danksharding on Ethereum, is critical for enabling secure scaling where nodes can participate in consensus without the prohibitive resource requirements of storing and processing the entire blockchain history, paving the way for higher throughput networks.

how-it-works
BLOCKCHAIN SCALING MECHANISM

How Does Data Availability Sampling Work?

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published and accessible, without downloading the entire dataset.

Data Availability Sampling (DAS) is a core scaling mechanism for blockchains using data availability layers or modular architectures. It solves the data availability problem, where a block producer might withhold transaction data after publishing a block header, making it impossible for others to verify or reconstruct the chain's state. Instead of requiring every node to download all data—a major bottleneck for scalability—DAS enables light clients to perform random checks on small portions of the data, providing high statistical confidence that the complete data is available.

The protocol relies on erasure coding, where the original block data is expanded into coded chunks with redundancy. A key property is that if any sufficient subset of these chunks is available, the entire original data can be recovered. Light nodes perform random sampling by requesting a small, fixed number of these chunks at random indices from the network. If all requested samples are successfully retrieved, the node gains confidence the data is available. The probability of a malicious block producer successfully hiding the data while passing many independent random checks becomes astronomically low.

For the system to be secure, sampling must be non-interactive and publicly verifiable. This is often achieved using KZG polynomial commitments or similar cryptographic schemes, which allow a node to verify that a provided data chunk correctly corresponds to the committed data in the block header without needing the full dataset. The sampling process is repeated over multiple rounds by many independent nodes, creating a robust, decentralized audit of data availability. This collective verification is fundamental to validiums and volitions, and is a prerequisite for secure ZK-rollup operation.

In practice, networks implementing DAS, such as Celestia or EigenDA, structure their node networks into full nodes that store all data and light nodes that perform sampling. As a light node completes its sampling rounds with successful responses, it accepts the block header. If a sample request fails, the node initiates a data availability challenge, alerting the network to potential malicious behavior. This design dramatically reduces the hardware requirements for participants while maintaining strong security guarantees, enabling blockchain scalability to thousands of transactions per second without compromising decentralization.

key-features
MECHANICS

Key Features of Data Availability Sampling

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published and accessible without downloading it entirely. This is a core innovation enabling secure and scalable blockchain scaling solutions like danksharding.

01

Probabilistic Guarantee

Instead of downloading an entire block (which can be large), a node randomly samples small chunks of data. By performing enough random samples, the node can achieve a statistically high confidence (e.g., 99.9%) that all data is available. This transforms a deterministic problem into a probabilistic one, drastically reducing the resource requirements for verification.

02

Erasure Coding

DAS relies on erasure coding (e.g., Reed-Solomon codes) to make data redundant. The original data is expanded into coded chunks, so that only a subset (e.g., 50%) is needed to reconstruct the whole. This ensures data remains recoverable even if some chunks are withheld, making it cryptoeconomically expensive for a malicious actor to successfully hide data.

03

2D KZG Commitments

A KZG polynomial commitment is a cryptographic proof that binds a prover to a specific piece of data without revealing it. In advanced DAS schemes like danksharding, a 2D KZG commitment is used. This creates a matrix of commitments, allowing samplers to verify the correctness of any single data chunk with a tiny, constant-sized proof, ensuring the sampled data is consistent with the block header.

04

Light Client Security

DAS is the foundation for secure light clients in high-throughput systems. It allows a phone or browser wallet to independently verify data availability, moving beyond the trust model of simply following the longest chain. This is critical for bridges and cross-chain applications, as it prevents acceptance of blocks where data is hidden—a common attack vector for stealing bridged funds.

05

Sampling Network

Effective DAS requires a decentralized network of samplers. Each participant (node) performs independent random sampling. Through a gossip protocol, samplers can alert the network if they cannot retrieve a requested chunk. If enough samples fail, the network can reject the block before it is finalized, creating a robust, collaborative defense against data withholding attacks.

06

Enabling Danksharding

DAS is the key innovation that makes danksharding on Ethereum feasible. Danksharding proposes massive blocks (~128 MB). Full nodes cannot process this, but through DAS, thousands of light nodes can collectively ensure the data is available for rollups to use. The Data Availability Committee (DAC) model is a simpler, trusted precursor, while DAS provides a trust-minimized, native solution.

ecosystem-usage
DATA AVAILABILITY SAMPLING

Ecosystem Usage & Implementations

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to efficiently verify that all data for a block is published and available without downloading the entire dataset. Its implementation is foundational for scaling blockchains securely.

04

Light Client Verification

The core user-facing benefit of DAS is enabling trust-minimized light clients. Instead of trusting a full node or a centralized RPC, a light client can:

  • Download only block headers.
  • Perform random sampling on small portions of the erasure-coded data.
  • Statistically guarantee (with high probability) that the entire data block is available. This is critical for mobile wallets and cross-chain bridges to operate securely without running full nodes.
05

Enabling Validiums and Volitions

DAS is the enabling technology for Validium and Volition scaling solutions. These are L2s where transaction data is posted off-chain to a DA layer (using DAS) instead of directly to Ethereum L1.

  • Validium: Security relies entirely on the DA layer's DAS proofs.
  • Volition: Users choose per-transaction whether data goes to Ethereum (as a ZK-Rollup) or a DA layer (as a Validium). This trade-off offers higher throughput and lower cost, with security dependent on the chosen DA layer.
06

The Sampling Process in Practice

A practical DAS implementation involves a defined workflow:

  1. Erasure Coding: The original data block is expanded using Reed-Solomon codes, creating redundant pieces.
  2. Distribution: These pieces are distributed across the network of full nodes or a dedicated committee.
  3. Random Sampling: Light clients request a small, random set of these pieces.
  4. Statistical Security: By successfully sampling a sufficient number of unique pieces (e.g., 30), the client can be statistically certain (e.g., >99.99%) the entire data is available. Failed sampling triggers a fraud proof challenge.
PROTOCOL MECHANISMS

Comparison with Other Data Availability Solutions

A technical comparison of data availability sampling against alternative approaches for ensuring data is published to the blockchain.

Feature / MetricData Availability Sampling (DAS)Committee-Based AttestationData Availability Committee (DAC)On-Chain Publication (Full Nodes)

Primary Security Model

Probabilistic Sampling & Erasure Coding

Cryptoeconomic Slashing of Committee

Multi-Signature Trust

Full Node Download & Verification

Scalability (Bandwidth per Node)

O(√N) to O(log N)

O(N) (Committee Size)

O(1) (Trusted Committee)

O(N) (Full Block)

Trust Assumptions

1-of-N Honest Light Nodes

Honest Majority of Committee

Honest Majority of Committee Members

None (Fully Trustless)

Fault Proof Latency

Minutes to Hours (Sampling Period)

1-2 Epochs (Slashing Delay)

Committee Decision Time

Immediate (Next Block)

Client Hardware Requirements

Light Client (Mobile Feasible)

Validator Node

Signature Verification

Archive Node

Data Redundancy

High (Dispersed via Erasure Coding)

Moderate (Replicated in Committee)

Low (Centralized Storage)

Maximum (Global Replication)

Typical Use Case

Ethereum DankSharding, Celestia

Early Rollup Designs

Private/Consortium Rollups

Base Layer L1 (e.g., Ethereum Mainnet)

security-considerations
DATA AVAILABILITY SAMPLING

Security Considerations & Guarantees

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published, without downloading the entire dataset. This is a foundational security primitive for scaling solutions like danksharding and modular blockchains.

01

Core Security Guarantee

DAS provides a probabilistic guarantee of data availability. By randomly sampling small chunks of data, a node can be statistically confident the entire data is available. The probability of missing withheld data decreases exponentially with the number of samples, making it computationally infeasible for an adversary to hide data.

02

The 51% Attack Mitigation

In traditional blockchains, a majority of validators could withhold block data while building on the chain, creating a fraud proof that cannot be verified. DAS prevents this by enabling any light client to catch data withholding with high probability, breaking the assumption that malicious validators can hide data from the network.

03

Erasure Coding Requirement

DAS relies on the block data being erasure-coded. Data is expanded with redundancy (e.g., from 32 chunks to 64). This ensures that even if 50% of the data is withheld, the original data can be fully reconstructed from the remaining samples. Sampling without erasure coding cannot guarantee recoverability.

04

Light Client Security Model

DAS shifts security assumptions for light clients. Instead of trusting a majority of full nodes (honest-majority assumption), they rely on the cryptographic soundness of sampling and erasure coding. A client performing 30 random samples has a >99.9% chance of detecting missing data, enabling secure operation without syncing the full chain.

05

Data Availability Committees (DACs) vs. DAS

A Data Availability Committee is a trusted, permissioned set of signers that attest to data availability. DAS, in contrast, is a trust-minimized cryptographic solution. While DACs offer simpler implementation, DAS eliminates trust assumptions and is the long-term goal for fully decentralized scaling (e.g., Ethereum's danksharding).

06

Implementation & Sampling Networks

Practical DAS requires a peer-to-peer network for sample distribution and retrieval. Projects like Celestia implement this via a Data Availability Network where light nodes request random chunks from full nodes. The security depends on network connectivity and incentives for full nodes to serve data honestly.

visual-explainer
DATA AVAILABILITY SAMPLING

Visual Explainer: The DAS Process

A step-by-step breakdown of how Data Availability Sampling (DAS) enables light clients to securely verify that block data is published without downloading it entirely.

Data Availability Sampling (DAS) is a cryptographic technique that allows network participants, such as light clients or validators, to verify with high statistical certainty that all data for a block is available by downloading only a small, random subset. This process is foundational to scalable blockchain designs like Ethereum's danksharding, as it decouples data verification from full data processing, enabling networks to securely scale block sizes far beyond what any single node could store. The core mechanism relies on erasure coding the block data into extended pieces, which are then arranged in a format that allows for efficient random sampling.

The process begins when a block producer creates a new block and commits to its data. This data is erasure coded, transforming the original N chunks of data into 2N chunks, where any N chunks are sufficient to reconstruct the whole. These chunks are arranged into a two-dimensional matrix, often using a Reed-Solomon code and a KZG polynomial commitment or a Merkle root, creating a verifiable data structure. The commitment to this extended data is then published as part of the block header, providing a cryptographic fingerprint that light clients can reference during sampling.

A light client performing DAS does not download the entire block. Instead, it randomly selects a fixed number of coordinates (e.g., 20-30) within the data matrix and requests the data chunks at those specific locations from the network. For each request, it receives the data and a cryptographic proof (like a Merkle proof) that the chunk is part of the committed data. If the client can successfully retrieve and verify all its randomly sampled chunks, it can conclude with overwhelming probability—often exceeding 99.9%—that the entire dataset is available. This probabilistic security model is mathematically robust against data withholding attacks.

The system's security is game-theoretically sound. For a malicious block producer to successfully hide even a small portion of the data (e.g., 1%) from a client performing many random samples, they would need to correctly guess which specific chunks the client will request—a probability that becomes astronomically small. If a client fails to retrieve a sampled chunk, it issues a data availability challenge, alerting the network to a potential fault. This triggers a protocol where full nodes can reconstruct the data from the remaining available chunks and prove the block producer's malfeasance, leading to their slashing.

In practice, DAS enables the creation of extremely data-dense blocks (e.g., 16-32 MB in danksharding) that no single node needs to store in full, while maintaining the security guarantee that the data exists for anyone who wishes to process it. This is critical for rollup scalability, as rollups post their transaction data to this available space. Light clients and bridges can trustlessly verify the availability of this data, securing cross-chain communication without relying on centralized intermediaries or expensive full nodes.

DEBUNKED

Common Misconceptions About Data Availability Sampling

Data Availability Sampling (DAS) is a critical scaling technology, but its technical nature leads to widespread misunderstandings. This glossary clarifies the core mechanics and limitations of DAS, separating the cryptographic reality from common hype and oversimplifications.

No, Data Availability Sampling (DAS) and data sharding are distinct but complementary concepts. Data sharding is the act of horizontally partitioning blockchain data into smaller, manageable pieces called shards. DAS is the light-client verification protocol that allows nodes to cryptographically confirm, with high probability, that all data within a shard is published and available without downloading it entirely. Sharding creates the partitioned data structure; DAS provides the trust-minimized method to ensure its availability. In architectures like Ethereum's danksharding roadmap, sharding provides the data 'fragments,' and DAS is the tool light nodes use to sample them.

DATA AVAILABILITY SAMPLING

Frequently Asked Questions (FAQ)

Data Availability Sampling (DAS) is a cryptographic technique that allows nodes to verify that all data for a block is published without downloading it entirely. These questions address its core purpose, mechanics, and role in scaling blockchains.

Data Availability Sampling (DAS) is a protocol that allows light nodes to probabilistically verify that all data for a new block is published and available for download, without having to download the entire dataset. It works by having the block producer erasure-code the block data, splitting it into smaller chunks. Light nodes then randomly sample a small, constant number of these chunks. If a node can successfully retrieve all its sampled chunks, it can be statistically confident that the entire data is available. This process is repeated over multiple rounds by many nodes to achieve high security guarantees. DAS is a core component of data availability layers like Celestia and Ethereum's proto-danksharding (EIP-4844).

ENQUIRY

Get In Touch
today.

Our experts will offer a free quote and a 30min call to discuss your project.

NDA Protected
24h Response
Directly to Engineering Team
10+
Protocols Shipped
$20M+
TVL Overall
NDA Protected Directly to Engineering Team
Data Availability Sampling (DAS) | Blockchain Glossary | ChainScore Glossary