Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
LABS
Glossary

Data Availability Sampling (DAS)

Data Availability Sampling (DAS) is a cryptographic technique that allows light clients to probabilistically verify the availability of block data by randomly sampling small chunks, eliminating the need to download the entire dataset.
Chainscore © 2026
definition
BLOCKCHAIN SCALING MECHANISM

What is Data Availability Sampling (DAS)?

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to efficiently verify that all transaction data for a block is published and available, without downloading the entire block.

Data Availability Sampling (DAS) is a core scaling mechanism for blockchain networks, particularly those using modular architectures that separate execution from consensus and data availability. Its primary function is to solve the data availability problem: ensuring that when a block producer publishes a new block, they have made all the underlying transaction data accessible to the network. Without this guarantee, a malicious producer could withhold data, making it impossible for others to validate or reconstruct the block's state, leading to security failures. DAS enables light clients or nodes to perform this verification with high confidence while consuming minimal resources.

The protocol works through a process of random sampling. When a block is produced, its data is erasure-coded—a process that expands the original data with redundant pieces. Light nodes then randomly select and download a small, fixed number of these coded pieces. Using cryptographic proofs, they can verify the authenticity of each piece. Statistically, if a node samples enough random pieces and finds them all available, it can conclude with near-certain probability that the entire dataset is available. This allows a large network of light nodes to collectively secure data availability with workloads that do not scale with the full size of the block data.

DAS is a foundational component for implementing data availability layers and volitions, and is central to the scalability roadmap of networks like Ethereum through proto-danksharding and full danksharding. By allowing nodes to securely operate without storing the entire blockchain history, DAS enables massive increases in block size and throughput while preserving the decentralized security model. It shifts the security assumption from "every node stores everything" to "the collective sampling power of the network ensures data is published," which is essential for scalable, secure Layer 2 rollups that post their transaction data to a base layer.

how-it-works
MECHANISM

How Does Data Availability Sampling Work?

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published and available, without downloading the entire dataset.

Data Availability Sampling (DAS) is a cornerstone protocol for scaling blockchains with data availability layers and modular architectures. Its core function is to solve the data availability problem: ensuring that block producers have actually published all the data for a new block, preventing them from hiding transaction data that could contain invalid state transitions. Instead of requiring every node to download an entire block's data—which is computationally expensive—DAS enables light clients or sampling nodes to check random, small chunks of the data. If a sufficient number of these random checks pass, the node can be statistically confident the entire dataset is available.

The process relies on encoding the block data using an erasure coding scheme like Reed-Solomon. This transforms the original data into a larger set of encoded pieces with redundancy. A key property is that the original data can be fully reconstructed from any sufficiently large subset of these pieces. Sampling nodes then request random data samples (small, unique pieces of the encoded data) from the network. They use cryptographic commitments, typically a Merkle root or a KZG polynomial commitment, to verify that each received sample is correct and part of the original block. If a node cannot retrieve a requested sample, it serves as proof that the data is not fully available.

For the system to be secure, sampling must be diligent and persistent. A single node performs dozens to hundreds of random sample requests per block. The probability of a node failing to detect missing data decreases exponentially with the number of successful samples. Furthermore, the protocol relies on a sufficiently large and honest subset of the sampling network. If enough independent nodes all successfully sample the data, the collective probability of a data withholding attack going undetected becomes astronomically low, providing probabilistic security that approaches certainty.

DAS is a critical innovation enabling scalable data availability layers, such as those proposed for Ethereum danksharding. It allows these layers to securely scale their block data size into the megabytes or even gigabytes, as the verification workload for individual nodes remains light and constant. This separates the cost of verifying data availability from the cost of storing it, a fundamental shift from monolithic blockchain designs. By providing a trust-minimized way to ensure data is published, DAS underpins the security of optimistic rollups and validiums, which depend on the availability of their transaction data for fraud proofs or forced inclusion.

key-features
CORE MECHANICS

Key Features of Data Availability Sampling (DAS)

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published and available, without downloading the entire dataset. This is the foundational technology enabling secure, scalable Layer 2 rollups and sharded blockchains.

01

Probabilistic Verification

Instead of downloading an entire block (which can be large), a light client or node downloads a small, random subset of data chunks (or erasure-coded shares). By sampling multiple independent chunks and receiving valid proofs, the node gains high statistical confidence that the entire dataset is available. This allows for secure verification with minimal data transfer.

02

Erasure Coding

Before sampling, block data is expanded using an erasure code (like Reed-Solomon). This process transforms the original data into a larger set of encoded pieces with redundancy. The key property: the original data can be reconstructed from any sufficient subset of these pieces. This ensures data remains recoverable even if some samples are missing or withheld.

03

2D Reed-Solomon & KZG Commitments

Advanced DAS implementations (e.g., Ethereum's Dankrad Feist) use a 2D Reed-Solomon scheme, arranging data in a matrix for more efficient sampling. Each row and column is committed to using a KZG polynomial commitment. This allows samplers to verify a single data chunk against a small, constant-sized cryptographic proof, ensuring the data is part of the originally committed block.

04

The Sampling Process

  1. Commit: The block producer publishes erasure-coded data and a KZG commitment (or Merkle root) to it.
  2. Sample: Each light node randomly selects multiple unique coordinates (e.g., specific row/column intersections in the 2D scheme).
  3. Request & Verify: The node requests the data chunk and a proof for each sample from the network.
  4. Statistical Security: After successfully verifying many random samples, the node concludes with high probability (>99.99%) that the entire data is available.
05

Enabling Data Availability Committees (DACs)

DAS is the trust-minimizing backbone for Data Availability Committees. Instead of blindly trusting signatures from a committee, nodes can use DAS to independently verify that the data the committee is attesting to is actually available on-chain. This shifts the security model from n-of-m trust to cryptographic and probabilistic guarantees.

06

Resource Requirements

  • Bandwidth: Minimal. Nodes download only a few kilobytes of samples per block, regardless of total block size (e.g., 50 KB vs. a 2 MB block).
  • Computation: Light. Verification involves a few elliptic curve pairings (for KZG) or hash operations (for Merkle proofs).
  • Time: Fast. Sampling can be done in parallel, allowing verification within seconds. This makes DAS feasible for resource-constrained devices, enabling truly decentralized validation.
ecosystem-usage
DATA AVAILABILITY SAMPLING

Ecosystem Usage & Implementations

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to verify the availability of large data blocks by checking only small, random samples. This section details its core implementations and how it secures modern blockchain architectures.

03

The Sampling Process: How Light Nodes Verify

A light node verifies data availability by:

  • Requesting random chunks (e.g., 32-byte samples) of an erasure-coded block.
  • Checking Merkle proofs to confirm each sample's authenticity against the block header.
  • Statistically guaranteeing that if a sufficient number of samples are valid, the entire data is available with near-certainty, preventing data withholding attacks.
04

Erasure Coding: Enabling Efficient Sampling

Reed-Solomon erasure coding is the prerequisite for DAS. It expands the original data block (e.g., 1 MB) into a larger encoded block (e.g., 2 MB) with redundancy. This allows light nodes to sample any unique subset of the encoded data—if even 50% is available, the original data can be fully reconstructed, making sampling highly efficient.

05

Key Security Guarantee: Data Availability Committees (DACs) vs. DAS

DAS provides a cryptoeconomic security guarantee superior to traditional Data Availability Committees (DACs). While DACs rely on a trusted multi-signature set, DAS allows a decentralized network of light nodes to cryptographically prove data is available without trust, eliminating single points of failure and enabling permissionless validation.

06

Enabling Scalable Layer 2 Rollups

DAS is the critical innovation allowing ZK-Rollups and Optimistic Rollups to scale securely. By posting transaction data to a DAS-secured layer (like Celestia or Ethereum with blobs), rollups ensure their data is publicly verifiable and downloadable for fraud proofs or state reconstruction, maintaining security while minimizing costs.

visual-explainer
DATA AVAILABILITY

Visual Explainer: The DAS Process

A step-by-step breakdown of how Data Availability Sampling (DAS) enables light clients to securely verify that block data is published without downloading it in full.

Data Availability Sampling (DAS) is a cryptographic technique that allows network participants, such as light clients, to probabilistically verify that all data for a new block has been published to the network. Instead of downloading the entire block—which can be several megabytes—a client performs multiple rounds of random sampling. In each round, it requests a small, randomly selected piece of the data, along with a Merkle proof attesting to its correctness. If the client can successfully retrieve all requested samples, it gains high statistical confidence that the complete dataset is available.

The process relies on encoding the block data using an erasure coding scheme, such as Reed-Solomon. This expands the original data with redundant parity chunks, creating a data availability (DA) layer. The key property is that the original data can be fully reconstructed from any sufficient subset of these encoded chunks. This redundancy is what makes sampling possible: even if some chunks are withheld by a malicious block producer, the remaining available chunks still contain the full information, and sampling will likely detect the missing data.

A critical component is the KZG polynomial commitment or a similar cryptographic commitment scheme. The block producer publishes a single, constant-sized commitment to the entire encoded data blob. For each random sample, the producer can generate a tiny proof that the fetched data chunk is consistent with this global commitment. This allows the light client to verify the sample's authenticity without trusting the node serving the data, ensuring the process remains secure in a decentralized, peer-to-peer network environment.

The sampling process is repeated over multiple rounds with different randomly selected indices. Security grows exponentially with the number of successful samples; a client performing 30-40 rounds can achieve confidence levels exceeding 99.9% that the data is fully available. This creates a scalability trilemma breakthrough: it decouples the verification workload from the size of the block data, enabling blockchain scalability without forcing every node to process the entire data load, a principle central to Ethereum's danksharding roadmap.

In practice, the system is designed with fault proofs in mind. If a light client does encounter a missing sample, it can generate a fraud proof to alert the network that the block producer is malicious. Furthermore, specialized full nodes or archival nodes that do download the full data can provide the missing samples or reconstruct the block, ensuring the network's liveness and data recovery capabilities are maintained even under adversarial conditions.

COMPARISON MATRIX

DAS vs. Alternative Data Availability Solutions

A technical comparison of Data Availability Sampling (DAS) with other primary methods for ensuring data is published and accessible for blockchain state verification.

Feature / MetricData Availability Sampling (DAS)Data Availability Committees (DACs)On-Chain Publication (e.g., Ethereum)

Core Mechanism

Light clients probabilistically sample small chunks of data

A trusted, permissioned committee signs attestations

Full data is published directly to consensus layer

Trust Model

Trust-minimized (cryptographic verification)

Trusted (reputational/legal trust in committee)

Trustless (secured by L1 consensus)

Scalability for Rollups

High (data scales with light client count)

Moderate (bottlenecked by committee coordination)

Low (limited by L1 block space)

Bandwidth Cost for Nodes

Low (< 100 KB per node via sampling)

Low (only committee members download full data)

High (all full nodes download 100% of data)

Data Redundancy

High (erasure-coded & distributed via network)

Moderate (depends on committee replication)

Maximum (all consensus nodes store data)

Time to Guarantee Availability

< 1 second (after sampling completes)

~2-5 seconds (committee attestation latency)

~12 seconds (L1 block confirmation)

Censorship Resistance

High (decentralized sampling network)

Low (controlled by committee)

Maximum (inherited from L1 consensus)

Primary Use Case

Modular blockchains & high-throughput L2s (e.g., Celestia)

Enterprise/consortium chains with trusted entities

Monolithic L1s & rollups using calldata (e.g., Ethereum)

security-considerations
DATA AVAILABILITY SAMPLING

Security Considerations & Guarantees

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published without downloading it entirely. Its security properties are foundational to scaling solutions like danksharding and modular blockchains.

01

The Core Security Guarantee

DAS provides a probabilistic guarantee of data availability. By randomly sampling small, fixed-size chunks (data blobs) of an erasure-coded block, a node can be statistically confident the entire data is available. The probability of missing withheld data decreases exponentially with more samples, making it computationally infeasible for an adversary to hide data.

02

Erasure Coding: The Foundational Layer

Before sampling, block data is expanded using erasure coding (e.g., Reed-Solomon). This creates redundant pieces so the original data can be reconstructed from any subset (e.g., 50% of the pieces). An adversary must therefore withhold over 50% of the data to succeed, which sampling is designed to detect. This transforms an availability problem into a detectability problem.

03

Threat Model: Data Withholding Attacks

The primary attack DAS mitigates is a data withholding attack, where a block producer creates a valid block but publishes only its header, hiding the transaction data. Without DAS, this could lead to:

  • Invalid state transitions going unchallenged.
  • Network forks if nodes accept blocks with unavailable data.
  • Censorship of specific transactions.
04

Sampling Light Clients & Security Thresholds

Light clients or validators perform sampling by requesting random chunks from the network. Security is defined by the sampling rate and network assumptions. For example, to achieve 99.9% confidence that data is available with 50% erasure coding, a node might need 30 random samples. The system assumes an honest majority of sampling nodes.

05

Implementation Risks & Assumptions

DAS security relies on several critical assumptions:

  • Honest-Majority Samplers: A sufficient number of nodes must perform sampling honestly.
  • Peer Discovery & Networking: Nodes must be able to find and connect to peers storing data chunks; sybil attacks on peer lists are a risk.
  • Erasure Coding Correctness: The initial encoding must be performed correctly, often verified by a Data Availability Committee (DAC) or a validity proof in advanced systems.
06

Relation to Fraud & Validity Proofs

DAS is often paired with fraud proofs (optimistic rollups) or validity proofs (zk-rollups).

  • With Fraud Proofs: DAS ensures the data needed to construct and verify a fraud proof is available.
  • With Validity Proofs: DAS guarantees the data referenced by the ZK-SNARK or ZK-STARK proof is accessible for reconstruction and verification. It does not verify state validity, only data publication.
CLARIFYING THE FUNDAMENTALS

Common Misconceptions About Data Availability Sampling (DAS)

Data Availability Sampling is a critical component of modern blockchain scaling, but its technical nature leads to widespread misunderstandings. This section debunks the most frequent misconceptions about how DAS works and its role in layer-2 and modular architectures.

No, Data Availability Sampling (DAS) is a specific cryptographic technique used to verify that data is available, while data availability (DA) is the broader property or guarantee that the data exists and can be retrieved. DAS is a mechanism; DA is the outcome. For example, a blockchain using erasure coding and light clients performing random sampling is employing DAS to achieve a strong DA guarantee. Other systems might achieve DA through different means, such as all full nodes storing the complete data.

DATA AVAILABILITY

Technical Deep Dive

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to verify that all transaction data for a block is published and accessible without downloading the entire dataset. This is a core scaling solution for blockchain architectures like Ethereum's danksharding and Celestia.

Data Availability Sampling (DAS) is a protocol that enables a node to probabilistically verify the full publication of a large data block by downloading and checking only a small, random subset of it. It works by having the block producer erasure-code the data, splitting it into coded chunks or data blobs. Light nodes then randomly select a handful of these chunks, request them from the network, and verify their correctness against a Merkle root commitment. If all sampled chunks are available and valid, the node gains high statistical confidence that the entire dataset is available. This process is repeated over multiple rounds by many nodes to achieve near-certain security.

Key Mechanism:

  • Erasure Coding: Data is expanded with redundancy, so the original can be reconstructed from any 50% of the chunks.
  • Random Sampling: Each node samples unique, random indices.
  • Attestation: Nodes sign attestations confirming data availability based on their samples.
DATA AVAILABILITY SAMPLING

Frequently Asked Questions (FAQ)

Data Availability Sampling (DAS) is a cryptographic technique that allows nodes to verify the availability of large datasets without downloading them in full. This is a foundational component for scaling blockchains through data availability layers and sharding.

Data Availability Sampling (DAS) is a cryptographic protocol that enables a light client or node to verify with high statistical confidence that all data for a block is published and available for download, without having to download the entire dataset. It works by having the node randomly select and download a small number of chunks from the erasure-coded block data. If all sampled chunks are successfully retrieved, the node can probabilistically conclude the entire data is available. This is crucial for scaling solutions like Ethereum's danksharding and modular data availability layers, as it allows for secure block validation without the burden of full data storage.

ENQUIRY

Get In Touch
today.

Our experts will offer a free quote and a 30min call to discuss your project.

NDA Protected
24h Response
Directly to Engineering Team
10+
Protocols Shipped
$20M+
TVL Overall
NDA Protected Directly to Engineering Team