Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
LABS
Glossary

Data Availability Sampling

Data Availability Sampling (DAS) is a cryptographic technique that allows light clients or nodes to probabilistically verify that all data for a block is published and available, without downloading the entire dataset.
Chainscore © 2026
definition
BLOCKCHAIN SCALING

What is Data Availability Sampling?

A cryptographic technique that allows light nodes to verify the availability of large data blocks without downloading them in full.

Data Availability Sampling (DAS) is a core mechanism in blockchain scaling solutions like Ethereum danksharding and Celestia that enables nodes to probabilistically confirm that all data for a block is published and accessible. Instead of downloading an entire block—which can be several megabytes or more—a light client performs multiple random checks on small, randomly selected pieces of the data. If the sampled pieces are consistently available, the client can be statistically confident the entire dataset is present, preventing a malicious block producer from hiding transaction data that could invalidate the block's state transitions.

The process relies on erasure coding the block data, typically using a Reed-Solomon code, which expands the original data with redundant parity chunks. This coding ensures that the original data can be fully reconstructed even if up to 50% of the chunks are missing. During sampling, nodes request these chunks by their coordinates within a two-dimensional matrix. A key security property is that if any part of the data is withheld, the probability of a sampling node detecting the unavailability increases exponentially with each sample, allowing for high security with a practical number of queries.

DAS is fundamental to the concept of a data availability layer, which separates the responsibility of ensuring data is published from the task of executing transactions. This separation allows for highly scalable blockchains where full nodes, called data availability committees or validators, handle the sampling and storage, while rollups or execution layers can process transactions with the guaranteed knowledge that the underlying data is secure and retrievable. It directly solves the data availability problem, a major obstacle in scaling blockchains securely.

key-features
DATA AVAILABILITY SAMPLING

Key Features

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to verify that all transaction data for a block is published and accessible without downloading the entire dataset.

01

Light Client Verification

Enables resource-constrained devices to participate in network security. Light clients perform random sampling by downloading small, random chunks of block data. If all samples are available, they can probabilistically guarantee the entire data is available, a principle known as probabilistic guarantee.

02

Erasure Coding

A prerequisite for efficient DAS. Block data is expanded using an erasure code (like Reed-Solomon), creating redundant data chunks. This allows the original data to be reconstructed even if a significant portion (e.g., 50%) of the chunks are missing, making data withholding attacks easily detectable.

03

2D KZG Commitments

A core cryptographic primitive used in implementations like Ethereum's Proto-Danksharding. A KZG polynomial commitment creates a short cryptographic proof for the entire data block. This proof allows any node to verify that a specific data chunk is part of the committed block without knowing the whole dataset.

04

Sampling & Attestation

The operational loop of DAS:

  • Random Sampling: A node requests multiple random chunks by their coordinates.
  • Availability Check: It verifies each chunk against the KZG commitment.
  • Attestation: If all checks pass, the node signs an attestation confirming data availability, which is aggregated into the consensus.
05

Security Threshold

DAS provides a cryptoeconomic security guarantee. An adversary must hide more than a certain percentage of data (e.g., >50% in a 2D scheme) to succeed. With enough nodes performing sufficient random samples, the probability of not detecting missing data drops exponentially, making attacks prohibitively expensive.

06

Enabling Scalability

DAS is the foundation for data availability layers and modular blockchains. It decouples data availability from execution verification, allowing block sizes to increase massively (e.g., 16 MB per slot in Ethereum danksharding) while keeping verification lightweight for all nodes.

how-it-works
BLOCKCHAIN SCALABILITY

How Does Data Availability Sampling Work?

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a new block is published and accessible, without downloading the entire dataset.

Data Availability Sampling (DAS) is a core mechanism in blockchain scaling architectures like Ethereum danksharding and Celestia. It solves the data availability problem, where a block producer might withhold transaction data after publishing a block header, making it impossible for others to verify or reconstruct the chain's state. DAS enables light clients or validators to check for data availability with high confidence by downloading only a small, random subset of the block's data.

The process relies on encoding the block data using an erasure coding scheme like Reed-Solomon. This transforms the original data into extended data chunks, where only a fraction (e.g., 50%) is needed to reconstruct the whole. A 2D Reed-Solomon matrix is often used, creating rows and columns of coded chunks. Nodes then perform multiple rounds of random sampling, requesting a handful of these chunks via a peer-to-peer network. If all requested samples are successfully retrieved, the node can be statistically confident the entire data is available.

The security model is probabilistic. Each successful sample increases confidence exponentially. For instance, after 30 successful random samples, a node might have 99.9% certainty that the data is available. If a sample request fails, it serves as proof that the block producer is malicious, and the block can be rejected. This creates a fisherman's game, where any honest node can catch and prove fraud by broadcasting a single missing data chunk.

DAS is fundamental to modular blockchain designs, separating execution from consensus and data availability. It allows rollups and other execution layers to post massive amounts of data cheaply to a base layer, knowing that light nodes can securely verify its availability. This enables scalable blocks without requiring every participant to process terabytes of data, paving the way for secure, high-throughput blockchain networks.

ecosystem-usage
DATA AVAILABILITY SAMPLING

Ecosystem Usage

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to verify that block data is available without downloading it entirely. This section details its practical implementations and the projects building on this core scaling primitive.

06

Light Client Verification & Bridges

DAS is the enabling technology for trust-minimized light clients and cross-chain bridges. A light client can perform a few random samples to be confident a block's data is available, enabling secure header synchronization without running a full node. This is critical for sovereign chains and bridge security models that rely on light client proofs.

  • Core Function: Replaces the need for data availability committees (DACs) with cryptographic guarantees.
  • Impact: Forms the foundation for a more secure and decentralized multi-chain ecosystem.
visual-explainer
VISUAL EXPLAINER

Data Availability Sampling

A visual guide to the cryptographic technique that allows light nodes to securely verify that all transaction data for a block is published, without downloading the entire dataset.

Data Availability Sampling (DAS) is a cryptographic protocol that enables nodes in a blockchain network to probabilistically verify that all data for a new block is published and accessible, without needing to download the entire block. This is achieved by having light clients or validators request small, random pieces of the erasure-coded data. If the sampled pieces are consistently available, it provides high statistical confidence that the entire dataset is available, preventing data withholding attacks where a malicious block producer might hide transaction data.

The process relies on erasure coding, where the original block data is expanded into a larger set of coded pieces with redundancy. A key property is that any sufficient subset of these pieces can reconstruct the original data. During sampling, a node requests a fixed number of random pieces (e.g., 30 samples). If the block producer is withholding data, the probability of the node successfully retrieving all requested samples decreases exponentially with each attempt, making deception statistically improbable.

DAS is a foundational component for scalable blockchain architectures like danksharding and modular rollups, as it decouples data verification from full data download. It allows the network to securely increase block sizes while keeping verification lightweight. Implementations, such as those in Ethereum's proto-danksharding (EIP-4844), use a 2D KZG polynomial commitment scheme to generate and commit to these data blobs, enabling efficient sampling and fraud proofs.

From a node's perspective, the sampling workflow is continuous: 1) Receive a block header with a commitment to the data. 2) Generate random coordinates (row and column indices) for the data matrix. 3) Request the data chunks at those coordinates from the network. 4) Verify the received chunks against the commitment. If a sample is unavailable, the node can initiate a data availability challenge, signaling to the network that the block may be invalid and should not be finalized.

The security model of DAS creates a clear trade-off: a node's confidence in data availability increases with the number of successful samples. For example, after 30 successful random samples, the probability of missing withheld data can be made astronomically low (e.g., less than 1 in a trillion). This allows resource-constrained devices to participate in consensus security, enabling truly trust-minimized light clients and paving the way for higher-throughput blockchains without sacrificing decentralization.

DATA VERIFICATION METHODS

DAS vs. Full Node Download

A comparison of the resource requirements and trust assumptions between Data Availability Sampling (DAS) and traditional full block data download.

Feature / MetricData Availability Sampling (DAS)Full Node Download

Core Function

Randomly samples small chunks to probabilistically verify data availability

Downloads and stores the entire block data for deterministic verification

Node Resource Requirement

Light client (low storage, low bandwidth)

Full node (high storage, high bandwidth)

Trust Model

Trustless; security scales with network size

Self-verified; requires no external trust

Minimum Data Downloaded

< 0.1% of block data (for high confidence)

100% of block data

Storage Overhead

Minimal (only sample proofs)

Full blockchain history (e.g., 1TB+ for Ethereum)

Verification Speed

~2-10 seconds (parallel sampling)

Limited by full block download speed

Scalability for Rollups

Enables scalable data availability layers (e.g., Celestia, EigenDA)

Limited by base layer block size and node capacity

security-considerations
DATA AVAILABILITY SAMPLING

Security Considerations

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to probabilistically verify that all data for a block is published without downloading it entirely. Its security model introduces unique trade-offs and attack vectors.

01

The 51% Attack Vector

DAS is designed to be secure against adversaries controlling less than 50% of the sampling nodes. However, a malicious block producer with over 50% of the network's sampling power can selectively hide data and pass random checks, creating a data withholding attack. This is mitigated by requiring a high number of independent samples (e.g., 30 samples for 99.9% confidence).

02

Sampling Node Sybil Resistance

The security of DAS depends on the assumption that sampling nodes are independent and numerous. An attacker could create many Sybil nodes to increase the probability of a malicious block not being caught. Defenses include:

  • Resource-based sampling (e.g., requiring a proof-of-work puzzle per sample).
  • Staking-based committees where samplers have economic skin in the game.
03

Data Availability Committee (DAC) Fallback

Many rollups using DAS employ a Data Availability Committee (DAC) as a fallback. This introduces a trust assumption—if the DAS layer fails, users must trust the DAC's signatures. The security of the system then reverts to the honesty of this multisig group, creating a potential centralization point and a different risk profile than pure cryptographic DAS.

04

Latency & Unavailability Proofs

A critical security mechanism is the generation of fraud proofs or data unavailability proofs. If a light node detects missing data, it must be able to generate a cryptographic proof that convinces the full network to reject the block. The speed and efficiency of this process is crucial; slow proofs can extend the window for chain reorganization or double-spend attacks.

05

Erasure Coding & Data Redundancy

DAS relies on erasure coding (e.g., Reed-Solomon) to expand the original data. The security property is that any 50% of the coded data can reconstruct the whole. An attacker must therefore hide more than 50% of the coded chunks to succeed. However, faulty implementation of the coding scheme or its polynomial commitments (like KZG) can create vulnerabilities where less data needs to be hidden.

06

Network-Level Attacks

Attackers can target the peer-to-peer network layer to degrade DAS. This includes:

  • Eclipse attacks to isolate sampling nodes from honest peers.
  • Bandwidth exhaustion attacks against nodes serving data samples.
  • Timing attacks to delay sample responses beyond a timeout window. Robust gossip protocols and peer diversity are essential countermeasures.
DATA AVAILABILITY SAMPLING

Common Misconceptions

Data Availability Sampling (DAS) is a critical cryptographic technique for scaling blockchains, but its mechanics are often misunderstood. This section clarifies the most frequent points of confusion about how DAS works, its guarantees, and its practical implications for rollups and validators.

No, Data Availability Sampling is a probabilistic verification method where nodes download only small, random chunks of block data to statistically confirm its availability, rather than downloading the entire block. This is the core innovation that enables light clients and validators to securely confirm that data for a block exists without the prohibitive bandwidth cost of full downloads. The security model relies on a large number of independent samples; if the data is withheld, a sample will eventually hit a missing piece with near-certain probability. This is fundamentally different from data availability committees or validators attesting to data they have fully seen.

DATA AVAILABILITY SAMPLING

Frequently Asked Questions

Data Availability Sampling (DAS) is a cryptographic technique that allows light nodes to verify that all transaction data for a block is published and accessible without downloading the entire dataset. This FAQ addresses its core mechanics, purpose, and role in modern blockchain scaling.

Data Availability Sampling (DAS) is a protocol that enables nodes to probabilistically verify the availability of all data in a large block by downloading and checking only a small, random subset of it. It works by having the block producer erasure-code the data, splitting it into many small chunks. Light nodes then randomly sample a handful of these chunks. If all sampled chunks are retrievable, the node gains high statistical confidence that the entire dataset is available. This process prevents block producers from successfully hiding any part of the transaction data, a scenario known as a data availability problem.

ENQUIRY

Get In Touch
today.

Our experts will offer a free quote and a 30min call to discuss your project.

NDA Protected
24h Response
Directly to Engineering Team
10+
Protocols Shipped
$20M+
TVL Overall
NDA Protected Directly to Engineering Team