Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
Free 30-min Web3 Consultation
Book Consultation
Smart Contract Security Audits
View Audit Services
Custom DeFi Protocol Development
Explore DeFi
Full-Stack Web3 dApp Development
View App Services
LABS
Glossary

Data Availability Proof (DAP)

A Data Availability Proof (DAP) is a cryptographic attestation that a specific dataset has been published and is accessible for download, without requiring the verifier to retrieve the entire data.
Chainscore © 2026
definition
BLOCKCHAIN SCALING

What is a Data Availability Proof (DAP)?

A cryptographic proof that guarantees a block's data is published and accessible, enabling secure scaling solutions like rollups.

A Data Availability Proof (DAP) is a cryptographic mechanism that allows a network node to verify, with high probability, that all data for a given block has been published and is retrievable by the network, without the node needing to download the entire dataset. This is a foundational component for layer-2 scaling solutions, particularly rollups, which rely on posting transaction data to a base layer like Ethereum. The core problem it solves is the Data Availability Problem: ensuring that a block producer (or sequencer) cannot withhold data, which could prevent others from verifying state transitions or detecting fraud. Without DAPs, a malicious actor could propose a block with invalid transactions and hide the data, making fraud proofs impossible.

The most common technical approach is Data Availability Sampling (DAS), where light clients perform multiple rounds of random sampling on small, erasure-coded pieces of the block data. If a sufficient number of samples are successfully retrieved, the client can be statistically confident the entire dataset is available. This is the principle behind data availability layers like Celestia and EigenDA, and is being integrated into Ethereum's roadmap via Proto-Danksharding (EIP-4844) and full Danksharding. These systems use advanced erasure coding, such as Reed-Solomon codes, to expand the data, ensuring it can be reconstructed even if some pieces are missing, which makes sampling an efficient and secure proxy for full data download.

Data Availability Proofs are critical for the security model of optimistic rollups and zk-rollups. For an optimistic rollup, verifiers must have access to all transaction data to compute the state and submit fraud proofs during the challenge period. For a zk-rollup, while validity is proven by a zero-knowledge proof (ZKP), the data is still needed for users to reconstruct their state and for future provers to generate proofs. DAPs thus decouple consensus and execution from data publishing, creating a modular blockchain stack where specialized layers handle each function. This enables higher throughput while maintaining the security assumption that the underlying data layer is honest and available.

key-features
DATA AVAILABILITY PROOF

Key Features

Data Availability Proofs (DAPs) are cryptographic protocols that allow a node to verify that all data for a block is published and accessible without downloading the entire dataset. This is a foundational security mechanism for scaling solutions like rollups.

04

Validity vs. Fraud Proofs

DAPs interact with two major rollup security models:

  • Validity Proofs (ZK-Rollups): Require 100% data availability for the state transition proof. DAPs ensure the proof's input data is published.
  • Fraud Proofs (Optimistic Rollups): Require data availability for the transaction data so verifiers can reconstruct state and challenge invalid transitions. Without DAPs, fraud proofs are impossible.
05

Data Availability Committees (DACs)

A simpler, more centralized alternative to cryptographic DAPs. A known set of entities sign attestations that they have received and stored a rollup's data. While less trust-minimized than pure cryptographic proofs, DACs offer practical data availability with lower complexity. Used by early Layer 2 solutions before advanced DAPs were production-ready.

06

Economic Security & Slashing

DAP systems often have an economic security layer. Block producers (e.g., sequencers or validators) post a bond. If they produce a block but withhold its data, they can be slashed—their bond is destroyed. This creates a strong financial disincentive against data withholding attacks, complementing the cryptographic guarantees.

how-it-works
MECHANISM

How Data Availability Proofs Work

A technical breakdown of the cryptographic and game-theoretic mechanisms that allow a blockchain network to verify data is available without downloading it.

A Data Availability Proof (DAP) is a cryptographic protocol that allows a network node to verify with high probability that all the data for a block is published and accessible, without needing to download the entire dataset. This is a critical component for scaling solutions like rollups and sharding, where only a small committee of validators may download full data, while light clients or other layers need assurance the data exists. The core challenge is preventing a malicious block producer from withholding even a small portion of data, which could conceal invalid transactions.

The most common technique is Data Availability Sampling (DAS). Here, light clients randomly request small, random pieces (samples) of the block's erasure-coded data. Erasure coding (e.g., Reed-Solomon) redundantly expands the original data, so that any 50% of the coded pieces can reconstruct the whole. If a malicious actor withholds even 1% of the data, the probability of a sampling client detecting the unavailability increases exponentially with each sample request. After a sufficient number of successful random samples, a client can be statistically confident the entire data is available.

This sampling process is secured by a cryptographic commitment, typically a Merkle root or a KZG polynomial commitment, to the full erasure-coded data. When a client requests a sample, the network provides the data chunk along with a Merkle proof linking it to the published commitment. This proves the sample is part of the genuine, committed dataset. Systems like Celestia and Ethereum's Proto-Danksharding (EIP-4844) implement this architecture, enabling secure scaling by separating data availability consensus from execution.

An alternative, game-theoretic approach is the Fraud Proof-based model used in some optimistic rollups. Here, a challenge period allows any honest node that has the full data to prove a block proposer acted maliciously by publishing a fraud proof. The security assumption is that at least one honest, fully-synced node exists. This model relies less on advanced cryptography but introduces a delay for finality and has different trust assumptions compared to cryptographic DAS.

Ultimately, Data Availability Proofs shift the security model from "every node stores everything" to "enough nodes store enough pieces." They enable blockchain scalability trilemma solutions by allowing layers to specialize: one layer ensures data is available and ordered (consensus and data availability layer), while another handles execution (execution layer). This modular separation, powered by DAPs, is foundational to modern modular blockchain architectures.

types-of-proofs
MECHANISMS

Types of Data Availability Proofs

Data Availability Proofs (DAPs) are cryptographic protocols that allow a node to verify that transaction data is published and accessible without downloading it entirely. Different mechanisms offer trade-offs in security, scalability, and trust assumptions.

01

Data Availability Sampling (DAS)

A probabilistic verification method where light clients randomly sample small, random chunks of the erasure-coded data block. By successfully querying for a sufficient number of unique samples, a client can achieve statistical certainty (e.g., 99.9%) that the entire data is available. This is the core scaling mechanism for Ethereum danksharding and Celestia.

  • Key Property: Enables trust-minimized scaling for light nodes.
  • Process: Client requests random indices; network must respond with correct data and Merkle proofs.
02

Data Availability Committees (DACs)

A trust-based model where a known, permissioned set of entities (the committee) cryptographically attests, typically via signatures, that they have stored a copy of the data and will make it available. Rollups like early versions of Arbitrum Nova and Polygon Avail in its initial phase used this model.

  • Key Property: Higher performance but introduces a trust assumption in the committee members.
  • Security: Relies on the honesty of a majority of committee members.
03

Validity Proofs with DA

In this architecture, the proof system itself guarantees data availability. A zk-Rollup's validity proof (ZK-SNARK/STARK) is published on-chain, but the proof's correctness is contingent on the availability of the underlying data for a fraud challenge period. Systems like zkSync Era and StarkNet use this model, often relying on a fallback DAC or the underlying L1 for data publication.

  • Key Property: Data availability is a prerequisite for the security of the cryptographic proof.
  • On-Chain Data: Typically publishes state diffs or calldata to Ethereum.
04

Fraud Proofs with DA

Used by Optimistic Rollups like Optimism and Arbitrum One. Transaction data is posted to a data availability layer (usually Ethereum as calldata). The system assumes data is available and state transitions are valid unless challenged. A verifier can submit a fraud proof during the challenge window, which must reference the available data to prove invalidity.

  • Key Property: Security window (e.g., 7 days) where data must remain available for fraud challenges.
  • Core Dependency: Absolute data availability is critical for the security model.
05

Erasure Coding (Reed-Solomon)

A foundational encoding technique used to make data availability sampling possible. A block's data is expanded using Reed-Solomon codes, creating redundant parity chunks. The property is that the original data can be reconstructed from any 50% (or other threshold) of the total chunks. This allows samplers to detect unavailability with high probability even if many chunks are missing.

  • Key Property: Provides redundancy and enables sampling.
  • Requirement: The encoding must be performed correctly, often verified by KZG polynomial commitments or fraud proofs.
06

Peer-to-Peer (P2P) Network Propagation

The underlying gossip network protocol that physically distributes data. While not a proof mechanism itself, it is the infrastructure upon which all DAPs rely. Nodes broadcast and relay data blocks and samples. The resilience and latency of this P2P mesh network directly impact the liveness and security guarantees of the data availability layer.

  • Key Property: The practical implementation layer for data distribution.
  • Metrics: Measured by block propagation time and network node diversity.
DATA AVAILABILITY ARCHITECTURE

DAPs in Scaling Solutions: Rollups vs. Validiums

Comparison of how two major Layer 2 scaling solutions handle data availability, a core component of their security models.

FeatureOptimistic RollupsZK-RollupsValidiums

Data Availability Layer

Ethereum L1

Ethereum L1

Off-Chain (Data Availability Committee or DAC)

Data Publication

Full transaction data posted to L1

State diffs and validity proofs posted to L1

Only validity proofs posted to L1; data held off-chain

Withdrawal Security (Censorship Resistance)

High (users can force tx inclusion via L1)

High (users can force tx inclusion via L1)

Reduced (relies on off-chain data providers)

Throughput (Max TPS)

~100-1,000

~2,000-10,000+

~10,000-20,000+

Transaction Finality Time

~1 week (challenge period)

< 10 minutes

< 10 minutes

Primary Security Model

Fraud proofs (challenge period)

Validity proofs (ZK-SNARKs/STARKs)

Validity proofs (ZK-SNARKs/STARKs)

L1 Data Cost per Batch

High

Medium

Very Low

Trust Assumptions

1-of-N honest validator

Cryptographic (trustless)

Cryptographic + Honest Data Availability Committee

ecosystem-usage
DATA AVAILABILITY PROOF (DAP)

Ecosystem Usage

Data Availability Proofs (DAPs) are cryptographic protocols that allow a verifier to confirm that the complete data for a block is published and accessible, without downloading it all. This is a foundational component for scaling solutions like rollups and sharding.

02

Data Availability Sampling (DAS)

A core technique where light clients or validators randomly sample small chunks of a block's data. By successfully sampling enough chunks, they can achieve statistical certainty that the entire dataset is available. This is the mechanism used by Ethereum's Proto-Danksharding (EIP-4844) and Celestia to scale data availability without requiring full nodes.

03

Erasure Coding & Merkle Proofs

To make sampling possible, block data is encoded using erasure codes (like Reed-Solomon), which expands the data with redundancy. The encoded data is then committed to using a Merkle root. Samplers request random pieces along with a Merkle proof to verify the piece is part of the original committed data.

04

Key Implementations

  • Ethereum (EIP-4844): Introduces blob-carrying transactions with a separate fee market for temporary data availability.
  • Celestia: A modular blockchain network specialized as a data availability layer.
  • Avail: A data availability layer built with Substrate, using validity proofs and KZG polynomial commitments.
  • EigenDA: A restaking-based data availability service built on Ethereum.
05

The Data Availability Problem

This is the challenge of ensuring that all data for a new block is actually published to the network. A malicious block producer could withhold data, making it impossible for honest validators to verify the block's correctness, leading to potential double-spends or invalid state transitions. DAPs solve this problem.

security-considerations
DATA AVAILABILITY PROOF

Security Considerations

Data Availability Proofs (DAPs) are cryptographic protocols that allow a node to verify that all data for a block is published and accessible without downloading it entirely. Their security is paramount for the integrity of scaling solutions like rollups.

01

Data Withholding Attacks

The primary security threat DAPs mitigate is a data withholding attack, where a block producer publishes a block header but withholds some transaction data. This prevents full nodes from reconstructing the state and verifying the block's validity. DAPs, like Erasure Coding and KZG Commitments, ensure that any missing data can be detected with high probability by sampling small, random chunks.

02

Sampling & Probabilistic Security

Light clients use Data Availability Sampling (DAS) to achieve security without downloading full blocks. By randomly sampling a small number of chunks (e.g., 30-50), a client can statistically guarantee (e.g., 99.9% confidence) that all data is available. The security scales with the number of samples; an adversary would need to control a massive portion of the network to hide data successfully.

03

Erasure Coding Redundancy

To enable sampling, block data is expanded using Erasure Coding (e.g., Reed-Solomon). This creates redundant data chunks so that the original data can be recovered even if up to 50% of the chunks are missing. The security guarantee depends on the soundness of this encoding: if it is implemented incorrectly, data could be unrecoverable even when the sampling check passes.

04

Commitment Schemes (KZG vs. Merkle)

A cryptographic commitment (like a KZG polynomial commitment or a Merkle root) binds the prover to the data. KZG commitments allow for efficient proofs that a specific data chunk is part of the encoded block. Their security relies on trusted setups or cryptographic assumptions (DL/AGM). Merkle proofs are simpler but require more data per sample.

05

Validator Incentives & Slashing

In Data Availability Committees (DACs) or consensus-based DA layers, validators must attest to data availability. Slashing conditions punish validators who sign for a block where data is later proven unavailable. Proper economic incentives are critical to ensure it is more costly to attack the system than to act honestly.

06

Layer-1 Fallback & Force Inclusion

Rollups using external DA must have a secure fallback mechanism to the Layer 1. This is often a force inclusion protocol, where a user can submit a transaction directly to the L1 contract if data is withheld, ensuring liveness. The security of the rollup depends on the timeliness and cost-effectiveness of this fallback.

visual-explainer
MECHANISM

Visual Explainer: The DAP Verification Flow

A step-by-step breakdown of how a Data Availability Proof (DAP) is generated and cryptographically verified to ensure data is retrievable.

A Data Availability Proof (DAP) is a cryptographic commitment that guarantees a block's data is published and accessible to the network, enabling light clients and other chains to trust data without downloading it entirely. The verification flow begins when a block producer creates a new block and encodes its data using an erasure coding scheme like Reed-Solomon. This process expands the original data into data shares, creating redundancy so the full data can be reconstructed even if some shares are missing. The producer then computes a Merkle root of these shares, which serves as the core data availability commitment included in the block header.

To enable efficient verification, the producer generates Data Availability Sampling (DAS) proofs. Light clients or validators on another chain randomly select and request a small number of these coded data shares. By successfully retrieving these random samples, they gain statistical confidence that the entire dataset is available. This sampling process is the heart of DAPs, as it allows resource-constrained participants to verify availability with minimal bandwidth and computation, a critical scaling solution for modular blockchain architectures like Ethereum's danksharding or Celestia.

The final verification step involves checking the cryptographic consistency of the samples against the published commitment. Each retrieved share must have a valid Merkle proof linking it back to the root in the block header. If a sufficient number of random samples are retrieved and validated, the verifier accepts the DAP, confirming data availability. If samples cannot be retrieved, it signals that the block producer is withholding data, preventing the network from accepting an invalid or malicious block. This flow decouples execution verification from data verification, forming a trust-minimized bridge for rollups and cross-chain communication.

DATA AVAILABILITY PROOF

Frequently Asked Questions (FAQ)

Essential questions and answers about Data Availability Proofs (DAPs), a critical component for scaling blockchains and ensuring data integrity.

A Data Availability Proof (DAP) is a cryptographic mechanism that allows a node to efficiently verify that all data for a block is published and accessible on the network, without needing to download the entire dataset. It works by having block producers commit to the data using a Merkle root and then generating erasure-coded fragments of the data. Light clients or other validators can then sample a small number of these random fragments; if they can successfully retrieve them, they have high statistical certainty the full data is available. This is foundational for scaling solutions like rollups, ensuring that anyone can reconstruct the state and challenge invalid transactions if needed.

ENQUIRY

Get In Touch
today.

Our experts will offer a free quote and a 30min call to discuss your project.

NDA Protected
24h Response
Directly to Engineering Team
10+
Protocols Shipped
$20M+
TVL Overall
NDA Protected Directly to Engineering Team