Original Author: @BlazingKevin_, the Researcher at Movemaker
Storage was once one of the top narratives in the industry. Filecoin, as the leading project in the last bull market, had a market cap exceeding $10 billion. Arweave, as a comparable storage protocol, promoted permanent storage as its selling point, reaching a market cap of $3.5 billion at its peak. However, as the availability of cold data storage has been debunked, the necessity of permanent storage has come into question, leaving a big question mark on whether the narrative of decentralized storage can succeed. The emergence of Walrus has stirred the long-dormant storage narrative, and now Aptos, in collaboration with Jump Crypto, has launched Shelby, aiming to elevate decentralized storage in the hot data sector. So, can decentralized storage make a comeback and provide widespread use cases? Or is it just another topic hype? This article analyzes the narrative evolution of decentralized storage through the development paths of Filecoin, Arweave, Walrus, and Shelby, attempting to find an answer: how far is the road to the popularization of decentralized storage?
Filecoin: Storage is the Surface, Mining is the Essence
Filecoin is one of the initially emerging altcoins, and its development direction naturally revolves around decentralization, which is a common trait of early altcoins—seeking the significance of decentralization in various traditional sectors. Filecoin is no exception; it connects storage with decentralization, naturally leading to the drawbacks of centralized storage: the trust assumption of centralized data storage service providers. Therefore, what Filecoin does is shift centralized storage to decentralized storage. However, certain aspects sacrificed in the process to achieve decentralization have become pain points that later projects like Arweave or Walrus aim to address. To understand why Filecoin is merely a mining coin, one must grasp the objective limitations of its underlying technology, IPFS, which is not suitable for hot data.
IPFS: Decentralized Architecture, Yet Stalled by Transmission Bottlenecks
IPFS (InterPlanetary File System) was launched around 2015, aiming to disrupt the traditional HTTP protocol through content addressing. The biggest drawback of IPFS is its extremely slow retrieval speed. In an era where traditional data service providers can achieve millisecond-level responses, retrieving a file via IPFS still takes several seconds, making it difficult to promote in practical applications and explaining why it is rarely adopted by traditional industries, aside from a few blockchain projects.
The underlying P2P protocol of IPFS is mainly suitable for "cold data," which refers to static content that does not change frequently, such as videos, images, and documents. However, when it comes to handling hot data, such as dynamic web pages, online games, or AI applications, the P2P protocol does not have a clear advantage over traditional CDNs.
However, despite IPFS itself not being a blockchain, its directed acyclic graph (DAG) design concept aligns closely with many public chains and Web3 protocols, making it inherently suitable as a foundational framework for blockchain. Therefore, even if it lacks practical value, it is sufficient as a foundational framework for carrying blockchain narratives. Early altcoin projects only needed a workable framework to embark on their journeys, but as Filecoin developed over time, the inherent flaws brought by IPFS began to hinder its progress.
The Mining Coin Logic Beneath the Storage Facade
The original design of IPFS was to allow users to store data while also being part of the storage network. However, without economic incentives, it is challenging for users to voluntarily use this system, let alone become active storage nodes. This means that most users will only store files on IPFS but will not contribute their storage space or store others' files. It is in this context that Filecoin emerged.
Filecoin's token economic model primarily involves three roles: users are responsible for paying fees to store data; storage miners earn token incentives for storing user data; and retrieval miners provide data when users need it and receive incentives.
This model has potential for malicious behavior. Storage miners may fill their allocated storage space with junk data to earn rewards. Since this junk data is not retrieved, even if it is lost, it does not trigger the penalty mechanism for storage miners. This allows storage miners to delete junk data and repeat the process. Filecoin's proof-of-replication consensus can only ensure that user data has not been deleted without permission, but it cannot prevent miners from filling their storage with junk data.
The operation of Filecoin largely relies on miners' continuous investment in the token economy rather than on end users' genuine demand for distributed storage. Although the project continues to iterate, at this stage, the ecological construction of Filecoin aligns more with the "mining coin logic" rather than the "application-driven" definition of storage projects.
Arweave: Success in Long-Termism, Failure in Long-Termism
If Filecoin's design goal is to build an incentivized, verifiable decentralized "data cloud" shell, then Arweave takes a different extreme in storage: providing the capability for permanent data storage. Arweave does not attempt to build a distributed computing platform; its entire system revolves around a core assumption—that important data should be stored once and remain on the network forever. This extreme long-termism makes Arweave fundamentally different from Filecoin in terms of mechanisms, incentive models, hardware requirements, and narrative perspectives.
Arweave takes Bitcoin as a learning object, attempting to continuously optimize its permanent storage network over long cycles measured in years. Arweave does not care about marketing, nor does it concern itself with competitors and market trends. It simply continues to iterate on its network architecture, indifferent to whether anyone pays attention, because this is the essence of the Arweave development team: long-termism. Thanks to long-termism, Arweave was fervently sought after in the last bull market; and because of long-termism, even after hitting rock bottom, Arweave may still survive several rounds of bull and bear markets. The question remains: will decentralized storage have a place for Arweave in the future? The existence value of permanent storage can only be proven over time.
Since the mainnet of Arweave transitioned from version 1.5 to the recent version 2.9, despite losing market discussion, it has been committed to enabling a broader range of miners to participate in the network at minimal cost and incentivizing miners to maximize data storage, thereby continuously enhancing the robustness of the entire network. Arweave's conservative approach, fully aware that it does not align with market preferences, does not embrace the miner community, and its ecosystem has completely stagnated, upgrading the mainnet at minimal cost while continuously lowering hardware thresholds without compromising network security.
Review of the Upgrade Path from 1.5 to 2.9
Version 1.5 of Arweave exposed a vulnerability where miners could rely on GPU stacking rather than actual storage to optimize block generation probabilities. To curb this trend, version 1.7 introduced the RandomX algorithm, limiting the use of specialized computing power and requiring general CPUs to participate in mining, thereby weakening the centralization of computing power.
In version 2.0, Arweave adopted SPoA, transforming data proofs into a concise path of Merkle tree structure and introducing format 2 transactions to reduce synchronization burdens. This architecture alleviated network bandwidth pressure, significantly enhancing node collaboration capabilities. However, some miners could still evade the responsibility of holding real data through centralized high-speed storage pool strategies.
To correct this bias, version 2.4 introduced the SPoRA mechanism, incorporating global indexing and slow hash random access, requiring miners to genuinely hold data blocks to participate in effective block generation, thereby mechanically weakening the effects of computing power stacking. The result was that miners began to focus on storage access speeds, driving the application of SSDs and high-speed read-write devices. Version 2.6 introduced hash chains to control block generation rhythm, balancing the marginal benefits of high-performance devices and providing fair participation space for small and medium miners.
Subsequent versions further strengthened network collaboration capabilities and storage diversity: version 2.7 added collaborative mining and mining pool mechanisms to enhance the competitiveness of small miners; version 2.8 introduced a composite packaging mechanism, allowing large-capacity low-speed devices to participate flexibly; and version 2.9 introduced a new packaging process with the replica29 format, significantly improving efficiency and reducing computational dependencies, completing the closed loop of data-oriented mining models.
Overall, Arweave's upgrade path clearly presents its long-term strategy oriented towards storage: continuously resisting the trend of computing power centralization while lowering participation thresholds to ensure the long-term operational viability of the protocol.
Walrus: Embracing Hot Data—Hype or Hidden Depths?
Walrus, in terms of design philosophy, is completely different from Filecoin and Arweave. Filecoin's starting point is to create a decentralized verifiable storage system at the cost of cold data storage; Arweave's starting point is to create a chain-based library of Alexandria capable of permanent data storage, at the cost of too few scenarios; Walrus's starting point is to optimize storage costs for hot data storage protocols.
Magic Modification of Erasure Codes: Cost Innovation or Old Wine in New Bottles?
In terms of storage cost design, Walrus believes that the storage costs of Filecoin and Arweave are unreasonable, as both adopt a fully replicated architecture, whose main advantage lies in each node holding a complete copy, providing strong fault tolerance and independence between nodes. This type of architecture ensures that even if some nodes go offline, the network still maintains data availability. However, this also means that the system requires multiple copies for redundancy to maintain robustness, thereby increasing storage costs. Especially in Arweave's design, the consensus mechanism itself encourages node redundant storage to enhance data security. In contrast, Filecoin is more flexible in cost control, but at the expense of potentially higher data loss risks for some low-cost storage. Walrus attempts to find a balance between the two, controlling replication costs while enhancing availability through structured redundancy, thereby establishing a new compromise path between data availability and cost efficiency.
Walrus's self-created Redstuff is the key technology for reducing node redundancy, originating from Reed-Solomon (RS) coding. RS coding is a very traditional erasure code algorithm, which allows for the doubling of a data set by adding redundant fragments (erasure code) that can be used to reconstruct the original data. From CD-ROMs to satellite communications to QR codes, it is frequently used in daily life.
Erasure codes allow users to obtain a block, for example, 1MB in size, and then "amplify" it to 2MB, where the additional 1MB is special data called erasure code. If any byte in the block is lost, users can easily recover those bytes through the code. Even if up to 1MB of the block is lost, the entire block can still be recovered. The same technology allows computers to read all data on a CD-ROM, even if it has been damaged.
Currently, the most commonly used is RS coding. The implementation method starts with k information blocks, constructs a related polynomial, and evaluates it at different x coordinates to obtain the encoded blocks. Using RS erasure codes, the probability of randomly sampling large blocks of lost data is very low.
For example: A file is divided into 6 data blocks and 4 parity blocks, totaling 10 pieces. As long as any 6 pieces are retained, the original data can be completely restored.
Advantages: Strong fault tolerance, widely used in CD/DVD, fault-tolerant RAID arrays, and cloud storage systems (such as Azure Storage, Facebook F4).
Disadvantages: Decoding calculations are complex and costly; not suitable for frequently changing data scenarios. Therefore, it is usually used for data recovery and scheduling in off-chain centralized environments.
In a decentralized architecture, Storj and Sia have adjusted traditional RS coding to meet the actual needs of distributed networks. Walrus has also proposed its own variant based on this—RedStuff coding algorithm—to achieve lower costs and more flexible redundancy storage mechanisms.
What is the biggest feature of RedStuff? By improving the erasure coding algorithm, Walrus can quickly and robustly encode unstructured data blocks into smaller shards, which are distributed across a network of storage nodes. Even if up to two-thirds of the shards are lost, the original data block can be quickly reconstructed using partial shards. This is made possible while maintaining a replication factor of only 4 to 5 times.
Therefore, it is reasonable to define Walrus as a lightweight redundancy and recovery protocol redesigned around decentralized scenarios. Compared to traditional erasure codes (like Reed-Solomon), RedStuff no longer pursues strict mathematical consistency but instead makes realistic trade-offs regarding data distribution, storage verification, and computational costs. This model abandons the immediate decoding mechanism required for centralized scheduling, opting instead to verify whether nodes hold specific data copies through on-chain Proof, thus adapting to a more dynamic and marginalized network structure.
The core design of RedStuff is to split data into two categories: primary shards and secondary shards. Primary shards are used to recover the original data, and their generation and distribution are strictly constrained, with a recovery threshold of f+1 and requiring 2f+1 signatures as availability endorsement; secondary shards are generated through simple operations like XOR combinations, serving to provide elastic fault tolerance and enhance the overall robustness of the system. This structure essentially lowers the requirements for data consistency—allowing different nodes to temporarily store different versions of data, emphasizing a practical path of "eventual consistency." Although similar to the lenient requirements for backtracking blocks in systems like Arweave, it has achieved some effect in reducing network burden, but it also weakens the guarantees of data immediacy and integrity.
It is important to note that while RedStuff achieves effective storage in low computational power and low bandwidth environments, it essentially remains a "variant" of an erasure code system. It sacrifices some data reading determinism in exchange for cost control and scalability in decentralized environments. However, whether this architecture can support large-scale, high-frequency interactive data scenarios remains to be seen. Furthermore, RedStuff has not truly broken through the long-standing computational bottleneck of erasure codes but has instead avoided the high coupling points of traditional architectures through structural strategies, with its innovation being more reflected in engineering-side combinatorial optimization rather than a fundamental algorithmic disruption.
Thus, RedStuff is more like a "reasonable modification" made for the current decentralized storage reality. It indeed brings improvements in redundancy costs and operational load, allowing edge devices and non-high-performance nodes to participate in data storage tasks. However, in large-scale applications, general computational adaptability, and business scenarios with higher consistency requirements, its capability boundaries remain quite evident. This makes Walrus's innovation more of an adaptive transformation of the existing technological system rather than a decisive breakthrough in promoting the paradigm shift of decentralized storage.
Sui and Walrus: Can High-Performance Public Chains Drive Storage Practicality?
From Walrus's official research article, we can see its target scenario: "The original design intention of Walrus is to provide solutions for storing large binary files (Blobs), which are the lifeblood of many decentralized applications."
Large blob data typically refers to large, structurally unfixed binary objects, such as videos, audio, images, model files, or software packages.
In the context of cryptocurrency, it more often refers to NFTs, images, and videos in social media content. This also constitutes the main application direction of Walrus.
Although the article also mentions potential uses for storing AI model datasets and data availability layers (DA), the phased retreat of Web3 AI has left very few related projects, and the number of projects that will truly adopt Walrus's protocol in the future may be very limited.
As for the DA layer direction, whether Walrus can serve as an effective alternative still needs to wait for mainstream projects like Celestia to reignite market attention to verify its feasibility.
Therefore, Walrus's core positioning can be understood as a hot storage system serving content assets like NFTs, emphasizing dynamic invocation, real-time updates, and version management capabilities.
This also explains why Walrus needs to rely on Sui: with the help of Sui's high-performance chain capabilities, Walrus can build a high-speed data retrieval network, significantly reducing operational costs without developing a high-performance public chain itself, thus avoiding direct competition with traditional cloud storage services in terms of unit costs.
According to official data, Walrus's storage costs are about one-fifth of traditional cloud services. Although it appears to be dozens of times more expensive than Filecoin and Arweave, its goal is not to pursue extremely low costs but to build a decentralized hot storage system that can be used in real business scenarios. Walrus itself operates as a PoS network, with the core responsibility of verifying the honesty of storage nodes, providing the most basic security guarantees for the entire system.
As for whether Sui truly needs Walrus, it currently remains more at the level of ecological narrative. If financial settlement is the primary use case, Sui does not urgently need off-chain storage support. However, if it hopes to support more complex on-chain scenarios in the future, such as AI applications, content assetization, and composable agents, the storage layer will be indispensable in providing context, context, and indexing capabilities. High-performance chains can handle complex state models, but these states need to be bound to verifiable data to build a trustworthy content network.
Shelby: Dedicated Fiber Network Completely Unleashes Web3 Application Scenarios
Among the biggest technical bottlenecks facing current Web3 applications, "read performance" has always been a difficult shortcoming to overcome.
Whether it is video streaming, RAG systems, real-time collaboration tools, or AI model inference engines, they all rely on low-latency, high-throughput hot data access capabilities. Decentralized storage protocols (from Arweave, Filecoin to Walrus) have made progress in data persistence and trustlessness, but because they operate on the public internet, they cannot escape the limitations of high latency, unstable bandwidth, and uncontrollable data scheduling.
Shelby attempts to solve this problem from the root.
First, the Paid Reads mechanism directly reshapes the "read operation" dilemma in decentralized storage. In traditional systems, reading data is almost free, and the lack of effective incentive mechanisms leads to service nodes generally being lazy in responding and cutting corners, resulting in actual user experiences lagging far behind Web2.
Shelby links user experience directly to service node income by introducing a pay-per-read model: the faster and more stably nodes return data, the more rewards they can earn.
This model is not an "ancillary economic design," but rather the core logic of Shelby's performance design—without incentives, there is no reliable performance; with incentives, there is sustainable improvement in service quality.
Secondly, one of the biggest technical breakthroughs proposed by Shelby is the introduction of a Dedicated Fiber Network, which is equivalent to building a high-speed rail network for the instant reading of Web3 hot data.
This architecture completely bypasses the public transport layer that Web3 systems generally rely on, directly deploying storage nodes and RPC nodes on a high-performance, low-congestion, physically isolated transport backbone. This not only significantly reduces the latency of cross-node communication but also ensures the predictability and stability of transmission bandwidth. The underlying network structure of Shelby is closer to the dedicated line deployment model between AWS internal data centers, rather than the "upload to a miner node" logic of other Web3 protocols.
Source: Shelby White Paper
This network-level architectural reversal makes Shelby the first decentralized hot storage protocol capable of truly supporting Web2-level user experiences. Users reading a 4K video on Shelby, invoking embedding data from a large language model, or backtracking a transaction log no longer need to endure the second-level delays commonly found in cold data systems but can achieve sub-second responses. For service nodes, the dedicated network not only enhances service efficiency but also significantly reduces bandwidth costs, making the "pay-per-read" mechanism genuinely economically viable, thus incentivizing the system to evolve towards higher performance rather than higher storage volume.
It can be said that the introduction of the dedicated fiber network is the key support for Shelby to "look like AWS, but at its core is Web3." It not only breaks the inherent opposition between decentralization and performance but also opens up real possibilities for Web3 applications in high-frequency reading, high-bandwidth scheduling, and low-cost edge access.
In addition, in terms of data persistence and cost, Shelby adopts an Efficient Coding Scheme built with Clay Codes, achieving low storage redundancy of up to 2x while still maintaining 11 nines of persistence and 99.9% availability. While most Web3 storage protocols still hover around a redundancy rate of 5x to 15x today, Shelby is not only technically more efficient but also more competitive in terms of cost. This also means that for dApp developers who truly value cost optimization and resource scheduling, Shelby offers a "both cheap and fast" practical option.
Summary
Looking at the evolution from Filecoin, Arweave, Walrus to Shelby, we can clearly see that: the narrative of decentralized storage has gradually shifted from a technological utopia of "existence is justification" to a realistic approach of "usability is justice." Early Filecoin drove hardware participation through economic incentives, but real user needs were long marginalized; Arweave chose extreme permanent storage but appeared increasingly isolated amid a silent application ecosystem; Walrus attempted to find a new balance between cost and performance, yet questions remain regarding the construction of practical scenarios and incentive mechanisms. It was not until Shelby emerged that decentralized storage first provided a systematic response to "Web2-level usability"—from the dedicated fiber network at the transport layer, to the efficient erasure coding design at the computation layer, and to the pay-per-read incentive mechanism, these capabilities, originally exclusive to centralized cloud platforms, began to be reconstructed in the Web3 world.
The emergence of Shelby does not mean the end of problems. It has not solved all challenges: issues such as developer ecosystems, permission management, and terminal access still lie ahead. However, its significance lies in opening a possible path of "performance without compromise" for the decentralized storage industry, breaking the binary paradox of "either censorship-resistant or user-friendly."
The path to the popularization of decentralized storage will ultimately not rely solely on conceptual hype or token speculation but must move towards an application-driven stage of "usable, integrable, and sustainable." In this stage, whoever can first address the real pain points of users will reshape the narrative of the next round of infrastructure. From mining logic to usage logic, Shelby's breakthrough may mark the end of one era—more importantly, the beginning of another.
About Movemaker
Movemaker is the first official community organization authorized by the Aptos Foundation and jointly initiated by Ankaa and BlockBooster, focusing on promoting the construction and development of the Aptos ecosystem in the Chinese-speaking region. As the official representative of Aptos in the Chinese-speaking area, Movemaker is committed to building a diverse, open, and prosperous Aptos ecosystem by connecting developers, users, capital, and numerous ecological partners.
Disclaimer:
This article/blog is for reference only, representing the author's personal views and does not reflect the position of Movemaker. This article does not intend to provide: (i) investment advice or investment recommendations; (ii) offers or solicitations to buy, sell, or hold digital assets; or (iii) financial, accounting, legal, or tax advice. Holding digital assets, including stablecoins and NFTs, carries high risks, with significant price volatility, and they may even become worthless. You should carefully consider whether trading or holding digital assets is suitable for you based on your financial situation. For specific issues, please consult your legal, tax, or investment advisor. The information provided in this article (including market data and statistics, if any) is for general reference only. Reasonable care has been taken in compiling this data and charts, but no responsibility is accepted for any factual errors or omissions expressed therein.
免责声明:本文章仅代表作者个人观点,不代表本平台的立场和观点。本文章仅供信息分享,不构成对任何人的任何投资建议。用户与作者之间的任何争议,与本平台无关。如网页中刊载的文章或图片涉及侵权,请提供相关的权利证明和身份证明发送邮件到support@aicoin.com,本平台相关工作人员将会进行核查。