Author: Jane Doe, Chen Li, Corresponding author: Youbi Investment Team
1 The intersection of AI and Crypto
On May 23, chip giant Nvidia released its first quarter financial report for fiscal year 2025. The financial report shows that Nvidia's revenue in the first quarter was US$26 billion. Among them, data center revenue increased by 427% over last year to a staggering US$22.6 billion. Behind Nvidia's financial performance, which can save the U.S. stock market on its own, is the demand for computing power that has exploded among global technology companies in order to compete in the AI track. The more top technology companies are ambitious in their layout in the AI track, the more their demand for computing power has increased exponentially. According to TrendForce's forecast, in 2024, the demand for high-end AI servers from the four major cloud service providers in the United States: Microsoft, Google, AWS and Meta are expected to account for 20.2%, 16.6%, 16% and 10.8% of global demand, respectively, totaling more than 60%.
Image source: https://investor.nvidia.com/financial-info/financial-reports/default.aspx
"Chip shortage" has become the annual buzzword in recent years. On the one hand, the training and inference of large language models (LLMs) require a lot of computing power support; and with the iteration of the model, the cost and demand of computing power increase exponentially. On the other hand, large companies like Meta will purchase a huge amount of chips, and the global computing power resources are tilted towards these technology giants, making it increasingly difficult for small businesses to obtain the computing power resources they need. The difficulties faced by small businesses come not only from the insufficient supply of chips caused by the surge in demand, but also from the structural contradictions of supply. At present, there are still a large number of idle GPUs on the supply side. For example, some data centers have a large amount of idle computing power (the utilization rate is only 12%-18%), and a large amount of computing power resources are idle due to the reduction of profits in crypto mining. Although these computing powers are not all suitable for professional application scenarios such as AI training, consumer-grade hardware can still play a huge role in other fields such as AI inference, cloud game rendering, and cloud phones. The opportunity to integrate and utilize this part of computing power resources is huge.
Turning our attention from AI to crypto, after three years of silence in the crypto market, it finally ushered in another round of bull market. Bitcoin prices have repeatedly hit new highs, and various memecoins have emerged in an endless stream. Although AI and Crypto have been popular as buzzwords for these years, artificial intelligence and blockchain as two important technologies are like two parallel lines, and they have not yet found an "intersection point". At the beginning of this year, Vitalik published an article titled "The promise and challenges of crypto + AI applications", discussing the future scenarios of the combination of AI and crypto. Vitalik mentioned a lot of visions in the article, including using encryption technologies such as blockchain and MPC to decentralize AI training and inference, which can open the black box of machine learning and make AI models more trustless. There is still a long way to go to realize these visions. But one of the use cases mentioned by Vitalik - using crypto's economic incentives to empower AI is also an important direction that can be realized in a short time. The decentralized computing network is one of the most suitable scenarios for AI + crypto at this stage.
2 Decentralized computing network
At present, there are already many projects developing on the track of decentralized computing network. The underlying logic of these projects is similar, which can be summarized as: Using tokens to incentivize computing power holders to participate in the network to provide computing power services, these scattered computing power resources can be aggregated into a decentralized computing power network of a certain scale. This can not only improve the utilization rate of idle computing power, but also meet customers' computing power needs at a lower cost, and achieve a win-win situation for both buyers and sellers.
In order to enable readers to gain an overall grasp of this track in a short period of time, this article will deconstruct specific projects and the entire track from the micro-macro perspective, aiming to provide readers with an analytical perspective to understand the core competitive advantages of each project and the overall development of the decentralized computing power track. The author will introduce and analyze five projects: Aethir, io.net, Render Network, Akash Network, Gensyn, and summarize and evaluate the project status and track development.
From the perspective of the analytical framework, if we focus on a specific decentralized computing power network, we can break it down into four core components:
Hardware network: Integrate the scattered computing power resources together, and realize the sharing and load balancing of computing power resources through nodes distributed around the world, which is the basic layer of the decentralized computing power network.
Two-sided market: Match computing power providers with demanders through reasonable pricing and discovery mechanisms, provide a secure trading platform, and ensure that transactions between supply and demand parties are transparent, fair and credible.
Consensus mechanism: Used to ensure that nodes in the network run correctly and complete their work. The consensus mechanism is mainly used to monitor two levels: 1) Monitor whether the node is running online and is in an active state that can accept tasks at any time; 2) Node work proof: After receiving the task, the node effectively and correctly completes the task, and the computing power is not used for other purposes and occupies processes and threads.
Token incentive: The token model is used to incentivize more participants to provide/use services, and use tokens to capture this network effect and realize community benefit sharing.
If you take a bird's eye view of the entire decentralized computing power track, Blockworks Research's research report provides a good analysis framework. We can divide the project positions of this track into three different layers.
Bare metal layer: It constitutes the base layer of the decentralized computing stack. Its main task is to collect raw computing resources and make them available for API calls.
Orchestration layer: It constitutes the middle layer of the decentralized computing stack. Its main task is coordination and abstraction. It is responsible for the scheduling, expansion, operation, load balancing and fault tolerance of computing power. Its main function is to "abstract" the complexity of the underlying hardware management and provide a more advanced user interface for end users to serve specific customer groups.
Aggregation layer: It constitutes the top layer of the decentralized computing stack. Its main task is integration. It is responsible for providing a unified interface so that users can implement multiple computing tasks in one place, such as AI training, rendering, zkML, etc. It is equivalent to the orchestration and distribution layer of multiple decentralized computing services.
Image source: Youbi Capital
Based on the above two analysis frameworks, we will make a horizontal comparison of the five selected projects and evaluate them from four levels: core business, market positioning, hardware facilities and financial performance.
2.1 Core Business
From the underlying logic, the decentralized computing power network is highly homogenized, that is, using tokens to incentivize idle computing power holders to provide computing power services. Around this underlying logic, we can understand the differences in the core business of the projects from three aspects:
Hardware Types
Common computing hardware resources include CPU, FPGA, GPU, ASIC and SoC. These hardware have significant differences in design goals, performance characteristics and application areas. In summary, CPU is better at general computing tasks, FPGA's advantage lies in high parallel processing and programmability, GPU performs well in parallel computing, ASIC is most efficient in specific tasks, and SoC integrates multiple functions in one, suitable for highly integrated applications. Which hardware to choose depends on the needs of the specific application, performance requirements and cost considerations. Most of the decentralized computing projects we discussed are for collecting GPU computing power, which is determined by the project business type and the characteristics of the GPU. Because GPU has unique advantages in AI training, parallel computing, multimedia rendering, etc.
Although most of these projects involve the integration of GPUs, different applications have different requirements for hardware specifications, so these hardware have heterogeneous optimization cores and parameters. These parameters include parallelism/serial dependencies, memory, latency, etc. For example, rendering workloads are actually more suitable for consumer-grade GPUs rather than more powerful data center GPUs, because rendering has high requirements for ray tracing, etc. Consumer-grade chips such as 4090s have enhanced RT cores and are specifically optimized for ray tracing tasks. AI training and inference require professional-level GPUs. Therefore, Render Network can collect consumer-grade GPUs such as RTX 3090s and 4090s from retail investors, while IO.NET needs more professional-level GPUs such as H100s and A100s to meet the needs of AI startups.
2.2 Market Positioning
In terms of project positioning, the bare metal layer, orchestration layer, and aggregation layer have different core issues to be solved, optimization priorities, and value capture capabilities.
The bare metal layer focuses on the collection and utilization of physical resources, while the orchestration layer focuses on the scheduling and optimization of computing power, optimizing the physical hardware according to the needs of the customer group. The aggregation layer is a general purpose layer that focuses on the integration and abstraction of different resources. From the perspective of the value chain, each project should start from the bare metal layer and strive to climb upward.
From the perspective of value capture, the value capture capability increases layer by layer from the bare metal layer, orchestration layer to the aggregation layer. The aggregation layer can capture the most value because the aggregation platform can obtain the largest network effect and directly reach the most users, which is equivalent to the traffic entrance of the decentralized network, thus occupying the highest value capture position in the entire computing resource management stack.
Correspondingly, it is also the most difficult to build an aggregation platform. The project needs to comprehensively solve many problems such as technical complexity, heterogeneous resource management, system reliability and scalability, network effect realization, security and privacy protection, and complex operation and maintenance management. These challenges are not conducive to the cold start of the project and depend on the development and timing of the track. It is not realistic to build an aggregation layer when the orchestration layer has not yet matured and taken up a certain market share.
Currently, Aethir, Render Network, Akash Network and Gensyn all belong to the orchestration layer, and they are designed to provide services for specific targets and customer groups. Aethir's main business is real-time rendering for cloud games, and provides certain development and deployment environments and tools for B-side customers; Render Network's main business is video rendering, Akash Network's mission is to provide a trading platform similar to Taobao, and Gensyn is deeply involved in the field of AI training. io.net is positioned as an aggregation layer, but the functions currently implemented by io are still some distance away from the complete functions of the aggregation layer. Although the hardware of Render Network and Filecoin has been collected, the abstraction and integration of hardware resources have not yet been completed.
2.3 Hardware Facilities
At present, not all projects have published detailed network data. Relatively speaking, the UI of io.net explorer is the best, on which you can see the number, type, price, distribution, network usage, node income and other parameters of GPU/CPU. However, at the end of April, the front end of io.net was attacked. Because io did not do Auth for the PUT/POST interface, hackers tampered with the front-end data. This has also sounded the alarm for the privacy and network data reliability of other projects.
In terms of the number and model of GPUs, io.net, as an aggregation layer, should have collected the most hardware. Aethir followed closely behind, and the hardware situation of other projects was not so transparent. From the GPU model, we can see that io has both professional-grade GPUs such as A100 and consumer-grade GPUs such as 4090, with a wide variety, which is in line with the positioning of io.net aggregation. io can choose the most suitable GPU according to specific task requirements. However, different models and brands of GPUs may require different drivers and configurations, and the software also needs to be complexly optimized, which increases the complexity of management and maintenance. At present, the allocation of various tasks of io is mainly based on the user's own choice.
Aethir released its own mining machine. In May, Aethir Edge, supported by Qualcomm, was officially launched. It will break the single centralized GPU cluster deployment method far away from users and deploy computing power to the edge. Aethir Edge will combine the cluster computing power of H100 to serve AI scenarios together. It can deploy trained models and provide users with inference computing services at the best cost. This solution is closer to users, provides faster services, and is more cost-effective.
From the perspective of supply and demand, taking Akash Network as an example, its statistics show that the total number of CPUs is about 16k and the number of GPUs is 378. According to the network rental demand, the utilization rates of CPUs and GPUs are 11.1% and 19.3% respectively. Among them, only the professional-grade GPU H100 has a relatively high rental rate, and most of the other models are idle. The situation faced by other networks is roughly the same as Akash. The overall demand for the network is not high. Except for popular chips such as A100 and H100, other computing power is mostly idle.
From the perspective of price advantage, compared with other traditional service providers, except for the giants in the cloud computing market, the cost advantage is not prominent.
2.4 Financial Performance
No matter how the token model is designed, a healthy tokenomics needs to meet the following basic conditions: 1) User demand for the network needs to be reflected in the coin price, that is, the token can capture value; 2) All participants, whether developers, nodes, or users, need to receive long-term and fair incentives; 3) Ensure decentralized governance to avoid excessive holdings by insiders; 4) Reasonable inflation and deflation mechanisms and token release cycles to avoid large fluctuations in coin prices affecting the robustness and sustainability of the network.
If we divide the token model into BME (burn and mint equilibrium) and SFA (stake for access), the sources of deflationary pressure of the tokens in these two models are different: the BME model burns tokens after users purchase services, so the deflationary pressure of the system is determined by demand. SFA requires service providers/nodes to stake tokens to obtain the qualification to provide services, so the deflationary pressure is brought by supply. The advantage of BME is that it is more suitable for non-standardized goods. However, if the demand of the network is insufficient, it may face the pressure of continuous inflation. The token models of various projects are different in details, but in general, Aethir prefers SFA, while io.net, Render Network and Akash Network prefer BME, and Gensyn is still unknown.
In terms of income, the demand of the network will be directly reflected in the overall income of the network (the income of miners is not discussed here, because miners receive subsidies from the project in addition to the rewards for completing tasks.) From the public data, io.net has the highest value. Although Aethir's revenue has not been announced, according to public information, they have announced that they have signed orders with many B-side customers.
In terms of coin price, only Render Network and Akash Network have conducted ICO. Aethir and io.net have also issued coins recently. The price performance needs to be observed and will not be discussed in detail here. Gensyn's plan is still unclear. From the two projects that have issued coins and the projects that have issued coins in the same track but are not included in the scope of this article, in general, the decentralized computing power network has a very impressive price performance, which to a certain extent reflects the huge market potential and high expectations of the community.
2.5 Summary
The decentralized computing power network track has developed rapidly overall, and many projects can rely on products to serve customers and generate certain income. The track has moved away from pure narrative and entered the development stage where it can provide preliminary services.
Weak demand is a common problem faced by decentralized computing networks, and long-term customer demand has not been well verified and explored. However, the demand side has not affected the price of the currency too much, and several projects that have already issued coins have performed well.
AI is the main narrative of decentralized computing networks, but it is not the only business. In addition to being used in AI training and inference, computing power can also be used for real-time rendering of cloud games, cloud mobile phone services, etc.
The hardware heterogeneity of computing power networks is relatively high, and the quality and scale of computing power networks need to be further improved.
For C-end users, the cost advantage is not very obvious. For B-end users, in addition to cost savings, they also need to consider the stability, reliability, technical support, compliance and legal support of services, etc., and Web3 projects generally do not do well in these aspects.
3 Closing thoughts
The explosive growth of AI has brought about a huge demand for computing power. Since 2012, the amount of computing power used in AI training tasks has been growing exponentially, currently doubling every 3.5 months (compared to Moore's Law, which was doubling every 18 months). The demand for computing power has grown more than 300,000 times since 2012, far outstripping Moore's Law's 12-fold growth. The GPU market is expected to grow at a 32% compound annual growth rate to more than $200 billion over the next five years. AMD's estimate is even higher, with the company predicting that the GPU chip market will reach $400 billion by 2027.
Image source: https://www.stateof.ai/
Because the explosive growth of artificial intelligence and other computationally intensive workloads (such as AR/VR rendering) has exposed the structural inefficiency in traditional cloud computing and leading computing markets. In theory, decentralized computing networks can provide more flexible, low-cost and efficient solutions by utilizing distributed idle computing resources, thereby meeting the huge market demand for computing resources. Therefore, the combination of crypto and AI has huge market potential, but it also faces fierce competition with traditional companies, high entry barriers and complex market environment. In general, looking at all crypto tracks, decentralized computing networks are one of the vertical fields in the crypto field that is most likely to obtain real demand.
Image source: https://vitalik.eth.limo/general/2024/01/30/cryptoai.html
The future is bright, but the road is tortuous. To achieve the above vision, we still need to solve many problems and challenges. In summary: At this stage, if we simply provide traditional cloud services, the profit margin of the project is very small. From the demand side, large enterprises generally build their own computing power, and pure C-end developers mostly choose cloud services. Whether small and medium-sized enterprises that truly use decentralized computing power network resources will have stable demand needs further exploration and verification. On the other hand, AI is a vast market with a very high ceiling and imagination space. For a broader market, decentralized computing service providers will also need to transform to model/AI services in the future, explore more crypto + AI usage scenarios, and expand the value that projects can create. But at present, there are still many problems and challenges to further develop into the field of AI:
The price advantage is not outstanding: From the previous data comparison, it can be seen that the cost advantage of decentralized computing network has not been reflected. The possible reason is that for professional chips H100, A100, etc. with high demand, the market mechanism determines that the price of this part of the hardware will not be cheap. In addition, although decentralized networks can collect idle computing resources, the lack of economies of scale brought by decentralization, high network and bandwidth costs, and huge management and operation complexity and other hidden costs will further increase the computing cost.
The particularity of AI training: There are huge technical bottlenecks in the current stage of using decentralized methods for AI trainingning. This bottleneck can be intuitively reflected in the GPU workflow. In the training of large language models, the GPU first receives the preprocessed data batches and performs forward propagation and back propagation calculations to generate gradients. Next, each GPU aggregates the gradients and updates the model parameters to ensure that all GPUs are synchronized. This process will be repeated until all batches are trained or the predetermined number of rounds is reached. This process involves a lot of data transmission and synchronization. What kind of parallel and synchronization strategies to use, how to optimize network bandwidth and latency, and reduce communication costs, etc., have not yet been well answered. At this stage, it is not realistic to use decentralized computing power networks to train AI.
Data security and privacy: In the training process of large language models, all links involving data processing and transmission, such as data allocation, model training, parameter and gradient aggregation, may affect data security and privacy. And data privacy is more important than model privacy. If the problem of data privacy cannot be solved, it cannot be truly scaled on the demand side.
From the most realistic perspective, a decentralized computing power network needs to take into account both current demand discovery and future market space. Find the right product positioning and target customer groups, such as targeting non-AI or Web3 native projects first, starting with relatively marginal needs, and building an early user base. At the same time, continue to explore various scenarios combining AI and crypto, explore the forefront of technology, and achieve transformation and upgrading of services.
References
https://www.stateof.ai/
https://vitalik.eth.limo/general/2024/01/30/cryptoai.html
https://foresightnews.pro/article/detail/34368
https://app.blockworksresearch.com/unlocked/compute-de-pi-ns-paths-to-adoption-in-an-ai-dominated-market?callback=%2Fresearch%2Fcompute-de-pi-ns-paths-to-adoption-in-an-ai-dominated-market https://research.web3caff.com/zh/archives/17351?ref=1554