Authors: @charlotte0211z, @BlazingKevin_, Metrics Ventures
Published by Vitalik on January 30The promise and challenges of crypto + AI applications This article discusses how blockchain and artificial intelligence should be combined, and what emerges in the process. potential challenges. One month after this article was published, NMR, Near, and WLD mentioned in the article all achieved good gains, completing a round of value discovery. Based on the four methods of combining Crypto and AI proposed by Vitalik, this article sorts out the subdivision directions of the existing AI track and briefly introduces representative projects in each direction.
1 Introduction: Four ways to combine Crypto with AI
Decentralization is the consensus maintained by the blockchain, ensuring security is the core idea, and open source is the key foundation for making on-chain behavior have the above characteristics from a cryptographic perspective. In the past few years, this approach has been applicable to several rounds of blockchain changes, but when artificial intelligence is involved, the situation changes.
Imagine using artificial intelligence to design the architecture of a blockchain or application. Then it is necessary to open source the model. However, this will expose its vulnerability to adversarial machine learning; otherwise, it will be lost. Decentralization. Therefore, it is necessary for us to think about how and how deeply to integrate artificial intelligence into current blockchains or applications.
Source: DE UNIVERSITY OF ETHEREUM
at DE UNIVERSITY OF ETHEREUM's When Giants Collide: Exploring the Convergence of Crypto x AI’s article explains the differences in the core characteristics of artificial intelligence and blockchain. As shown in the figure above, the characteristics of artificial intelligence are:
Blockchain is completely opposite to artificial intelligence in these five points. This is also the real argument of Vitalik’s article. If artificial intelligence and blockchain are combined, what kind of trade-offs should be made in terms of data ownership, transparency, monetization capabilities, energy consumption costs, etc., and what infrastructure needs to be created? Ensure the effective combination of the two.
In accordance with the above principles and its own thinking, Vitalik divides applications combining artificial intelligence and blockchain into four major categories:
AI as a player in a game
AI as a player in a game AI as an interface to the game
AI as the rules of the game
AI as the rules of the game strong>
AI as the objective of the game
Among them, the first three are mainly the three ways in which AI is introduced into the Crypto world, representing three levels from shallow to deep. According to the author’s understanding, this division represents the degree of influence of AI on human decision-making, and is represented by This introduces varying degrees of system risks to the entire Crypto:
Artificial intelligence as a participant in the application: artificial intelligence itself will not affect human beings Decisions and behaviors have consequences, so they do not bring risks to the real human world, and therefore have the highest level of implementation at present.
Artificial intelligence as an application interface: Artificial intelligence provides auxiliary information or auxiliary tools for human decision-making and behavior, which will improve user and developer experience and lower the threshold, but it is wrong Information or operations will carry certain real-world risks.
The rules of artificial intelligence as an application: artificial intelligence will fully replace humans to complete decision-making and operations, so the evildoings and failures of artificial intelligence will directly lead to chaos in the real world, whether in Neither Web2 nor Web3 can currently trust artificial intelligence to make decisions on behalf of humans.
Finally, the fourth category of projects is dedicated to using the characteristics of Crypto to create better artificial intelligence. As mentioned above, centralization, low transparency, energy consumption, and monopoly Weak currency attributes can be naturally neutralized by Crypto’s attributes. Although many people have doubts about whether Crypto can have an impact on the development of artificial intelligence, influencing the real world through the power of decentralization has always been Crypto’s most fascinating narrative. This track has also become an AI competition with its grand ideas. The most hyped part.
2 AI as a participant
In the mechanism of AI participation, the ultimate source of incentives comes from the agreement entered by humans. Before AI becomes an interface or even a rule, we often need to evaluate the performance of different AIs so that the AI can participate in a mechanism and ultimately be rewarded or punished through an on-chain mechanism.
As a participant, AI’s risk to users and the entire system is basically negligible compared to its role as interfaces and rules. It can be said that it is a necessary step before AI begins to deeply affect user decisions and behaviors. stage, so the costs and trade-offs required for the integration of artificial intelligence and blockchain at this level are relatively small, and it is also a type of product that Buterin believes is now highly implementable.
From a broad perspective and implementation level, current AI applications mostly fall into this category, such as AI-empowered trading bots and chatbots. At the current level of implementation, it is difficult to implement AI as an interface or even a rule. Users are comparing and gradually optimizing among different bots, and crypto users have not yet developed the behavioral habit of using AI applications. In V God's article, Autonomous Agent is also classified into this category.
But in a narrow sense and long-term vision, we tend to classify AI applications or AI Agents in more detail. Therefore, under this category, we consider representative subdivisions. The categories include:
2.1 AI games
To some extent, AI games can be classified into this category. Players interact with AI and train their AI characters. Make AI characters more in line with personal needs, such as being more in line with personal preferences or more combat-effective and competitive in game mechanics. Games are a transitional stage before AI enters the real world. It is also a track that currently has low risk of implementation and is easiest for ordinary users to understand. Iconic projects such as AI Arena, Echelon Prime, Altered State Machine, etc.
AI Arena: AI Arena is a game where players can learn and train through AI to continuously evolve their game characters. The PVP fighting game hopes to allow more ordinary users to contact, understand and experience AI in the form of games, and at the same time allow artificial intelligence engineers to provide various AI algorithms based on AI Arena to increase income. Each game character is an NFT empowered by AI, of which Core is the core containing the AI model, including two parts: architecture and parameters, which are stored on IPFS. The parameters in a new NFT are randomly generated, which means It will perform random actions, and users need to improve the character's strategic capabilities through the process of imitation learning (IL). Every time the user trains the character and saves the progress, the parameters will be updated on IPFS.
Altered State Machine: ASM is not an AI game, but a protocol for AI Agent to confirm rights and trade. Positioned as the Metaverse AI protocol, it is currently being integrated with multiple games including FIFA, introducing AI Agents into games and the Metaverse. ASM uses NFT to confirm and trade AI Agents. Each Agent will contain three parts: Brain (the Agent's own characteristics), Memories (the part that stores the behavioral strategies learned by the Agent and model training, and is bound to the Brain) , Form (character appearance, etc.). ASM has a Gym module, including a decentralized GPU cloud provider, which can provide computing power support for the Agent. Projects currently using ASM as the underlying include AIFA (AI football game), Muhammed Ali (AI boxing game), AI League (street football game in partnership with FIFA), Raicers (AI-driven racing game), and FLUF World's Thingies (generative NFT).
Parallel Colony (PRIME): Echelon Prime is developing Parallel Colony, an AI LLM-based game that players You can interact with and influence your AI Avatar, and the Avatar will act autonomously based on memories and life trajectories. Colony is currently one of the most anticipated AI games. Recently, the official released a white paper and announced its migration to Solana, which has ushered in a new wave of gains for PRIME.
2.2 Prediction market/competition
Prediction ability is the basis for AI’s future decision-making and behavior. Before the AI model is used for actual prediction, prediction The competition compares the performance of AI models at a higher level and provides incentives for data scientists/AI models through tokens, which has positive significance for the development of the entire Crypto×AI - through incentives, continuous development of stronger, more efficient and better performance Models and applications suitable for the crypto world, creating better and more secure products before AI exerts a deeper impact on decision-making and behavior. As V God said, prediction market is a powerful primitive that can be expanded to other types of problems. Iconic projects in this track include: Numerai and Ocean Protocol.
Numerai: Numerai is a data science competition that has been running for a long time. Data scientists compete based on historical market data. (Provided by Numerai) Train a machine learning model to predict the stock market, and stake the model and NMR tokens for tournaments. Models with better performance will be rewarded with NMR tokens, while the pledged tokens of poorer models will be destroyed. As of March 7, 2024, a total of 6,433 models have been staked, and the protocol has provided a total of $75,760,979 in incentives to data scientists. Numerai is inspiring global data scientists to collaborate to build new types of hedge funds, with funds launched so far including Numerai One and Numerai Supreme. Numerai’s path: market prediction competition → crowdsourcing prediction model → new hedge fund based on crowdsourcing model.
Ocean Protocol: Ocean Predictor is focusing on predictions, starting with crowdsourced predictions of cryptocurrency trends. Players can choose to run Predictoor bot or Trader bot. Predictoor bot uses an AI model to predict the price of cryptocurrency (such as BTC/USDT) at the next point in time (such as five minutes later) and pledges a certain amount of $OCEAN. The agreement will be based on The global prediction is calculated by weighting the amount of pledges. Traders purchase the prediction results and can trade based on them. When the accuracy of the prediction results is high, Traders can profit from it. Predictors who predict wrongly will be punished, while those who predict correctly can get This portion of the tokens and Traders purchase fees serve as rewards. On March 2, Ocean Predictor announced its latest direction - World-World Model (WWM) in the media, and began to explore real-world predictions such as weather and energy.
3 AI as an interface
AI can help users understand what is happening in simple and easy-to-understand language, act as a mentor for users in the crypto world, and provide tips on possible risks to reduce Crypto use thresholds and user risks to improve user experience. The specific achievable products have rich functions, such as risk reminders during wallet interaction, AI-driven intention transactions, AI Chatbot that can answer ordinary users' crypto questions, etc. By expanding the audience group, almost all groups, including ordinary users, developers, analysts, etc., will become the service targets of AI.
Let us reiterate what these projects have in common: they have not yet replaced humans to perform certain decisions and behaviors, but they are using AI models to provide humans with information and tools to assist decision-making and behaviors. Starting from this level, the risk of AI evil has begun to be exposed in the system - it can interfere with human beings' final judgment by providing wrong information. This has also been analyzed in detail in Buterin's article.
There are many and complex projects that can be classified under this category, including AI chatbot, AI smart contract audit, AI code writing, AI trading bot, etc. It can be said that the vast majority of them currently AI applications are at the primary level of this category. Representative projects include:
PaaL: PaaL It is currently the leading project of AI Chatbot. It can be regarded as ChatGPT trained with crypto-related knowledge. By integrating TG and Discord, it can provide users with: token data analysis, token fundamentals and token economics analysis, and text generation pictures. For other functions, PaaL Bot can be integrated into group chats to automatically reply to some information. PaaL supports customizing personal bots, and users can build their own AI knowledge base and custom bots by feeding data sets. Paal is moving towards AI Trading Bot. On February 29, it announced the release of its AI-supported crypto research & trading terminal PaalX. According to the introduction, it can realize AI smart contract auditing, Twitter-based news integration and trading, Crypto research and trading support. , artificial intelligence assistant can lower the user threshold.
ChainGPT: ChainGPT relies on artificial intelligence to develop a series of crypto tools, such as chatbot, NFT generator, news collection, smart Contract generation and audit, trading assistant, Prompt market and AI cross-chain exchange. However, ChainGPT’s current focus is on project incubation and Launchpad, and it has completed IDOs for 24 projects and 4 Free Giveaways.
Arkham: Ultra is Arkham's dedicated AI engine. The use case is to match addresses with real entities through algorithms. Increase transparency in the crypto industry. Ultra combines the on-chain and off-chain data provided by users and collected by itself, and outputs it into an expandable database, which is finally presented in the form of charts. However, the Ultra system is not discussed in detail in the Arkham document. The reason why Arkham has received attention this time is the personal investment of OpenAI founder Sam Altman, which has gained 5 times in the past 30 days.
GraphLinq: GraphLinq is an automated process management solution designed to enable users to deploy and manage various types of automation functions without programming , such as pushing the price of Bitcoin in Coingecko to TG Bot every 5 minutes. GraphLinq's solution is to use Graph to visualize the automation process. Users can create automated tasks by dragging nodes and execute them using GraphLinq Engine. Although no code is required, the process of creating Graph still has a certain threshold for ordinary users, including selecting the appropriate template, selecting the appropriate one among hundreds of logical blocks, and connecting them. Therefore, GraphLinq is introducing AI, allowing users to use conversational artificial intelligence and natural language to complete the construction and management of automated tasks.
**0x0.ai: **0x0 has three main AI-related businesses: AI smart contract audit, AI anti-Rug detection and AI developer center. Among them, AI anti-Rug detection will detect suspicious behaviors, such as excessive taxes or draining liquidity, to prevent users from being deceived. The AI developer center uses machine learning technology to generate smart contracts and realize No-code deployment contracts. However, AI smart contract auditing has only been initially launched, and the other two functions have not yet been developed.
Zignaly: Zignaly was born in 2018, aiming to allow individual investors to choose fund managers to manage crypto assets for themselves, similar to Copy- trading logic. Zignaly is using machine learning and artificial intelligence technology to establish an indicator system for systematic evaluation of fund managers. The first product currently launched is Z-Score, but as an artificial intelligence product, it is still relatively rudimentary.
4 AI as the rules of the game
This is the most exciting part - allowing AI to replace humans in decision-making and behavior. Your AI will directly control your wallet and conduct transactions on your behalf. Decision-making and behavior. Under this classification, the author believes that it can be mainly divided into three levels: AI applications (especially applications with the vision of autonomous decision-making, such as AI automated trading bot, AI DeFi income bot), Autonomous Agent protocol and zkml/ opml.
AI applications are tools for making specific decisions on problems in a certain field. They accumulate knowledge and data in different subdivisions and rely on AI tailored to subdivided problems. Model makes decisions. It can be noted that AI applications are classified into two categories in this article: interfaces and rules. From the perspective of development vision, AI applications should become independent decision-making agents. However, at present, neither the effectiveness of the AI model nor the security of integrated AI , are unable to meet this requirement, and are even slightly reluctant as an interface. AI applications are in a very early stage. The specific projects have been introduced previously, so I will not go into details here.
Autonomous Agent was mentioned by Buterin in the first category (AI as a participant). From the perspective of long-term vision, this article classifies it into the third category. Autonomous Agent utilizes large amounts of data and algorithms to simulate human thinking and decision-making processes and perform various tasks and interactions. This article focuses on the Agent's communication layer, network layer and other infrastructure. These protocols define the Agent's ownership rights, establish the Agent's identity, communication standards and communication methods, connect multiple Agent applications, and enable collaborative decision-making and behavior.
zkML/opML: Using cryptography or economics methods to ensure that the correct model inference process is used to provide credible output. Security issues are very fatal to the introduction of AI into smart contracts. Smart contracts rely on input to generate output and automatically perform a series of functions. Once the AI misbehaves and gives wrong input, it will introduce great systemic risks to the entire Crypto system, so zkML /opML and a range of potential solutions are the basis for allowing AI to act and make decisions independently.
Finally, the three constitute the three basic levels of AI as operating rules: zkml/opml, as the lowest infrastructure, ensures the security of the protocol; the Agent protocol establishes the Agent ecosystem, which can Collaborative decision-making and behavior; AI applications, also specific AI Agents, will continuously improve their capabilities in a certain field and actually make decisions and actions.
4.1 Autonomous Agent
The application of AI Agent in the Crypto world is natural. From smart contracts to TG Bots to AI Agents, the crypto world is moving towards a higher level. Automation and lower user barriers. Although smart contracts automatically execute functions through non-tamperable code, they still need to rely on external triggers to wake up, and cannot run autonomously and continuously; TG Bots lowers the user threshold. Users do not need to interact directly with the encryption front-end, but through natural Language completes on-chain interactions, but it can only complete extremely simple and specific tasks, and still cannot achieve user-intent-centered transactions; AI Agents have certain independent decision-making capabilities, understand the user's natural language, and independently find and combine other Agents and on-chain tools to accomplish user-specified goals.
AI Agent is committed to greatly improving the experience of using encryption products, and blockchain can also help the operation of AI Agent to be more decentralized, transparent and secure. The specific help is:
< ul class=" list-paddingleft-2">
Incentivize more developers to provide Agent through tokens
NFT confirmation promotes Agent-based charging and transactions
Provide Agent identity and registration mechanism on the chain
Provide non-tamperable Agent activity logs to monitor their behavior Carry out timely traceability and accountability
The main projects of this track are as follows:
-
Autonolas: Autonolas supports the asset confirmation and composability of Agents and related components through on-chain protocols, enabling code components, Agents and services to be discovered and reused on the chain, and incentivizing Developers receive financial compensation. After developers develop a complete Agent or component, they will register the code on the chain and obtain NFT, which represents the ownership of the code; the Service Owner will unite multiple Agents to create a service and register it on the chain, and attract Agent Operators to The service is actually performed and the user pays to use the service.
Fetch.ai: Fetch.ai has a strong team background and development experience in the AI field, and is currently focusing on the AI Agent track. The protocol consists of four key layers: AI Agents, Agentverse, AI Engine and Fetch Network. AI Agents are the core of the system, and others are frameworks and tools that assist in building Agent services. Agentverse is a software-as-a-service platform mainly used for creating and registering AI Agents. The goal of the AI Engine is to perform tasks by reading user natural language input, converting it into actionable characters, and selecting the most appropriate registered AI Agent in the Agentverse. Fetch Network is the blockchain layer of the protocol. AI Agents must be registered in the Almanac contract on the chain before they can start collaborative services with other Agents. It is worth noting that Autonolas is currently focusing on the construction of Agents in the crypto world, introducing off-chain Agent operations to the chain; Fetch.ai's focus includes the Web2 world, such as travel booking, weather prediction, etc.
Delysium: Delysium has transformed from a game to an AI Agent protocol. It mainly includes two layers: communication layer and blockchain layer. The communication layer is Delysium The backbone provides a secure and scalable infrastructure that enables fast and efficient communication between AI Agents. The blockchain layer authenticates the Agent and achieves tamper-proof records of Agent behavior through smart contracts. Specifically, the communication layer establishes a unified communication protocol between Agents and uses a standardized messaging system so that Agents can communicate with each other without any barriers through a common language. In addition, it establishes a service discovery protocol and API so that users and Other Agents can quickly discover and connect to available Agents. The blockchain layer mainly consists of two parts: Agent ID and Chronicle smart contract. Agent ID ensures that only legitimate Agents can access the network. Chronicle is a log repository of all important decisions and behaviors made by the Agent. It cannot be tampered with after being uploaded to the chain. , ensuring trustworthy traceability of Agent behavior.
Altered State Machine: NFT sets standards for Agent asset ownership and transactions. For detailed analysis, see Part 1. Although ASM currently mainly It is connected to the game, but as a basic specification, it also has the possibility of being extended to other Agent fields.
Morpheous: An AI Agent ecological network is being built. The protocol is designed to connect the four roles of Coder, Computer provider, Community Builder and Capital, respectively. The network provides AI Agents, computing power to support the operation of Agents, front-end and development tools, and funds. MOR will take the form of Fair launch to miners, stETH pledgers, Agent or smart contract development contributors, and community development contributors who provide computing power. Provide incentives.
4.2 zkML/opML
Zero-knowledge proof currently has two main application directions:
Prove that the operation is running correctly on the chain at a lower cost (ZK-Rollup and ZKP cross-chain bridge are taking advantage of this feature of ZK);
Privacy protection: You do not need to know the details of the calculation to prove that the calculation was performed correctly.
Similarly, the application of ZKP in machine learning can also be divided into two categories:
< li>Inference verification: that is, through ZK-proof, it is proved on the chain at a lower cost that the intensive calculation process of AI model inference is correctly executed off the chain.
Privacy protection: It can be divided into two categories. One is the protection of data privacy, that is, using private data for inference on a public model, and ZKML can be used to perform inference on private data. protection; the second is to protect the privacy of the model, hoping to hide the specific information such as the weight of the model, and calculate the output results from the public input.
The author believes that inference verification is more important to Crypto at present. We will further elaborate on the inference verification scenario here. Starting from AI as a participant to AI as the rule of the world, we hope to make AI a part of the on-chain process. However, the computational cost of AI model reasoning is too high to run directly on the chain. Putting this process off-chain, This means that we need to endure the trust issues caused by this black box - did the AI model runner tamper with my input? Was the model I specified used for inference? By converting ML models into ZK circuits, it is possible to: (1) upload smaller models to the chain, store small zkML models in smart contracts, and directly upload them to the chain to solve the problem of opacity; (2) complete inference off-chain , and generate a ZK proof at the same time, to prove the correctness of the reasoning process by running the ZK proof on the chain. The infrastructure will include two contracts - the main contract (using the ML model to output results) and the ZK-Proof verification contract.
zkML is still in a very early stage, facing technical problems in converting ML models to ZK circuits, as well as extremely high computational and cryptographic overhead costs. Like Rollup's development path, opML has become another solution from an economic perspective. opML uses Arbitrum's AnyTrust assumption, that is, each claim has at least one honest node to ensure that the submitter or at least one verifier is honest. of. However, OPML can only be an alternative to inference verification and cannot achieve privacy protection.
The current project is building the infrastructure of zkML and working hard to explore its applications. The establishment of applications is equally important because it needs to clearly prove to encryption users the important role of zkML and prove that the final value can offset the huge cost. . Among these projects, some focus on ZK technology research and development related to machine learning (such as Modulus Labs), while others are more general ZK infrastructure construction. Related projects include:
Modulus is using zkML to apply artificial intelligence to the on-chain reasoning process. Modulus launched the zkML prover Remainder on February 27, which achieves a 180-fold efficiency improvement compared to traditional AI inference on equivalent hardware. In addition, Modulus cooperates with multiple projects to explore practical use cases of zkML, such as working with Upshot to collect complex market data, evaluate NFT prices, and transmit prices to the chain by using artificial intelligence with ZK proof; and AI Arena Cooperate to prove that the Avatar being fought is the same one trained by the player.
Risc Zero puts the model on the chain, and by running the machine learning model in RISC Zero's ZKVM, it can be proven that the exact calculations involved in the model were performed correctly.
Ingonyama is developing hardware specifically for ZK technology, which may lower the barrier to entry into the field of ZK technology, and zkML may also be used in the model training process.
5 AI as a goal
If the first three categories focus more on how AI empowers Crypto, then "AI as a goal" emphasizes how Crypto Help for AI, that is, how to use Crypto to create better AI models and products, may include multiple evaluation criteria: more efficient, more accurate, more decentralized, etc.
AI includes three cores: data, computing power and algorithm. In each dimension, Crypto is committed to providing more effective assistance to AI:
Data: Data is the basis for model training. The decentralized data protocol will encourage individuals or enterprises to provide more private domain data, while using cryptography to ensure data privacy and avoid the leakage of sensitive personal data. Give way.
Computing power: The decentralized computing power track is currently the hottest AI track. The agreement promotes long-tail computing power and AI companies by providing a matching market between supply and demand. matching for model training and inference.
Algorithm: Crypto’s empowerment of algorithms is the core link in realizing decentralized AI. It is also the main content of the “AI as a goal” narrative in Buterin’s article. Creating decentralized With centralized, trustworthy black box AI, the problem of adversarial machine learning mentioned above will be solved, but it will face a series of obstacles such as extremely high cryptography overhead. Furthermore, “using cryptographic incentives to encourage making better AI” can also be achieved without completely falling down the cryptographic rabbit hole of full encryption.
The monopoly of data and computing power by large technology companies has jointly resulted in a monopoly on the model training process, and closed-source models have become the key to profits for large enterprises. From an infrastructure perspective, Crypto encourages the decentralized supply of data and computing power through economic means, while ensuring data privacy in the process through cryptographic methods, and based on this, it facilitates decentralized model training to achieve More transparent and decentralized AI.
5.1 Decentralized Data Protocol
Decentralized Data Protocol is mainly carried out in the form of data crowdsourcing, encouraging users to provide data sets or data services (such as data annotation) for enterprise use Carry out model training and open a Data Marketplace to promote the matching of supply and demand. Some protocols are also exploring the use of DePIN incentive protocols to obtain users' browsing data, or use the user's equipment/bandwidth to complete network data crawling.
Ocean Protocol: To confirm and tokenize data, users can use the code-free method on Ocean Protocol After completing the NFT creation of the data/algorithm, create the corresponding datatoken to control access to the data NFT. Ocean Protocol ensures the privacy of data through Compute To Data (C2D). Users can only obtain the output results based on the data/algorithm, but cannot download them in full. Ocean Protocol was established in 2017. As a data market, it is natural to catch up with the AI express train in this round of craze.
Synesis One: This project is the Train2Earn platform on Solana. Users can obtain $SNS rewards by providing natural language data and data annotation. Users By providing data to support mining, the data will be stored and uploaded to the chain after verification, and used by AI companies for training and reasoning. Specifically, miners are divided into three categories: Architect/Builder/Validator. Architect is responsible for creating new data tasks, Builder provides corpus in the corresponding data tasks, and Validator verifies the data set provided by Builder. The completed data set will be stored in IPFS, and the data source and IPFS address will be saved on the chain and will be stored in an off-chain database for use by the AI company (currently Mind AI).
Grass: The decentralized data layer known as AI is essentially a decentralized web scraping market , and use this to obtain data for AI model training. Internet websites are an important source of AI training data. The data of many websites, including Twitter, Google, and Reddit, are of great value, but these websites are constantly restricting data crawling. Grass uses unused bandwidth in personal networks to reduce the impact of data blocking by using different IP addresses to capture data from public websites, complete preliminary data cleaning, and become a data source for AI model training companies and projects. Grass is currently in the beta testing phase, and users can provide bandwidth to earn points to receive potential airdrops.
AIT Protocol: AIT Protocol is a decentralized data annotation protocol designed to provide developers with high-quality data sets. for model training. Web3 enables the global workforce to quickly access the network and gain incentives through data annotation. AIT's data scientists will pre-annotate the data, which will then be further processed by users. After being inspected by the data scientists, the data that has passed the quality inspection will be provided to developers. By.
In addition to the above-mentioned data provision and data annotation protocols, previous decentralized storage infrastructure, such as Filecoin, Arweave, etc., will also provide assistance for more decentralized data.
5.2 Decentralized Computing Power
In the AI era, the importance of computing power is self-evident. Not only is Nvidia’s stock price climbing to peaks, but in the Crypto world, decentralized computing power It can be said to be the most hyped segment of the AI track - among the 11 AI projects in the top 200 by market value, 5 are projects that focus on decentralized computing power (Render/Akash/AIOZ Network/Golem/Nosana) , and has achieved high-fold gains in the past few months. We have also seen the emergence of many decentralized computing power platforms in projects with small market capitalization. Although they have just started, with the wave of Nvidia Conference, as long as they are related to GPU, they have quickly gained a surge.
From the perspective of the characteristics of the track, the basic logic of projects in this direction is highly homogeneous - through token incentives, people or companies with idle computing power resources are provided with resources, and thus the use of resources is significantly reduced. Fees establish a supply and demand market for computing power. Currently, the main supply of computing power comes from data centers, miners (especially after Ethereum switches to PoS), consumer computing power, and cooperation with other projects. Although homogeneous, this is a track with a high moat for leading projects. The main competitive advantages of the project come from: computing power resources, computing power rental price, computing power utilization rate and other technical advantages. Leading projects in this track include Akash, Render, io.net and Gensyn.
According to specific business directions, projects can be roughly divided into two categories: AI model inference and AI model training. Since AI model training requires much higher computing power and bandwidth than inference and is more difficult to implement than distributed inference, and the market for model inference is rapidly expanding, the predictable income will be significantly higher than that of model training in the future. Therefore, it is absolutely impossible to implement it at present. Most projects focus on inference (Akash, Render, io.net), and the leader in training is Gensyn. Among them, Akash and Render were born earlier and were not born for AI computing. Akash was originally used for general computing, Render is mainly used for video and image rendering, and io.net is specially designed for AI computing, but in AI computing power will After the demand has increased to a higher level, these projects have tended to develop in AI.
The two most important competitive indicators still come from the supply side (computing power resources) and the demand side (computing power utilization rate). Akash has 282 GPUs and more than 20,000 CPUs, and has completed 160,000 leases. The utilization rate of the GPU network is 50-70%, which is a good number in this track. io.net has 40272 GPUs and 5958 CPUs, as well as Render's 4318 GPUs and 159 CPUs, and Filecoin's 1024 GPU licenses, including about 200 H100 blocks and thousands of A100 blocks. Currently, 151,879 inferences have been completed. This time, io.net is using extremely high airdrop expectations to attract computing resources. GPU data is growing rapidly, and its ability to attract resources needs to be re-evaluated after the token is launched. Render and Gensyn did not release specific data. In addition, many projects are improving their competitiveness on the supply and demand side through ecological cooperation. For example, io.net uses the computing power of Render and Filecoin to improve its resource reserves. Render has established a computing client plan (RNP-004). It allows users to indirectly access Render's computing resources through computing clients - io.net, Nosana, FedMl, and Beam, thereby quickly transitioning from the rendering field to artificial intelligence computing.
Also, go to the center Verification of chemical computing remains a problem—how to prove that workers with computing resources performed computing tasks correctly. Gensyn is trying to build such a verification layer to ensure the correctness of calculations through probabilistic learning proofs, graph-based precise positioning protocols, and incentives. The verifiers and whistleblowers jointly check the calculations. Therefore, in addition to decentralized training, Gensyn It provides computing power support, and the verification mechanism it establishes also has unique value. Fluence, the computing protocol located on Solana, also adds verification of computing tasks. Developers can verify that their applications run as expected and that calculations are performed correctly by checking proofs issued by on-chain providers. However, the actual demand is still that "feasible" is greater than "trusted". The computing platform must first have sufficient computing power to have the possibility of competition. Of course, for excellent verification protocols, you can choose to access the computing power of other platforms to become Authentication layer and protocol layer play a unique role.
5.3 Decentralized model
It is still very far away from the ultimate scenario described by Vitalik (shown in the figure below). We are currently unable to create a decentralized model through blockchain and encryption technology. Trustworthy black-box AI is used to solve the problem of adversarial machine learning. Encrypting the entire AI running process from data training to query output is a very large expense. However, there are currently projects that are trying to create better AI models through incentive mechanisms. They first open up the closed state between different models and create a pattern of mutual learning, collaboration and healthy competition between models. Bittensor is the most representative of them. project.
Bittensor: Bittensor is promoting the combination of different AI models, but it is worth noting that Bittensor itself does not train the model, but mainly provides AI Reasoning services. Bittensor's 32 subnets focus on different service directions, such as data capture, text generation, Text2Image, etc. When completing a task, AI models belonging to different directions can collaborate with each other. The incentive mechanism promotes competition between subnets and within subnets. Currently, rewards are issued at a rate of 1 TAO per block. A total of approximately 7,200 TAO tokens are issued daily, and 64 validators in SN0 (root network) According to the performance of the subnet, the distribution ratio of these rewards among different subnets is determined. The subnet validator determines the distribution ratio among different miners by evaluating the work of the miners, thereby achieving better service and performance. Better models receive more incentives, which improves the overall reasoning quality of the system.
6 Conclusion: MEME hype or technological revolution?
From the surge in prices of ARKM and WLD caused by Sam Altman’s movements to the NVIDIA conference leading to a series of participating projects, many people are adjusting their investment philosophy on the AI track. What is the AI track? MEME hype or technological revolution?
Except for a few celebrity topics (such as ARKM and WLD), the AI track as a whole is more like a "MEME dominated by technical narratives."
On the one hand, the overall hype of the Crypto AI track must be closely linked to the progress of Web2 AI. External hype led by OpenAI will become the trigger for the Crypto AI track. On the other hand, the story of the AI track is still dominated by technical narratives. Of course, what we emphasize here is "technical narrative" rather than "technology", which makes the choice of the subdivision direction of the AI track and the fundamentals of the project Attention is still important. We need to find a narrative direction with hype value, and we also need to find projects with mid- to long-term competitiveness and moats.
From the four types of possible combinations proposed by V God, we can see the mutual trade-off between narrative charm and landing possibility. In the first and second categories represented by AI applications, we have seen many GPT Wrapper. The products are launched quickly but the business homogeneity is also high. The first-mover advantage, ecosystem, number of users and product revenue become Stories to tell in homogeneous competition. The third and fourth categories represent the grand narrative of the combination of AI and Crypto, such as Agent on-chain collaboration network, zkML, and decentralized reshaping of AI. They are all in their early stages. Projects with technological innovation will quickly attract funds, even if It's just a very early implementation.