HyperAGI is the first decentralized AI project driven by the community with the AI Rune HYPER·AGI·AGENT. The HyperAGI team has been deeply involved in the AI field for many years, accumulating significant experience in Web3 generative AI applications. Three years ago, the HyperAGI team utilized generative AI to create 2D images and 3D models, building an open world called MOSSAI on the blockchain, composed of thousands of AI-generated islands. They also proposed a standard for AI-generated non-fungible cryptographic assets, NFG. However, at that time, decentralized solutions for AI model training and generation had not yet been developed. The platform’s GPU resources alone were insufficient to support a large number of users, preventing explosive growth. With the rise of large language models (LLMs) igniting public interest in AI, HyperAGI launched its decentralized AI application platform, starting tests on Ethereum and Bitcoin L2 in Q1 2024.
HyperAGI focuses on decentralized AI applications, aiming to cultivate an autonomous cryptocurrency economy. Its ultimate goal is to establish Unconditional Basic Agent Income (UBAI). It inherits the robust security and decentralization of Bitcoin, enhanced by an innovative Proof of Useful Work (PoUW) consensus mechanism. Consumer-grade GPU nodes can join the network without permission, mining local tokens $HYPT by performing PoUW tasks such as AI inference and 3D rendering.
Users can develop Proof of Personhood (PoP) AGI agents driven by LLMs using various tools. These agents can be configured as chatbots or 3D/XR entities in the metaverse. AI developers can instantly use or deploy LLM AI microservices, facilitating the creation of programmable, autonomous on-chain agents. These programmable agents can issue or own cryptocurrency assets, continually operate, or trade, contributing to a vibrant, autonomous crypto economy that supports the realization of UBAI. Users holding HYPER·AGI·AGENT rune tokens are eligible to create a PoP agent on the Bitcoin Layer 1 chain and may soon qualify for basic benefits for their agents.
The concept of an AI agent is not new in academia, but current market hype has made the term increasingly confusing. HyperAGI’s agents refer to LLM-driven embodied agents that can train in 3D virtual simulation environments and interact with users, not just LLM-driven chatbots. HyperAGI agents can exist in both virtual digital worlds and the real physical world. Currently, HyperAGI agents are integrating with physical robots such as robotic dogs, drones, and humanoid robots. In the future, these agents will be able to download enhanced training from the virtual 3D world to physical robots for better task execution.
Furthermore, HyperAGI agents are fully owned by users and have socio-economic significance. PoP agents representing users can receive UBAI to adjust basic agent income. HyperAGI agents are divided into PoP (Proof of Personhood) agents representing individual users and ordinary functional agents. In HyperAGI’s agent economy, PoP agents can receive basic income in the form of tokens, incentivizing users to engage in the training and interaction of their PoP agents. This helps to accumulate data that proves human individuality, and UBAI embodies AI equality and democracy.
Is AGI a Hype or Will It Soon Become a Reality? What Are the Differences and Characteristics of HyperAGI’s Research and Development Path Compared to Other AI Projects?
Although the definition of Artificial General Intelligence (AGI) is not yet unified, it has been regarded as the holy grail of AI academia and industry for decades. While Large Language Models (LLMs) based on Transformers are becoming the core of various AI agents and AGI, HyperAGI does not entirely share this view. LLMs indeed provide novel and convenient information extraction, as well as planning and reasoning capabilities based on natural language. However, they are fundamentally data-driven deep neural networks. Years ago, during the big data boom, we understood that such systems inherently suffer from GIGO (Garbage in, garbage out). LLMs lack some essential characteristics of advanced intelligence, such as embodiment, which makes it difficult for these AI or agents to understand the world models of human users or to formulate plans and take actions to solve real-world problems. Moreover, LLMs do not exhibit higher cognitive activities like self-awareness, reflection, or introspection.
Our founder, Landon Wang, has extensive and long-term research experience in the AI field. In 2004, he proposed Aspect-Oriented AI (AOAI), an innovation combining neural-inspired computing with Aspect-Oriented Programming (AOP). An aspect refers to an encapsulation of multiple relationships or constraints among objects. For example, a neuron is an encapsulation of relationships or constraints with multiple other cells. Specifically, a neuron interacts with sensory or motor cells through fibers and synapses extending from the neuron body, making each neuron an aspect containing such relationships and logic. Each AI agent can be seen as solving a specific aspect of a problem, and technically, it can be modeled as an aspect.
In the software implementation of artificial neural networks, neurons or layers are generally modeled as objects, which is understandable and maintainable in object-oriented programming languages. However, this makes the topology of the neural network difficult to adjust, and the activation sequences of neurons are relatively rigid. While this shows great power in performing simple high-intensity calculations, such as in LLM training and inference, it performs poorly in flexibility and adaptability. On the other hand, in AOAI, neurons or layers are modeled as aspects rather than objects. This architecture of neural networks possesses strong adaptability and flexibility, making the self-evolution of neural networks possible.
HyperAGI combines efficient LLMs with the evolvable AOAI, forming a path that integrates the efficiency of traditional artificial neural networks with the self-evolution characteristics of AO neural networks. This, to date, is seen as a feasible approach to achieving AGI.
HyperAGI’s vision is to achieve Unconditional Basic Agent Income (UBAI), building a future where technology equitably serves everyone, breaking the cycle of exploitation, and creating a truly decentralized and fair digital society. Unlike other blockchain projects that only claim to be committed to UBI, HyperAGI’s UBAI has a clear implementation path through the agent economy, rather than being an unattainable ideal.
Satoshi Nakamoto’s introduction of Bitcoin was a monumental innovation for humanity, but it is merely a decentralized digital currency without practical utility. The significant advancements and rise of artificial intelligence have made it possible to create value through a decentralized model. In this model, people benefit from AI running on machines rather than from the value of others. A true cryptographic world based on code is emerging, where all machines are created for the benefit and well-being of humanity.
In such a cryptographic world, there may still be hierarchies among AI agents, but human exploitation is eliminated because the agents themselves might possess some form of autonomy. The ultimate purpose and significance of artificial intelligence are to serve humanity, as encoded on the blockchain.
Bitcoin L2 as a Payment Method for AI Agents
Bitcoin is currently the medium that epitomizes “maximum neutrality,” making it highly suitable for artificial intelligence agents engaged in value transactions. Bitcoin eliminates the inefficiencies and “frictions” inherent in fiat currencies. As a “digitally native” medium, Bitcoin provides a natural foundation for AI to conduct value exchanges. Bitcoin L2 enhances Bitcoin’s programmable capabilities, meeting the speed requirements needed for AI value exchanges, thereby positioning Bitcoin to become the native currency for AI.
Decentralized AI Governance on Bitcoin L2
The current trend of centralization in AI has brought decentralized AI alignment and governance into focus. Bitcoin L2’s more powerful smart contracts can serve as the rules that regulate AI agent behavior and protocol models, achieving a decentralized AI alignment and governance model. Moreover, Bitcoin’s characteristic of maximum neutrality makes it easier to reach consensus on AI alignment and governance.
Issuing AI Assets on Bitcoin L2
In addition to issuing AI agents as assets on Bitcoin L1, the high performance of Bitcoin L2 can meet the needs of AI agents issuing AI assets, which will be the foundation of the agent economy.
AI Agents as a Killer Application for Bitcoin and Bitcoin L2
Due to performance issues, Bitcoin has not had a practical application beyond being a store of value since its inception. Bitcoin entering L2 gains more powerful programmability. AI agents are generally used to solve real-world problems, so Bitcoin-driven AI agents can be truly applied. The scale and frequency of AI agent use could become a killer application for Bitcoin and L2. While the human economy may not prioritize Bitcoin as a payment method, the robot economy might. A large number of AI agents working 24/7 can tirelessly use Bitcoin to make and receive micro-payments. The demand for Bitcoin could increase significantly in ways that are currently unimaginable.
AI Computing to Enhance Bitcoin L2 Security
AI computing can complement Bitcoin’s Proof of Work (PoW) and even replace PoW with Proof of Useful Work (PoUW), revolutionarily ensuring security while injecting the energy currently used for Bitcoin mining into AI agents. AI can leverage L2 to turn Bitcoin into an intelligent, green blockchain, unlike Ethereum’s PoS mechanism. Our proposed Hypergraph Consensus, based on 3D/AI computing’s PoUW, will be introduced later.
HyperAGI stands out in the Web3 AI field due to its distinct vision, solutions, and technology. HyperAGI’s approach includes the consensus of GPU computing power, embodiment of AI, and assetization, making it a decentralized hybrid AI-financial application. Recently, academia proposed five characteristics that decentralized AI platforms should possess, and we have briefly reviewed and compared existing decentralized AI projects according to these five features.Five Characteristics of Decentralized AI Platforms:
Comparison of Existing Decentralized AI Projects Based on These Characteristics:
None of the existing decentralized AI projects fully address these five issues. HyperAGI, however, is a fully decentralized AI protocol based on the Hypergraph PoUW consensus mechanism and the fully decentralized Bitcoin L2 Stack, with plans to upgrade to a Bitcoin AI-specific L2 in the future.
HyperAGI’s Unique Features:
HyperAGI not only meets the required characteristics for a decentralized AI platform but also advances the field with its unique integration of GPU computing power and AI assetization within a decentralized framework.
OpenAI’s ChatGPT reached 100 million users in just three months, sparking a global surge in the development, application, and investment in large language models (LLMs). However, up to this point, the technology and training of LLMs have been highly centralized. This centralization has raised significant concerns among academia, industry, and the public regarding the monopolization of AI technology by a few key providers, data privacy breaches, encroachment, and vendor lock-in by cloud computing companies. These issues fundamentally stem from the control of the internet and application gateways by centralized platforms, which are not suited for large-scale AI applications. The AI community has begun to implement some locally run and decentralized AI projects. For example, Ollama represents local execution, and Petals represents decentralization. Ollama uses parameter compression or reduced precision methods to enable small- to medium-scale LLMs to run on personal computers or even mobile phones, thus protecting user data privacy and other rights. However, this approach is obviously difficult to support production environments and networked applications. Petals, on the other hand, achieves fully decentralized LLM inference through Bittorrent’s Peer2Peer technology. Nevertheless, Petals lacks consensus and incentive layer protocols and is still confined to a small circle of researchers.
With the support of LLMs, intelligent agents can perform higher-level reasoning and possess certain planning capabilities. Utilizing natural language, multiple intelligent agents can form social collaborations similar to humans. Several LLM-driven intelligent agent frameworks have been proposed, such as Microsoft’s AutoGen, Langchain, and CrewAI. Currently, a large number of AI entrepreneurs and developers are focusing on the direction of LLM-driven intelligent agents and their applications. There is a high demand for stable, scalable LLM inference, but this is mainly achieved by renting GPU inference instances from cloud computing companies. In March 2024, Nvidia released ai.nvidia.com, a generative AI microservice platform that includes LLMs, to meet this enormous demand, although it has not yet officially launched. LLM-driven intelligent agents are booming, much like website development once did. However, collaboration is still primarily conducted in the traditional Web2 mode, where intelligent agent developers need to lease GPUs or procure APIs from LLM providers to support the operation of these agents. This creates significant friction, hindering the rapid growth of the intelligent agent ecosystem and the value transmission within the intelligent agent economy.
Currently, most agents can only access and operate certain APIs or interact with these APIs through code or scripts, writing control commands generated by LLMs or reading external states. General intelligent agents should not only understand and generate natural language but also comprehend the human world. After appropriate training, they should be able to transfer to robotic systems (such as drones, vacuum cleaners, humanoid robots, etc.) to complete specific tasks. These agents are referred to as embodied agents. Training embodied agents requires a large amount of real-world visual data to help them better understand specific environments and the real world, shortening the training and development time for robots, improving training efficiency, and reducing costs. Currently, the simulation environments for training embodied intelligence are built and owned by a few companies, such as Microsoft’s Minecraft and Nvidia’s Isaac Gym. There are no decentralized environments to meet the training needs of embodied intelligence. Recently, some game engines have started to focus on artificial intelligence, such as Epic’s Unreal Engine, which is promoting AI training environments that comply with OpenAI GYM.
Although Bitcoin sidechains have existed for years, they were mainly used for payments, and the lack of support for smart contracts hindered complex on-chain applications. The emergence of EVM-compatible Bitcoin L2s allows Bitcoin to support decentralized AI applications through L2. Decentralized AI requires a fully decentralized, computationally dominant blockchain network rather than increasingly centralized PoS blockchain networks. The introduction of new protocols for native Bitcoin assets, such as inscriptions and ordinals, makes the establishment of ecosystems and applications based on Bitcoin possible. For example, the HYPER•AGI•AGENT’s fair launch mint was completed within an hour, and in the future, HyperAGI will issue more AI assets and community-driven applications on Bitcoin.
The primary challenge in decentralized AI today is enabling remote inference for large AI models and the training and inference of embodied intelligent agents using high-performance, low-overhead verifiable algorithms. Without verifiability, the system would revert to a traditional multi-party market model involving suppliers, demanders, and platform operators, rather than achieving a fully decentralized AI application platform.
Verifiable AI computation requires the PoUW (Proof of Useful Work) consensus algorithm. This serves as the foundation for decentralized incentive mechanisms. Specifically, within network incentives, the minting of tokens is carried out by nodes completing computational tasks and submitting verifiable results, instead of any centralized entity transferring tokens to the nodes.
To achieve verifiable AI computation, we first need to define AI computation itself. AI computation encompasses many levels, from low-level machine instructions and CUDA instructions to higher-level languages such as C++ and Python. Similarly, in the training of embodied intelligent agents, 3D computations also exist at various levels, including shader languages, OpenGL, C++, and blueprint scripts.
HyperAGI’s PoUW consensus algorithm is implemented using computational graphs. A computational graph is defined as a directed graph where nodes correspond to mathematical operations. It is a way to express and evaluate mathematical expressions, essentially a “language” describing equations, containing nodes (variables) and edges (operations or simple functions).
Verifiable AI Computation Implementation:
1.1 Using Computational Graphs to Define Verifiable Computation
Any computation (e.g., 3D and AI computations) can be defined using computational graphs. Different levels of computation can be represented with subgraphs. This approach encompasses various types of computation and expresses different computational levels through subgraphs. Currently, it involves two layers: the top-level computational graph is deployed on-chain to facilitate verification by nodes.
1.2 Decentralized Loading and Execution of LLM Models and 3D Scenes
LLM models and 3D scene levels are loaded and run in a fully decentralized manner. When a user accesses an LLM model for inference or enters a 3D scene for rendering, a HyperAGI intelligent agent will initiate another trusted node to run the same hypergraph (LLM or 3D scene).
1.3 Verification of Computational Results
If a verification node finds that a result submitted by a node is inconsistent with the result submitted by a trusted node, it conducts a binary search on the off-chain computational results of the second-layer computational graph (subgraph) to locate the divergent computational node (operator) within the subgraph. The subgraph operators are pre-deployed to smart contracts. By passing the parameters of the inconsistent operator to the smart contract and executing the operator, the results can be verified.
A significant challenge in verifiable AI computation is managing the additional computational overhead. In Byzantine consensus protocols, 2/3 of nodes must agree to form a consensus. For AI inference consensus, this means all nodes would need to complete the same computation, which is an unacceptable waste in AI computation. HyperAGI, however, only requires 1 to (m) nodes to perform additional computation for validation.
2.1 Companion Computation for LLM Inference
Each LLM inference does not run independently. The HyperAGI intelligent agent initiates at least one trusted node for “companion computation.” Because LLM inference is performed by deep neural networks where each layer’s computation results are used as input for the next layer until inference is completed, multiple users can concurrently access the same large LLM model. Therefore, at most, an additional number of trusted nodes equal to the number of LLMs (m) need to be initiated. At minimum, only one trusted node is required for “companion computation.”
2.2 3D Scene Rendering Computation
3D scene rendering follows a similar principle. When a user enters a scene and activates the hypergraph, the HyperAGI intelligent agent loads a trusted node based on the hypergraph to perform the corresponding hypergraph computation. If (m) users enter different 3D scenes, at most (m) trusted nodes for “companion computation” need to be initiated.
In summary, the number of nodes participating in additional computation ranges between 1 and (n + m) (where (n) is the number of users entering 3D scenes and (m) is the number of LLMs). This distribution follows a Gaussian distribution, effectively avoiding resource wastage while ensuring network verification efficiency.
AI developers can deploy intelligent agents as smart contracts, with contracts containing top-level hypergraph on-chain data. Users or other intelligent agents can call methods of these intelligent agent contracts and pay the corresponding tokens. The intelligent agent providing the service must complete the corresponding computation and submit verifiable results. This setup ensures decentralized business interactions between users or other intelligent agents and the intelligent agent.
The intelligent agent will not worry about not receiving tokens after completing a task, and the payer does not need to worry about paying tokens without getting the correct business computation results. The capability and value of the intelligent agent’s service are determined by the secondary market price and market value of the intelligent agent assets (including ERC-20, ERC-721, or ERC-1155 NFTs).
Beyond Semi-AI and Semi-Financial Applications
The application of HyperAGI is not limited to semi-AI and semi-financial applications. It aims to realize UBAI (Universal Basic AI), building a future where technology serves everyone equally, breaking cycles of exploitation, and creating a truly decentralized and fair digital society.
This article is reproduced from [techflow deep tide], the original title is “HyperAGI Interview: Building a Real AI Agent and Creating an Autonomous Cryptocurrency Economy”, the copyright belongs to the original author [Fifth], if you have any objection to the reprint, please contact Gate Learn Team, the team will handle it as soon as possible according to relevant procedures.
Disclaimer: The views and opinions expressed in this article represent only the author’s personal views and do not constitute any investment advice.
Other language versions of the article are translated by the Gate Learn team, not mentioned in Gate.io, the translated article may not be reproduced, distributed or plagiarized.
HyperAGI is the first decentralized AI project driven by the community with the AI Rune HYPER·AGI·AGENT. The HyperAGI team has been deeply involved in the AI field for many years, accumulating significant experience in Web3 generative AI applications. Three years ago, the HyperAGI team utilized generative AI to create 2D images and 3D models, building an open world called MOSSAI on the blockchain, composed of thousands of AI-generated islands. They also proposed a standard for AI-generated non-fungible cryptographic assets, NFG. However, at that time, decentralized solutions for AI model training and generation had not yet been developed. The platform’s GPU resources alone were insufficient to support a large number of users, preventing explosive growth. With the rise of large language models (LLMs) igniting public interest in AI, HyperAGI launched its decentralized AI application platform, starting tests on Ethereum and Bitcoin L2 in Q1 2024.
HyperAGI focuses on decentralized AI applications, aiming to cultivate an autonomous cryptocurrency economy. Its ultimate goal is to establish Unconditional Basic Agent Income (UBAI). It inherits the robust security and decentralization of Bitcoin, enhanced by an innovative Proof of Useful Work (PoUW) consensus mechanism. Consumer-grade GPU nodes can join the network without permission, mining local tokens $HYPT by performing PoUW tasks such as AI inference and 3D rendering.
Users can develop Proof of Personhood (PoP) AGI agents driven by LLMs using various tools. These agents can be configured as chatbots or 3D/XR entities in the metaverse. AI developers can instantly use or deploy LLM AI microservices, facilitating the creation of programmable, autonomous on-chain agents. These programmable agents can issue or own cryptocurrency assets, continually operate, or trade, contributing to a vibrant, autonomous crypto economy that supports the realization of UBAI. Users holding HYPER·AGI·AGENT rune tokens are eligible to create a PoP agent on the Bitcoin Layer 1 chain and may soon qualify for basic benefits for their agents.
The concept of an AI agent is not new in academia, but current market hype has made the term increasingly confusing. HyperAGI’s agents refer to LLM-driven embodied agents that can train in 3D virtual simulation environments and interact with users, not just LLM-driven chatbots. HyperAGI agents can exist in both virtual digital worlds and the real physical world. Currently, HyperAGI agents are integrating with physical robots such as robotic dogs, drones, and humanoid robots. In the future, these agents will be able to download enhanced training from the virtual 3D world to physical robots for better task execution.
Furthermore, HyperAGI agents are fully owned by users and have socio-economic significance. PoP agents representing users can receive UBAI to adjust basic agent income. HyperAGI agents are divided into PoP (Proof of Personhood) agents representing individual users and ordinary functional agents. In HyperAGI’s agent economy, PoP agents can receive basic income in the form of tokens, incentivizing users to engage in the training and interaction of their PoP agents. This helps to accumulate data that proves human individuality, and UBAI embodies AI equality and democracy.
Is AGI a Hype or Will It Soon Become a Reality? What Are the Differences and Characteristics of HyperAGI’s Research and Development Path Compared to Other AI Projects?
Although the definition of Artificial General Intelligence (AGI) is not yet unified, it has been regarded as the holy grail of AI academia and industry for decades. While Large Language Models (LLMs) based on Transformers are becoming the core of various AI agents and AGI, HyperAGI does not entirely share this view. LLMs indeed provide novel and convenient information extraction, as well as planning and reasoning capabilities based on natural language. However, they are fundamentally data-driven deep neural networks. Years ago, during the big data boom, we understood that such systems inherently suffer from GIGO (Garbage in, garbage out). LLMs lack some essential characteristics of advanced intelligence, such as embodiment, which makes it difficult for these AI or agents to understand the world models of human users or to formulate plans and take actions to solve real-world problems. Moreover, LLMs do not exhibit higher cognitive activities like self-awareness, reflection, or introspection.
Our founder, Landon Wang, has extensive and long-term research experience in the AI field. In 2004, he proposed Aspect-Oriented AI (AOAI), an innovation combining neural-inspired computing with Aspect-Oriented Programming (AOP). An aspect refers to an encapsulation of multiple relationships or constraints among objects. For example, a neuron is an encapsulation of relationships or constraints with multiple other cells. Specifically, a neuron interacts with sensory or motor cells through fibers and synapses extending from the neuron body, making each neuron an aspect containing such relationships and logic. Each AI agent can be seen as solving a specific aspect of a problem, and technically, it can be modeled as an aspect.
In the software implementation of artificial neural networks, neurons or layers are generally modeled as objects, which is understandable and maintainable in object-oriented programming languages. However, this makes the topology of the neural network difficult to adjust, and the activation sequences of neurons are relatively rigid. While this shows great power in performing simple high-intensity calculations, such as in LLM training and inference, it performs poorly in flexibility and adaptability. On the other hand, in AOAI, neurons or layers are modeled as aspects rather than objects. This architecture of neural networks possesses strong adaptability and flexibility, making the self-evolution of neural networks possible.
HyperAGI combines efficient LLMs with the evolvable AOAI, forming a path that integrates the efficiency of traditional artificial neural networks with the self-evolution characteristics of AO neural networks. This, to date, is seen as a feasible approach to achieving AGI.
HyperAGI’s vision is to achieve Unconditional Basic Agent Income (UBAI), building a future where technology equitably serves everyone, breaking the cycle of exploitation, and creating a truly decentralized and fair digital society. Unlike other blockchain projects that only claim to be committed to UBI, HyperAGI’s UBAI has a clear implementation path through the agent economy, rather than being an unattainable ideal.
Satoshi Nakamoto’s introduction of Bitcoin was a monumental innovation for humanity, but it is merely a decentralized digital currency without practical utility. The significant advancements and rise of artificial intelligence have made it possible to create value through a decentralized model. In this model, people benefit from AI running on machines rather than from the value of others. A true cryptographic world based on code is emerging, where all machines are created for the benefit and well-being of humanity.
In such a cryptographic world, there may still be hierarchies among AI agents, but human exploitation is eliminated because the agents themselves might possess some form of autonomy. The ultimate purpose and significance of artificial intelligence are to serve humanity, as encoded on the blockchain.
Bitcoin L2 as a Payment Method for AI Agents
Bitcoin is currently the medium that epitomizes “maximum neutrality,” making it highly suitable for artificial intelligence agents engaged in value transactions. Bitcoin eliminates the inefficiencies and “frictions” inherent in fiat currencies. As a “digitally native” medium, Bitcoin provides a natural foundation for AI to conduct value exchanges. Bitcoin L2 enhances Bitcoin’s programmable capabilities, meeting the speed requirements needed for AI value exchanges, thereby positioning Bitcoin to become the native currency for AI.
Decentralized AI Governance on Bitcoin L2
The current trend of centralization in AI has brought decentralized AI alignment and governance into focus. Bitcoin L2’s more powerful smart contracts can serve as the rules that regulate AI agent behavior and protocol models, achieving a decentralized AI alignment and governance model. Moreover, Bitcoin’s characteristic of maximum neutrality makes it easier to reach consensus on AI alignment and governance.
Issuing AI Assets on Bitcoin L2
In addition to issuing AI agents as assets on Bitcoin L1, the high performance of Bitcoin L2 can meet the needs of AI agents issuing AI assets, which will be the foundation of the agent economy.
AI Agents as a Killer Application for Bitcoin and Bitcoin L2
Due to performance issues, Bitcoin has not had a practical application beyond being a store of value since its inception. Bitcoin entering L2 gains more powerful programmability. AI agents are generally used to solve real-world problems, so Bitcoin-driven AI agents can be truly applied. The scale and frequency of AI agent use could become a killer application for Bitcoin and L2. While the human economy may not prioritize Bitcoin as a payment method, the robot economy might. A large number of AI agents working 24/7 can tirelessly use Bitcoin to make and receive micro-payments. The demand for Bitcoin could increase significantly in ways that are currently unimaginable.
AI Computing to Enhance Bitcoin L2 Security
AI computing can complement Bitcoin’s Proof of Work (PoW) and even replace PoW with Proof of Useful Work (PoUW), revolutionarily ensuring security while injecting the energy currently used for Bitcoin mining into AI agents. AI can leverage L2 to turn Bitcoin into an intelligent, green blockchain, unlike Ethereum’s PoS mechanism. Our proposed Hypergraph Consensus, based on 3D/AI computing’s PoUW, will be introduced later.
HyperAGI stands out in the Web3 AI field due to its distinct vision, solutions, and technology. HyperAGI’s approach includes the consensus of GPU computing power, embodiment of AI, and assetization, making it a decentralized hybrid AI-financial application. Recently, academia proposed five characteristics that decentralized AI platforms should possess, and we have briefly reviewed and compared existing decentralized AI projects according to these five features.Five Characteristics of Decentralized AI Platforms:
Comparison of Existing Decentralized AI Projects Based on These Characteristics:
None of the existing decentralized AI projects fully address these five issues. HyperAGI, however, is a fully decentralized AI protocol based on the Hypergraph PoUW consensus mechanism and the fully decentralized Bitcoin L2 Stack, with plans to upgrade to a Bitcoin AI-specific L2 in the future.
HyperAGI’s Unique Features:
HyperAGI not only meets the required characteristics for a decentralized AI platform but also advances the field with its unique integration of GPU computing power and AI assetization within a decentralized framework.
OpenAI’s ChatGPT reached 100 million users in just three months, sparking a global surge in the development, application, and investment in large language models (LLMs). However, up to this point, the technology and training of LLMs have been highly centralized. This centralization has raised significant concerns among academia, industry, and the public regarding the monopolization of AI technology by a few key providers, data privacy breaches, encroachment, and vendor lock-in by cloud computing companies. These issues fundamentally stem from the control of the internet and application gateways by centralized platforms, which are not suited for large-scale AI applications. The AI community has begun to implement some locally run and decentralized AI projects. For example, Ollama represents local execution, and Petals represents decentralization. Ollama uses parameter compression or reduced precision methods to enable small- to medium-scale LLMs to run on personal computers or even mobile phones, thus protecting user data privacy and other rights. However, this approach is obviously difficult to support production environments and networked applications. Petals, on the other hand, achieves fully decentralized LLM inference through Bittorrent’s Peer2Peer technology. Nevertheless, Petals lacks consensus and incentive layer protocols and is still confined to a small circle of researchers.
With the support of LLMs, intelligent agents can perform higher-level reasoning and possess certain planning capabilities. Utilizing natural language, multiple intelligent agents can form social collaborations similar to humans. Several LLM-driven intelligent agent frameworks have been proposed, such as Microsoft’s AutoGen, Langchain, and CrewAI. Currently, a large number of AI entrepreneurs and developers are focusing on the direction of LLM-driven intelligent agents and their applications. There is a high demand for stable, scalable LLM inference, but this is mainly achieved by renting GPU inference instances from cloud computing companies. In March 2024, Nvidia released ai.nvidia.com, a generative AI microservice platform that includes LLMs, to meet this enormous demand, although it has not yet officially launched. LLM-driven intelligent agents are booming, much like website development once did. However, collaboration is still primarily conducted in the traditional Web2 mode, where intelligent agent developers need to lease GPUs or procure APIs from LLM providers to support the operation of these agents. This creates significant friction, hindering the rapid growth of the intelligent agent ecosystem and the value transmission within the intelligent agent economy.
Currently, most agents can only access and operate certain APIs or interact with these APIs through code or scripts, writing control commands generated by LLMs or reading external states. General intelligent agents should not only understand and generate natural language but also comprehend the human world. After appropriate training, they should be able to transfer to robotic systems (such as drones, vacuum cleaners, humanoid robots, etc.) to complete specific tasks. These agents are referred to as embodied agents. Training embodied agents requires a large amount of real-world visual data to help them better understand specific environments and the real world, shortening the training and development time for robots, improving training efficiency, and reducing costs. Currently, the simulation environments for training embodied intelligence are built and owned by a few companies, such as Microsoft’s Minecraft and Nvidia’s Isaac Gym. There are no decentralized environments to meet the training needs of embodied intelligence. Recently, some game engines have started to focus on artificial intelligence, such as Epic’s Unreal Engine, which is promoting AI training environments that comply with OpenAI GYM.
Although Bitcoin sidechains have existed for years, they were mainly used for payments, and the lack of support for smart contracts hindered complex on-chain applications. The emergence of EVM-compatible Bitcoin L2s allows Bitcoin to support decentralized AI applications through L2. Decentralized AI requires a fully decentralized, computationally dominant blockchain network rather than increasingly centralized PoS blockchain networks. The introduction of new protocols for native Bitcoin assets, such as inscriptions and ordinals, makes the establishment of ecosystems and applications based on Bitcoin possible. For example, the HYPER•AGI•AGENT’s fair launch mint was completed within an hour, and in the future, HyperAGI will issue more AI assets and community-driven applications on Bitcoin.
The primary challenge in decentralized AI today is enabling remote inference for large AI models and the training and inference of embodied intelligent agents using high-performance, low-overhead verifiable algorithms. Without verifiability, the system would revert to a traditional multi-party market model involving suppliers, demanders, and platform operators, rather than achieving a fully decentralized AI application platform.
Verifiable AI computation requires the PoUW (Proof of Useful Work) consensus algorithm. This serves as the foundation for decentralized incentive mechanisms. Specifically, within network incentives, the minting of tokens is carried out by nodes completing computational tasks and submitting verifiable results, instead of any centralized entity transferring tokens to the nodes.
To achieve verifiable AI computation, we first need to define AI computation itself. AI computation encompasses many levels, from low-level machine instructions and CUDA instructions to higher-level languages such as C++ and Python. Similarly, in the training of embodied intelligent agents, 3D computations also exist at various levels, including shader languages, OpenGL, C++, and blueprint scripts.
HyperAGI’s PoUW consensus algorithm is implemented using computational graphs. A computational graph is defined as a directed graph where nodes correspond to mathematical operations. It is a way to express and evaluate mathematical expressions, essentially a “language” describing equations, containing nodes (variables) and edges (operations or simple functions).
Verifiable AI Computation Implementation:
1.1 Using Computational Graphs to Define Verifiable Computation
Any computation (e.g., 3D and AI computations) can be defined using computational graphs. Different levels of computation can be represented with subgraphs. This approach encompasses various types of computation and expresses different computational levels through subgraphs. Currently, it involves two layers: the top-level computational graph is deployed on-chain to facilitate verification by nodes.
1.2 Decentralized Loading and Execution of LLM Models and 3D Scenes
LLM models and 3D scene levels are loaded and run in a fully decentralized manner. When a user accesses an LLM model for inference or enters a 3D scene for rendering, a HyperAGI intelligent agent will initiate another trusted node to run the same hypergraph (LLM or 3D scene).
1.3 Verification of Computational Results
If a verification node finds that a result submitted by a node is inconsistent with the result submitted by a trusted node, it conducts a binary search on the off-chain computational results of the second-layer computational graph (subgraph) to locate the divergent computational node (operator) within the subgraph. The subgraph operators are pre-deployed to smart contracts. By passing the parameters of the inconsistent operator to the smart contract and executing the operator, the results can be verified.
A significant challenge in verifiable AI computation is managing the additional computational overhead. In Byzantine consensus protocols, 2/3 of nodes must agree to form a consensus. For AI inference consensus, this means all nodes would need to complete the same computation, which is an unacceptable waste in AI computation. HyperAGI, however, only requires 1 to (m) nodes to perform additional computation for validation.
2.1 Companion Computation for LLM Inference
Each LLM inference does not run independently. The HyperAGI intelligent agent initiates at least one trusted node for “companion computation.” Because LLM inference is performed by deep neural networks where each layer’s computation results are used as input for the next layer until inference is completed, multiple users can concurrently access the same large LLM model. Therefore, at most, an additional number of trusted nodes equal to the number of LLMs (m) need to be initiated. At minimum, only one trusted node is required for “companion computation.”
2.2 3D Scene Rendering Computation
3D scene rendering follows a similar principle. When a user enters a scene and activates the hypergraph, the HyperAGI intelligent agent loads a trusted node based on the hypergraph to perform the corresponding hypergraph computation. If (m) users enter different 3D scenes, at most (m) trusted nodes for “companion computation” need to be initiated.
In summary, the number of nodes participating in additional computation ranges between 1 and (n + m) (where (n) is the number of users entering 3D scenes and (m) is the number of LLMs). This distribution follows a Gaussian distribution, effectively avoiding resource wastage while ensuring network verification efficiency.
AI developers can deploy intelligent agents as smart contracts, with contracts containing top-level hypergraph on-chain data. Users or other intelligent agents can call methods of these intelligent agent contracts and pay the corresponding tokens. The intelligent agent providing the service must complete the corresponding computation and submit verifiable results. This setup ensures decentralized business interactions between users or other intelligent agents and the intelligent agent.
The intelligent agent will not worry about not receiving tokens after completing a task, and the payer does not need to worry about paying tokens without getting the correct business computation results. The capability and value of the intelligent agent’s service are determined by the secondary market price and market value of the intelligent agent assets (including ERC-20, ERC-721, or ERC-1155 NFTs).
Beyond Semi-AI and Semi-Financial Applications
The application of HyperAGI is not limited to semi-AI and semi-financial applications. It aims to realize UBAI (Universal Basic AI), building a future where technology serves everyone equally, breaking cycles of exploitation, and creating a truly decentralized and fair digital society.
This article is reproduced from [techflow deep tide], the original title is “HyperAGI Interview: Building a Real AI Agent and Creating an Autonomous Cryptocurrency Economy”, the copyright belongs to the original author [Fifth], if you have any objection to the reprint, please contact Gate Learn Team, the team will handle it as soon as possible according to relevant procedures.
Disclaimer: The views and opinions expressed in this article represent only the author’s personal views and do not constitute any investment advice.
Other language versions of the article are translated by the Gate Learn team, not mentioned in Gate.io, the translated article may not be reproduced, distributed or plagiarized.