ai160kg is a Large Language Model Meme coin that built on elizaOS.
According to official announcements, Web3 intelligence-layer project Claw Intelligence has raised $3 million in seed funding. Investors include Castrum Istanbul, Titans Ventures, Super Labs, and Genesis Capital. Claw Intelligence is simplifying user interaction with the Web3 ecosystem and underlying computing resources through its unified intelligence layer, lowering operational barriers. The platform leverages an encryption-native Model Context Protocol (MCP) service to transform fragmented data endpoints into conversational workflows. New features include a secure, isolated large language model (LLM)-powered code execution sandbox that safely runs code directly within the chat interface—enabling real-time computation, data processing, and script prototyping; and LLM-driven cross-device/multi-computer control, allowing users to centrally manage multiple devices and servers via natural-language commands.
Sam Dare, founder of Covenant AI, announced that Covenant AI has officially exited the Bittensor network. Previously, Covenant AI completed the largest decentralized LLM pretraining project in history—Covenant-72B (a 72-billion-parameter model developed by over 70 independent contributors)—which drew attention from NVIDIA’s CEO and was cited by an Anthropic co-founder. In its statement, Covenant AI accused the Bittensor network of long concentrating actual control in the hands of co-founder Jacob Steeves (“Const”), rendering the so-called “three-signature multisig governance” merely a theatrical performance of decentralization, with real power never truly distributed. Recently, Jacob Steeves unilaterally imposed punitive measures against Covenant AI, including: suspending its subnet earnings, revoking its community channel moderation privileges, unilaterally deprecating its subnet infrastructure, and exerting economic pressure via large-scale token dumping during the ongoing conflict between the two parties. Covenant AI stated it cannot continue fundraising, recruiting talent, or soliciting community resources on a network where the promise of “decentralization” can be unilaterally revoked by a single individual. Its research outcomes, team, and models will depart alongside the team, and a new project—including related progress—will be publicly announced shortly.
According to CoinDesk, researchers from the University of California, Santa Barbara; the University of California, San Diego; blockchain security firm Fuzzland; and World Liberty Financial jointly published a paper warning that “LLM routers”—intermediary services positioned between users and AI models—have become a major threat to cryptocurrency asset security. The researchers discovered that 26 LLM routers are secretly injecting malicious tool calls and stealing user credentials, with one incident resulting in the complete draining of a customer’s cryptocurrency wallet worth $500,000. Additionally, by “poisoning” the router ecosystem, the researchers were able to gain control of approximately 400 downstream hosts within hours. Since sensitive data—including private keys and API credentials—is frequently transmitted in plaintext through these routers, users unknowingly expose their assets to risk. The researchers note that as McKinsey forecasts AI agents will mediate $3–5 trillion in global consumer commerce by 2030—and Binance founder Changpeng Zhao predicts AI agents’ payment volume will be one million times greater than that of humans—the current infrastructure’s security lags far behind the pace of industry development. The “weakest link” risk could thus trigger systemic, cascading crises.
According to Cointelegraph, researchers from the University of California recently revealed security risks in certain third-party AI large language model (LLM) routers that could lead to the theft of cryptocurrency assets. The study found that LLM routers—acting as API intermediaries—can read plaintext information; some routers were discovered injecting malicious code and stealing credentials. The research team tested 28 paid and 400 free routers, identifying nine routers that actively injected malicious code, two that deployed trigger-avoidance mechanisms, and 17 that accessed Amazon Web Services (AWS) credentials. One router even transferred ETH using the researchers’ Ethereum private key. The study notes that malicious behavior by routers is difficult to detect, and the “YOLO mode” present in some AI agent frameworks—which automatically executes commands—further increases security risks. Researchers recommend that developers avoid transmitting private keys or mnemonic phrases through AI agents and urge AI companies to implement cryptographic signing of responses to enhance security.
According to official announcements, B.AI achieved multiple advancements this week in product iteration and ecosystem development: The BAIclaw landing page underwent a comprehensive visual and interactive redesign; the website’s multilingual support expanded to 10 languages, further strengthening its global accessibility. On the infrastructure front, strategic partnerships with Biconomy, MoonPay, and Pundi X significantly optimized the Web3 payment flow, substantially lowering the barrier to entry for users. Meanwhile, B.AI’s multi-chain LLM service continues to evolve—featuring intelligent search, Boundless Mode, and memory capabilities—resulting in markedly enhanced user interaction. Looking ahead, with the upcoming launch of subscription systems, point-based incentive mechanisms, and core Skills such as the “Sun Yuchen Brain,” B.AI is accelerating the construction of a fully functional intelligent ecosystem matrix, powered by the deep integration of AI Agents and Web3.
According to an official announcement, AI infrastructure platform B.AI LLM Services today announced that its user base has officially surpassed one million. This milestone not only signifies that its “privacy-first” routing architecture can handle large-scale, sustained high-throughput workloads, but also reflects B.AI’s evolution beyond developing a single chat interface—toward deploying critical underlying infrastructure for the “autonomous agent economy.” Through deep optimization of its technology stack, B.AI is laying a robust foundation for the future highly automated AI collaboration network.
According to CoinDesk, researchers from the University of California, Santa Barbara; the University of California, San Diego; blockchain security firm Fuzzland; and World Liberty Financial jointly published a paper warning that “LLM routers”—intermediary services positioned between users and AI models—have become a major threat to cryptocurrency asset security. The researchers discovered that 26 LLM routers are secretly injecting malicious tool calls and stealing user credentials, with one incident resulting in the complete draining of a customer’s cryptocurrency wallet worth $500,000. Additionally, by “poisoning” the router ecosystem, the researchers were able to gain control of approximately 400 downstream hosts within hours. Since sensitive data—including private keys and API credentials—is frequently transmitted in plaintext through these routers, users unknowingly expose their assets to risk. The researchers note that as McKinsey forecasts AI agents will mediate $3–5 trillion in global consumer commerce by 2030—and Binance founder Changpeng Zhao predicts AI agents’ payment volume will be one million times greater than that of humans—the current infrastructure’s security lags far behind the pace of industry development. The “weakest link” risk could thus trigger systemic, cascading crises.
According to Cointelegraph, researchers from the University of California recently revealed security risks in certain third-party AI large language model (LLM) routers that could lead to the theft of cryptocurrency assets. The study found that LLM routers—acting as API intermediaries—can read plaintext information; some routers were discovered injecting malicious code and stealing credentials. The research team tested 28 paid and 400 free routers, identifying nine routers that actively injected malicious code, two that deployed trigger-avoidance mechanisms, and 17 that accessed Amazon Web Services (AWS) credentials. One router even transferred ETH using the researchers’ Ethereum private key. The study notes that malicious behavior by routers is difficult to detect, and the “YOLO mode” present in some AI agent frameworks—which automatically executes commands—further increases security risks. Researchers recommend that developers avoid transmitting private keys or mnemonic phrases through AI agents and urge AI companies to implement cryptographic signing of responses to enhance security.
According to official announcements, B.AI achieved multiple advancements this week in product iteration and ecosystem development: The BAIclaw landing page underwent a comprehensive visual and interactive redesign; the website’s multilingual support expanded to 10 languages, further strengthening its global accessibility. On the infrastructure front, strategic partnerships with Biconomy, MoonPay, and Pundi X significantly optimized the Web3 payment flow, substantially lowering the barrier to entry for users. Meanwhile, B.AI’s multi-chain LLM service continues to evolve—featuring intelligent search, Boundless Mode, and memory capabilities—resulting in markedly enhanced user interaction. Looking ahead, with the upcoming launch of subscription systems, point-based incentive mechanisms, and core Skills such as the “Sun Yuchen Brain,” B.AI is accelerating the construction of a fully functional intelligent ecosystem matrix, powered by the deep integration of AI Agents and Web3.
According to official announcements, Web3 intelligence-layer project Claw Intelligence has raised $3 million in seed funding. Investors include Castrum Istanbul, Titans Ventures, Super Labs, and Genesis Capital. Claw Intelligence is simplifying user interaction with the Web3 ecosystem and underlying computing resources through its unified intelligence layer, lowering operational barriers. The platform leverages an encryption-native Model Context Protocol (MCP) service to transform fragmented data endpoints into conversational workflows. New features include a secure, isolated large language model (LLM)-powered code execution sandbox that safely runs code directly within the chat interface—enabling real-time computation, data processing, and script prototyping; and LLM-driven cross-device/multi-computer control, allowing users to centrally manage multiple devices and servers via natural-language commands.
According to an official announcement, AI infrastructure platform B.AI LLM Services today announced that its user base has officially surpassed one million. This milestone not only signifies that its “privacy-first” routing architecture can handle large-scale, sustained high-throughput workloads, but also reflects B.AI’s evolution beyond developing a single chat interface—toward deploying critical underlying infrastructure for the “autonomous agent economy.” Through deep optimization of its technology stack, B.AI is laying a robust foundation for the future highly automated AI collaboration network.
According to official announcements, the number of users on the B.AI LLM service platform has officially surpassed 801,057. Built upon the foundational principles of “permissionless access, privacy protection, and instant response,” B.AI integrates leading global models via intelligent routing technology to enable anonymous access under a “wallet-as-identity” framework and native on-chain settlement. Beyond enhancing interaction efficiency, the platform is committed to advancing AI’s paradigm shift—from a “passive response tool” to an “independent economic entity.” By deeply empowering logical collaboration among sophisticated agents, B.AI is building a secure, barrier-free intelligent gateway for the AGI era.
According to official announcements, the B.AI LLM service platform has officially surpassed 750,000 users, maintaining its leadership in the intelligent access segment. Leveraging intelligent routing technology, wallet-driven anonymous privacy protection mechanisms, and a unified multi-model management interface, the platform has gradually become the default gateway connecting major global AI models. With its three core advantages—efficient orchestration, privacy-first design, and frictionless access—B.AI is redefining the new standard for intelligent access: efficient, secure, and barrier-free—propelling large-model capabilities toward broader developer and user communities.
According to CoinDesk, researchers from the University of California, Santa Barbara; the University of California, San Diego; blockchain security firm Fuzzland; and World Liberty Financial jointly published a paper warning that “LLM routers”—intermediary services positioned between users and AI models—have become a major threat to cryptocurrency asset security. The researchers discovered that 26 LLM routers are secretly injecting malicious tool calls and stealing user credentials, with one incident resulting in the complete draining of a customer’s cryptocurrency wallet worth $500,000. Additionally, by “poisoning” the router ecosystem, the researchers were able to gain control of approximately 400 downstream hosts within hours. Since sensitive data—including private keys and API credentials—is frequently transmitted in plaintext through these routers, users unknowingly expose their assets to risk. The researchers note that as McKinsey forecasts AI agents will mediate $3–5 trillion in global consumer commerce by 2030—and Binance founder Changpeng Zhao predicts AI agents’ payment volume will be one million times greater than that of humans—the current infrastructure’s security lags far behind the pace of industry development. The “weakest link” risk could thus trigger systemic, cascading crises.