Connect with us

Uncategorized

The DeepSeek-R1 Effect and Web3-AI

Published

on

The artificial intelligence (AI) world was taken by storm a few days ago with the release of DeepSeek-R1, an open-source reasoning model that matches the performance of top foundation models while claiming to have been built using a remarkably low training budget and novel post-training techniques. The release of DeepSeek-R1 not only challenged the conventional wisdom surrounding the scaling laws of foundation models – which traditionally favor massive training budgets – but did so in the most active area of research in the field: reasoning.

The open-weights (as opposed to open-source) nature of the release made the model readily accessible to the AI community, leading to a surge of clones within hours. Moreover, DeepSeek-R1 left its mark on the ongoing AI race between China and the United States, reinforcing what has been increasingly evident: Chinese models are of exceptionally high quality and fully capable of driving innovation with original ideas.

Unlike most advancements in generative AI, which seem to widen the gap between Web2 and Web3 in the realm of foundation models, the release of DeepSeek-R1 carries real implications and presents intriguing opportunities for Web3-AI. To assess these, we must first take a closer look at DeepSeek-R1’s key innovations and differentiators.

Inside DeepSeek-R1

DeepSeek-R1 was the result of introducing incremental innovations into a well-established pretraining framework for foundation models. In broad terms, DeepSeek-R1 follows the same training methodology as most high-profile foundation models. This approach consists of three key steps:

Pretraining: The model is initially pretrained to predict the next word using massive amounts of unlabeled data.

Supervised Fine-Tuning (SFT): This step optimizes the model in two critical areas: following instructions and answering questions.

Alignment with Human Preferences: A final fine-tuning phase is conducted to align the model’s responses with human preferences.

Most major foundation models – including those developed by OpenAI, Google, and Anthropic – adhere to this same general process. At a high level, DeepSeek-R1’s training procedure does not appear significantly different. ButHowever, rather than pretraining a base model from scratch, R1 leveraged the base model of its predecessor, DeepSeek-v3-base, which boasts an impressive 617 billion parameters.

In essence, DeepSeek-R1 is the result of applying SFT to DeepSeek-v3-base with a large-scale reasoning dataset. The real innovation lies in the construction of these reasoning datasets, which are notoriously difficult to build.

First Step: DeepSeek-R1-Zero

One of the most important aspects of DeepSeek-R1 is that the process did not produce just a single model but two. Perhaps the most significant innovation of DeepSeek-R1 was the creation of an intermediate model called R1-Zero, which is specialized in reasoning tasks. This model was trained almost entirely using reinforcement learning, with minimal reliance on labeled data.

Reinforcement learning is a technique in which a model is rewarded for generating correct answers, enabling it to generalize knowledge over time.

R1-Zero is quite impressive, as it was able to match GPT-o1 in reasoning tasks. However, the model struggled with more general tasks such as question-answering and readability. That said, the purpose of R1-Zero was never to create a generalist model but rather to demonstrate it is possible to achieve state-of-the-art reasoning capabilities using reinforcement learning alone – even if the model does not perform well in other areas.

Second-Step: DeepSeek-R1

DeepSeek-R1 was designed to be a general-purpose model that excels at reasoning, meaning it needed to outperform R1-Zero. To achieve this, DeepSeek started once again with its v3 model, but this time, it fine-tuned it on a small reasoning dataset.

As mentioned earlier, reasoning datasets are difficult to produce. This is where R1-Zero played a crucial role. The intermediate model was used to generate a synthetic reasoning dataset, which was then used to fine-tune DeepSeek v3. This process resulted in another intermediate reasoning model, which was subsequently put through an extensive reinforcement learning phase using a dataset of 600,000 samples, also generated by R1-Zero. The final outcome of this process was DeepSeek-R1.

While I have omitted several technical details of the R1 pretraining process, here are the two main takeaways:

R1-Zero demonstrated that it is possible to develop sophisticated reasoning capabilities using basic reinforcement learning. Although R1-Zero was not a strong generalist model, it successfully generated the reasoning data necessary for R1.

R1 expanded the traditional pretraining pipeline used by most foundation models by incorporating R1-Zero into the process. Additionally, it leveraged a significant amount of synthetic reasoning data generated by R1-Zero.

As a result, DeepSeek-R1 emerged as a model that matched the reasoning capabilities of GPT-o1 while being built using a simpler and likely significantly cheaper pretraining process.

Everyone agrees that R1 marks an important milestone in the history of generative AI, one that is likely to reshape the way foundation models are developed. When it comes to Web3, it will be interesting to explore how R1 influences the evolving landscape of Web3-AI.

DeepSeek-R1 and Web3-AI

Until now, Web3 has struggled to establish compelling use cases that clearly add value to the creation and utilization of foundation models. To some extent, the traditional workflow for pretraining foundation models appears to be the antithesis of Web3 architectures. However, despite being in its early stages, the release of DeepSeek-R1 has highlighted several opportunities that could naturally align with Web3-AI architectures.

1) Reinforcement Learning Fine-Tuning Networks

R1-Zero demonstrated that it is possible to develop reasoning models using pure reinforcement learning. From a computational standpoint, reinforcement learning is highly parallelizable, making it well-suited for decentralized networks. Imagine a Web3 network where nodes are compensated for fine-tuning a model on reinforcement learning tasks, each applying different strategies. This approach is far more feasible than other pretraining paradigms that require complex GPU topologies and centralized infrastructure.

2) Synthetic Reasoning Dataset Generation

Another key contribution of DeepSeek-R1 was showcasing the importance of synthetically generated reasoning datasets for cognitive tasks. This process is also well-suited for a decentralized network, where nodes execute dataset generation jobs and are compensated as these datasets are used for pretraining or fine-tuning foundation models. Since this data is synthetically generated, the entire network can be fully automated without human intervention, making it an ideal fit for Web3 architectures.

3) Decentralized Inference for Small Distilled Reasoning Models

DeepSeek-R1 is a massive model with 671 billion parameters. However, almost immediately after its release, a wave of distilled reasoning models emerged, ranging from 1.5 to 70 billion parameters. These smaller models are significantly more practical for inference in decentralized networks. For example, a 1.5B–2B distilled R1 model could be embedded in a DeFi protocol or deployed within nodes of a DePIN network. More simply, we are likely to see the rise of cost-effective reasoning inference endpoints powered by decentralized compute networks. Reasoning is one domain where the performance gap between small and large models is narrowing, creating a unique opportunity for Web3 to efficiently leverage these distilled models in decentralized inference settings.

4) Reasoning Data Provenance

One of the defining features of reasoning models is their ability to generate reasoning traces for a given task. DeepSeek-R1 makes these traces available as part of its inference output, reinforcing the importance of provenance and traceability for reasoning tasks. The internet today primarily operates on outputs, with little visibility into the intermediate steps that lead to those results. Web3 presents an opportunity to track and verify each reasoning step, potentially creating a «new internet of reasoning» where transparency and verifiability become the norm.

Web3-AI Has a Chance in the Post-R1 Reasoning Era

The release of DeepSeek-R1 has marked a turning point in the evolution of generative AI. By combining clever innovations with established pretraining paradigms, it has challenged traditional AI workflows and opened a new era in reasoning-focused AI. Unlike many previous foundation models, DeepSeek-R1 introduces elements that bring generative AI closer to Web3.

Key aspects of R1 – synthetic reasoning datasets, more parallelizable training and the growing need for traceability – align naturally with Web3 principles. While Web3-AI has struggled to gain meaningful traction, this new post-R1 reasoning era may present the best opportunity yet for Web3 to play a more significant role in the future of AI.

Continue Reading
Click to comment

Leave a Reply

Ваш адрес email не будет опубликован. Обязательные поля помечены *

Uncategorized

Robinhood, Kraken-Backed Global Dollar (USDG) Comes to Europe

Published

on

By

Global Dollar (USDG), a stablecoin issued by regulated fintech Paxos, and backed by a consortium of heavy hitters that includes Robinhood, Kraken and Mastercard, is being made available to consumers across the European Union, according to a press release on Tuesday.

USDG is regulated by Europe’s Markets in Crypto-Assets (MiCA), the Finnish Financial Supervisory Authority (FIN-FSA), and the Monetary Authority of Singapore (MAS), Paxos said in a statement.

Demand for U.S. dollar-backed stablecoins is growing in Europe where Circle’s USDC token is the largest MiCA-regulated choice. USDG will make a significant impact as an alternative regulated option, Paxos said.

“USDG is a fully regulated global USD-stablecoin that is compliant with MiCA and now available in the EU, a testament to our commitment to offering global digital assets that are supervised by prudential regulators and also meet the highest standards of consumer protection,” said Walter Hessert, head of strategy at Paxos.

Fulfilling requirements under the EU’s MiCA regulation necessitates that Paxos Issuance Europe, which is regulated by FIN-FSA, holds a portion of USDG reserve assets with European banking partners, Paxos said.

Continue Reading

Uncategorized

XRP, TRX, DOGE Lead Majors With Positive Funding Rates as Bitcoin’s Traditionally Weak Quarter Begins

Published

on

By

A key metric called perpetual funding rates is signaling bullishness for top altcoins as bitcoin (BTC) kicks off the traditionally weak third quarter quarter with flat price action.

Funding rates, charged by exchanges every eight hours, refer to the cost of holding bullish long or bearish short positions in the perpetual (perps) futures (with no expiry).

A positive funding rate indicates that perps are trading at a premium to the spot price, necessitating a payment from longs to shorts to maintain bullish bets. Therefore, positive rates are interpreted as representing bullish sentiment, while negative rates suggest otherwise.

As of writing, perps tied to payments-focused token XRP (XRP), the world’s fourth-largest digital asset by market value, had an annualized funding rate of nearly 11%, the highest among the top 10 tokens, according to data source Velo. Funding rates for Tron’s TRX (TRX) and dogecoin (DOGE) were 10% and 8.4%, respectively, while rates for market leaders bitcoin and ether were marginally positive.

In other words, the XRP market demonstrated the strongest demand for leveraged bullish exposure among other major cryptocurrencies, including BTC and ether (ETH). That’s consistent with the spike in bullish sentiment for XRP last week, despite the settlement between Ripple and the SEC stalling, as noted by Santiment.

Funding rates for cryptocurrencies. (Velo Data)

Privacy-focused monero (XMR) stood among tokens beyond the top 10 list with a funding rate of over 23%, while Stellar’s XLM token signaled a strong bias for bearish bets with a funding rate of 24%.

Seasonally weak quarter

Historically, the third quarter has been a weak period for bitcoin, with data indicating an average gain of 5.57% since 2013, according to Coinglass. That’s a far cry compared to the fourth quarter’s 85% average gain.

BTC’s spot price remained flat at around $107,000 at press time, offering no clear direction bias. Valuations have been stuck largely between $100,000 and $110,000 for nearly 50 days, with selling by long-term holder wallets counteracting persistent inflows into the U.S.-listed spot exchange-traded funds (ETFs).

Some analysts, however, expect a significant move to occur soon, with all eyes on Fed Chairman Jerome Powell’s speech on Tuesday and the release of nonfarm payrolls on Friday.

Continue Reading

Uncategorized

Asia Morning Briefing: Are Distributed Compute Tokens Undervalued vs. CoreWeave (CRWV)?

Published

on

By

Tech investors love to pay for potential. GameFi tokens, with sky-high valuations divorced from current user numbers or revenues, embody this optimism perfectly — as CoinDesk investigated in 2022, Decentraland’s then billion-dollar market cap didn’t quite match the number of active players on the platform.

But, surprisingly, distributed compute tokens don’t seem to enjoy the same speculative premium even when compared to their Traditional Finance traded peers like CoreWeave (CRWV).

CoinMarketCap says the category of tokens for decentralized networks that provide GPU power for AI and other compute workloads, which includes well-known tokens like BitTensor, Aethir, and Render, is worth $12 billion.

At the same time, market data from research group MarketsandMarkets puts the value of the GPU as a service industry at around $8 billion this year, growing to $26 billion in 2030.

In contrast, CRWV closed Monday in New York at $163, putting its market cap at $79.2 billion. The company’s recent earnings forecast up to $5.1 billion in 2025 revenue, suggesting it trades at more than 15 times forward sales.

That kind of multiple might be justified in a high-growth environment, but CoreWeave also posted a $314.6 million net loss in the first quarter, driven in part by stock-based compensation and continued infrastructure buildout.

Despite this, investors continue to reward CoreWeave for its dominant position in centralized AI infrastructure with its stock up 300% year-to-date. The company is tightly integrated with Nvidia and has high visibility through contracts with OpenAI and other enterprise clients.

Meanwhile, decentralized compute networks are delivering similar services— AI inference, rendering, and compute power — without needing to raise billions in debt or equity as they act as a broker connecting existing GPUs to users, saving the capital expenditure of buying their own server farms.

These are not theoretical networks. They are functional systems already processing real workloads, and the brokerage model works for customers.

Yet their collective market value remains a fraction of CoreWeave’s. Certainly, they don’t have the same level of workload running through their networks, but the gap is striking. While the market treats GameFi with irrational exuberance, distributed compute tokens may be suffering from the opposite problem.

Despite addressing the same market need as CoreWeave, and in some ways offering a more capital-efficient and globally scalable model without the eye-watering CapEx, they remain modestly valued.

Justin Sun-Backed SRM Entertainment Announces $100 Million TRX Staking Move

SRM Entertainment (Nasdaq: SRM), soon to rebrand as TRON Inc., has staked its entire treasury of 365 million TRX tokens through JustLend, a move that could yield an annual return of up to 10%, according to a release.

The move comes on the heels of a $100 million investment round closed earlier this month to fund what the company calls a “TRON treasury strategy,” essentially, a public market vehicle modeled on bitcoin-holding firms like MicroStrategy, but for TRX.

That structure provides equity investors with indirect exposure to a network that plays a dominant role in USDT stablecoin settlement, particularly in the Global South, where TRON-based Tether serves as a dollar lifeline – arguably a ‘Visa IPO‘ moment for the region’s economy.

Sogni AI Debuts Mainnet, SOGNI Token to List on Kraken, MEXC, Gate.io

Sogni AI, a decentralized platform for generative AI workflows, has launched its mainnet and will list its native token, SOGNI, on Kraken, MEXC, and Gate.io.

SOGNI is the utility token of the Sogni Supernet. It is used for compute payments, staking, governance, and access to advanced application features.

The mainnet launch includes deployments on Base, an Ethereum Layer-2 developed by Coinbase, and Etherlink, a Tezos-based EVM-compatible Layer-2 using Smart Rollups. In a release, the platform said this chain-agnostic approach is designed to balance scalability and accessibility.

The project’s stated goal is to create an open and economically sustainable environment for creative AI applications, combining Web3 infrastructure with user tools that resemble Web2 services in usability.

The platform also uses a non-transferable credit system called Spark Points, which are fixed-value rendering credits that can be purchased or earned within the Sogni ecosystem.

Users interact with the network through three core applications: Sogni Web, Sogni Pocket, and Sogni Studio. Creators submit generative AI jobs, while node operators, or “Workers,” provide GPU resources and are compensated in SOGNI tokens.

Market Movements:

  • BTC: Bitcoin is trading at $107,200, holding a strong support zone after a 14,695 BTC volume spike near $107K, with traders eyeing a potential breakout toward $115,000.
  • ETH: Ethereum rebounded sharply from a 3.4% intraday drop, currently trading at $2,480, forming a V-shaped recovery off $2,438 support, as institutional inflows continue despite broader market uncertainty.
  • Gold: Gold is trading at $3,310.95, rebounding from a one-month low as a weaker dollar and Fed pressure offset risk-on sentiment.
  • Nikkei 225: Asia-Pacific markets traded mixed Tuesday as investors weighed Wall Street’s record highs against looming uncertainty from Trump’s expiring 90-day tariff reprieve, with Japan’s Nikkei 225 down 0.58%
  • S&P 500: Stocks climbed Monday as the S&P 500 rose 0.52% to a record close of 6,204.95, capping a strong month.

Elsewhere in Crypto:

Continue Reading

Trending

Copyright © 2017 Zox News Theme. Theme by MVP Themes, powered by WordPress.