Connect with us

Uncategorized

The DeepSeek-R1 Effect and Web3-AI

Published

on

The artificial intelligence (AI) world was taken by storm a few days ago with the release of DeepSeek-R1, an open-source reasoning model that matches the performance of top foundation models while claiming to have been built using a remarkably low training budget and novel post-training techniques. The release of DeepSeek-R1 not only challenged the conventional wisdom surrounding the scaling laws of foundation models – which traditionally favor massive training budgets – but did so in the most active area of research in the field: reasoning.

The open-weights (as opposed to open-source) nature of the release made the model readily accessible to the AI community, leading to a surge of clones within hours. Moreover, DeepSeek-R1 left its mark on the ongoing AI race between China and the United States, reinforcing what has been increasingly evident: Chinese models are of exceptionally high quality and fully capable of driving innovation with original ideas.

Unlike most advancements in generative AI, which seem to widen the gap between Web2 and Web3 in the realm of foundation models, the release of DeepSeek-R1 carries real implications and presents intriguing opportunities for Web3-AI. To assess these, we must first take a closer look at DeepSeek-R1’s key innovations and differentiators.

Inside DeepSeek-R1

DeepSeek-R1 was the result of introducing incremental innovations into a well-established pretraining framework for foundation models. In broad terms, DeepSeek-R1 follows the same training methodology as most high-profile foundation models. This approach consists of three key steps:

Pretraining: The model is initially pretrained to predict the next word using massive amounts of unlabeled data.

Supervised Fine-Tuning (SFT): This step optimizes the model in two critical areas: following instructions and answering questions.

Alignment with Human Preferences: A final fine-tuning phase is conducted to align the model’s responses with human preferences.

Most major foundation models – including those developed by OpenAI, Google, and Anthropic – adhere to this same general process. At a high level, DeepSeek-R1’s training procedure does not appear significantly different. ButHowever, rather than pretraining a base model from scratch, R1 leveraged the base model of its predecessor, DeepSeek-v3-base, which boasts an impressive 617 billion parameters.

In essence, DeepSeek-R1 is the result of applying SFT to DeepSeek-v3-base with a large-scale reasoning dataset. The real innovation lies in the construction of these reasoning datasets, which are notoriously difficult to build.

First Step: DeepSeek-R1-Zero

One of the most important aspects of DeepSeek-R1 is that the process did not produce just a single model but two. Perhaps the most significant innovation of DeepSeek-R1 was the creation of an intermediate model called R1-Zero, which is specialized in reasoning tasks. This model was trained almost entirely using reinforcement learning, with minimal reliance on labeled data.

Reinforcement learning is a technique in which a model is rewarded for generating correct answers, enabling it to generalize knowledge over time.

R1-Zero is quite impressive, as it was able to match GPT-o1 in reasoning tasks. However, the model struggled with more general tasks such as question-answering and readability. That said, the purpose of R1-Zero was never to create a generalist model but rather to demonstrate it is possible to achieve state-of-the-art reasoning capabilities using reinforcement learning alone – even if the model does not perform well in other areas.

Second-Step: DeepSeek-R1

DeepSeek-R1 was designed to be a general-purpose model that excels at reasoning, meaning it needed to outperform R1-Zero. To achieve this, DeepSeek started once again with its v3 model, but this time, it fine-tuned it on a small reasoning dataset.

As mentioned earlier, reasoning datasets are difficult to produce. This is where R1-Zero played a crucial role. The intermediate model was used to generate a synthetic reasoning dataset, which was then used to fine-tune DeepSeek v3. This process resulted in another intermediate reasoning model, which was subsequently put through an extensive reinforcement learning phase using a dataset of 600,000 samples, also generated by R1-Zero. The final outcome of this process was DeepSeek-R1.

While I have omitted several technical details of the R1 pretraining process, here are the two main takeaways:

R1-Zero demonstrated that it is possible to develop sophisticated reasoning capabilities using basic reinforcement learning. Although R1-Zero was not a strong generalist model, it successfully generated the reasoning data necessary for R1.

R1 expanded the traditional pretraining pipeline used by most foundation models by incorporating R1-Zero into the process. Additionally, it leveraged a significant amount of synthetic reasoning data generated by R1-Zero.

As a result, DeepSeek-R1 emerged as a model that matched the reasoning capabilities of GPT-o1 while being built using a simpler and likely significantly cheaper pretraining process.

Everyone agrees that R1 marks an important milestone in the history of generative AI, one that is likely to reshape the way foundation models are developed. When it comes to Web3, it will be interesting to explore how R1 influences the evolving landscape of Web3-AI.

DeepSeek-R1 and Web3-AI

Until now, Web3 has struggled to establish compelling use cases that clearly add value to the creation and utilization of foundation models. To some extent, the traditional workflow for pretraining foundation models appears to be the antithesis of Web3 architectures. However, despite being in its early stages, the release of DeepSeek-R1 has highlighted several opportunities that could naturally align with Web3-AI architectures.

1) Reinforcement Learning Fine-Tuning Networks

R1-Zero demonstrated that it is possible to develop reasoning models using pure reinforcement learning. From a computational standpoint, reinforcement learning is highly parallelizable, making it well-suited for decentralized networks. Imagine a Web3 network where nodes are compensated for fine-tuning a model on reinforcement learning tasks, each applying different strategies. This approach is far more feasible than other pretraining paradigms that require complex GPU topologies and centralized infrastructure.

2) Synthetic Reasoning Dataset Generation

Another key contribution of DeepSeek-R1 was showcasing the importance of synthetically generated reasoning datasets for cognitive tasks. This process is also well-suited for a decentralized network, where nodes execute dataset generation jobs and are compensated as these datasets are used for pretraining or fine-tuning foundation models. Since this data is synthetically generated, the entire network can be fully automated without human intervention, making it an ideal fit for Web3 architectures.

3) Decentralized Inference for Small Distilled Reasoning Models

DeepSeek-R1 is a massive model with 671 billion parameters. However, almost immediately after its release, a wave of distilled reasoning models emerged, ranging from 1.5 to 70 billion parameters. These smaller models are significantly more practical for inference in decentralized networks. For example, a 1.5B–2B distilled R1 model could be embedded in a DeFi protocol or deployed within nodes of a DePIN network. More simply, we are likely to see the rise of cost-effective reasoning inference endpoints powered by decentralized compute networks. Reasoning is one domain where the performance gap between small and large models is narrowing, creating a unique opportunity for Web3 to efficiently leverage these distilled models in decentralized inference settings.

4) Reasoning Data Provenance

One of the defining features of reasoning models is their ability to generate reasoning traces for a given task. DeepSeek-R1 makes these traces available as part of its inference output, reinforcing the importance of provenance and traceability for reasoning tasks. The internet today primarily operates on outputs, with little visibility into the intermediate steps that lead to those results. Web3 presents an opportunity to track and verify each reasoning step, potentially creating a «new internet of reasoning» where transparency and verifiability become the norm.

Web3-AI Has a Chance in the Post-R1 Reasoning Era

The release of DeepSeek-R1 has marked a turning point in the evolution of generative AI. By combining clever innovations with established pretraining paradigms, it has challenged traditional AI workflows and opened a new era in reasoning-focused AI. Unlike many previous foundation models, DeepSeek-R1 introduces elements that bring generative AI closer to Web3.

Key aspects of R1 – synthetic reasoning datasets, more parallelizable training and the growing need for traceability – align naturally with Web3 principles. While Web3-AI has struggled to gain meaningful traction, this new post-R1 reasoning era may present the best opportunity yet for Web3 to play a more significant role in the future of AI.

Continue Reading
Click to comment

Leave a Reply

Ваш адрес email не будет опубликован. Обязательные поля помечены *

Uncategorized

Riot Platforms Hits Post-Halving Bitcoin Production High as It Expands AI Capacity

Published

on

By

Riot Platforms (RIOT) reported strong operational performance in March 2025, highlighted by continued expansion into the artificial intelligence (AI) and high-performance computing (HPC) sector.

The company’s bitcoin (BTC) production last month rose to 533 BTC, the most since the reward halving almost a year ago. The figure represents a month-on-month increase of 13% and 25% more than a year before. Bitcoin holdings grew to 19,223 BTC.

Riot said it plans to «aggressively pursue» development of its Corsicana facility to capitalize on rising demand for compute infrastructure used in AI and HPC.

A recently completed feasibility study by industry consultant Altman Solon confirmed the significant potential of the site to support up to 600 megawatts of additional capacity for AI/HPC applications. Key advantages include 1.0 gigawatt of secured power, 400 MW of which is already operational, 265 acres of land with substantial development potential and close proximity to Dallas — a major hub for AI and cloud computing.

The study noted the site’s ability to support both inference and cloud-based workloads, strengthening its appeal to AI/HPC tenants.

Riot maintained a steady deployed hash rate of 33.7 EH/s, while its average operating hash rate grew 3% month-over-month to 30.3 EH/s—representing a 254% increase year-over-year. Although power credits declined due to seasonal factors, Riot kept its all-in power cost low at 3.8 cents per kWh, and improved fleet efficiency to 21.0 J/TH, a 22% improvement from the previous year.

Riot’s shares fell 5.5% Friday, while the Nasdaq 100 index dropped 2.8%. They have lost 35% year-to-date.

Disclaimer: This article was generated with AI tools and reviewed by our editorial team to ensure accuracy and adherence to our standards. For more information, see CoinDesk’s full AI Policy. This article may include information from external sources, which are listed below when applicable.

Continue Reading

Uncategorized

A Blueprint for Digital Assets in America

Published

on

By

In 2008, an anonymous person or group of people known only as “Satoshi Nakamoto” released a now-seminal document, the Bitcoin White paper, introducing a peer-to-peer system for value of exchange without intermediaries.

With this revolutionary concept, the idea of a “digital asset” was born. Soon after, developers and entrepreneurs expanded on this concept, developing systems where value was exchanged not just for its own sake, but for services and digital products.

Over the past decade, innovators have built permissionless, decentralized networks for computing services, file storage, asset exchange, cellular coverage, Wi-Fi connectivity, mapping tools, lending services, and more. Because digital assets can be used for services that anyone can offer and anyone can access, the use-cases – both financial and non-financial – are potentially endless.

Despite this promise, these networks have courted criticism. The Biden-Harris Administration attempted to block this innovative advance through a relentless campaign of lawsuits and enforcement actions without providing the regulatory clarity the digital asset ecosystem and its innovators and users so desperately needed.

The Securities and Exchange Commission (SEC) failed to clarify how existing securities laws apply and — more importantly — don’t apply to digital asset transactions. This lack of regulatory clarity stifled the digital asset ecosystem, pushing growth out of the United States to jurisdictions that have established clear rules of the road.

To address these failures, Congress began exploring ways to modernize the regulatory structure to accommodate the unique characteristics of digital assets and how they could be used in our financial system. These efforts culminated in a series of bills aimed at clarifying how digital assets could be used in the financial system, ensuring investor protection and fostering innovation.

In the 118th Congress, the House Committees on Financial Services and Agriculture launched a historic joint effort to address digital asset regulation. This led to the first-ever passage of bipartisan digital asset market structure legislation in a chamber of Congress. This collaboration enabled Congress to address longstanding challenges in the ecosystem and lay the foundation for a fit for purpose framework under the leadership of President Trump.

This Congress, both the House and Senate are committed to creating a clear path forward for the digital asset ecosystem. As we move ahead, it is crucial that the framework is both balanced and iron-clad for the future. To accomplish this, we have set out principles for digital asset legislation.

Six principles

First, legislation must promote innovation. We seek to protect opportunities for innovators to create and utilize digital assets, while ensuring users can lawfully transact with one another.

Second, legislation must provide clarity for the classification of assets. Users of digital assets should clearly understand the nature of their holdings, including whether they qualify as securities or non-securities.

Third, legislation must codify a framework for the issuance of new digital assets. The framework should permit issuers to raise capital through the sale of new digital assets under the jurisdiction of the SEC. It should protect retail investors and require developers to disclose relevant information to help users understand the unique characteristics of digital asset networks.

Fourth, the legislation must establish the regulation of spot market exchanges and intermediaries. Centralized, custodial exchanges and intermediaries facilitating transactions with non-security digital assets should adhere to similar requirements as other financial firms.

Congress should provide the Commodity Futures Trading Commission (CFTC) with the authority to impose requirements over these entities necessary to protect customers, limit conflicts of interest, ensure appropriate execution of customer orders, and provide disclosures.

Fifth, the legislation must establish best practices for the protection of customer assets. Entities registered with the SEC or CFTC should be required to segregate customer funds and hold them with qualified custodians. Customer funds should also be protected during bankruptcy.

Sixth, and finally, the legislation must protect innovative decentralized projects and activities. Congress should ensure that decentralized protocols, which pose different risks and benefits, are not subject to regulations designed for centralized, custodial firms. In safeguarding decentralized activities, Congress must also protect an individual’s right to self-custody their digital assets.

We look forward to both Committees continuing our legislative work together to fulfill President Trump’s request to make America the “crypto capital of the planet.” In May, our Committees will host our second joint hearing to discuss digital asset market structure legislation.

Our goal is to bring much-needed regulatory clarity to this rapidly evolving industry, ensuring that America continues to lead in shaping the future of digital finance.

Continue Reading

Uncategorized

OKX Fined $1.2M by Malta for Breaching Money Laundering Rules

Published

on

By

OKX’s Europe company—also known as OKCoin Europe, a subsidiary of crypto exchange OKX—was fined 1.05 million euros ($1.2 million) by Malta’s financial watchdog on Thursday for breaching the country’s money laundering rules.

The Financial Intelligence Analysis Unit (FIAU) said the company failed to assess the money laundering and financing of terrorism risks emanating from the products it offers and had violated parts of the country’s Prevention of Money Laundering and Financing of Terrorism Regulations.

«Regulatory compliance is a top priority for OKX, and we remain committed to meeting and exceeding global regulatory standards,» OKX said in a statement.

The company also said it had addressed gaps identified in its compliance framework following the authority’s 2023 review. In the new notice, FIAU also commended the company on making significant improvements over the past 18 months.

OKX secured the coveted Markets in Crypto Assets license (MiCA) from Malta earlier this year, which will enable it to offer crypto services across the European Union.

«The company was expected to assess the nature of risks prevalent in the services it was offering,» the authority said in its notice.

FIAU said the exchange should assess risks tied to the use of stablecoins, mixers that obscure the origins of transactions, privacy coins, tokens designed for anonymity, and tokens on decentralized exchanges.

OKX recently temporarily suspended its decentralized exchange aggregator following reports that European regulators had been looking at how it had been used to launder funds from a recent hack of the Bybit exchange.

Bloomberg first reported the story.

Continue Reading

Trending

Copyright © 2017 Zox News Theme. Theme by MVP Themes, powered by WordPress.