Skip to content

GitLab

  • Menu
Projects Groups Snippets
    • Loading...
  • Help
    • Help
    • Support
    • Community forum
    • Submit feedback
    • Contribute to GitLab
  • Sign in
  • C caybora
  • Project information
    • Project information
    • Activity
    • Labels
    • Members
  • Issues 1
    • Issues 1
    • List
    • Boards
    • Service Desk
    • Milestones
  • Merge requests 0
    • Merge requests 0
  • CI/CD
    • CI/CD
    • Pipelines
    • Jobs
    • Schedules
  • Deployments
    • Deployments
    • Environments
  • Monitor
    • Monitor
    • Incidents
  • Packages & Registries
    • Packages & Registries
    • Package Registry
    • Infrastructure Registry
  • Analytics
    • Analytics
    • Value stream
  • Wiki
    • Wiki
  • Snippets
    • Snippets
  • Activity
  • Create a new issue
  • Jobs
  • Issue Boards
Collapse sidebar
  • Chasity Brifman
  • caybora
  • Issues
  • #1

Closed
Open
Created Feb 11, 2025 by Chasity Brifman@chasitybrifmanMaintainer

DeepSeek R1's Implications: Winners and Losers in the Generative AI Value Chain


R1 is mainly open, on par with leading exclusive designs, appears to have actually been trained at substantially lower cost, and is more affordable to utilize in regards to API gain access to, all of which point to an innovation that may change competitive dynamics in the field of Generative AI.

  • IoT Analytics sees end users and AI applications companies as the greatest winners of these recent advancements, while exclusive design companies stand to lose the most, based on value chain analysis from the Generative AI Market Report 2025-2030 (published January 2025).

Why it matters

For suppliers to the generative AI value chain: Players along the (generative) AI worth chain may need to re-assess their value propositions and line up to a possible reality of low-cost, lightweight, open-weight models. For generative AI adopters: DeepSeek R1 and other frontier models that may follow present lower-cost options for AI adoption.
Background: DeepSeek's R1 design rattles the marketplaces

DeepSeek's R1 model rocked the stock markets. On January 23, 2025, China-based AI start-up DeepSeek launched its open-source R1 reasoning generative AI (GenAI) design. News about R1 rapidly spread, and by the start of stock trading on January 27, 2025, the marketplace cap for lots of major innovation companies with big AI footprints had fallen considerably given that then:

NVIDIA, a US-based chip designer and designer most known for its information center GPUs, dropped 18% in between the market close on January 24 and the market close on February 3. Microsoft, the leading hyperscaler in the cloud AI race with its Azure cloud services, dropped 7.5% (Jan 24-Feb 3). Broadcom, a semiconductor business concentrating on networking, broadband, and custom-made ASICs, dropped 11% (Jan 24-Feb 3). Siemens Energy, a German energy technology vendor that provides energy services for information center operators, dropped 17.8% (Jan 24-Feb 3).
Market individuals, and particularly investors, reacted to the narrative that the model that DeepSeek released is on par with innovative designs, was allegedly trained on just a number of countless GPUs, and is open source. However, since that preliminary sell-off, reports and analysis shed some light on the initial buzz.

The insights from this short article are based on

Download a sample to read more about the report structure, choose meanings, select market data, extra data points, and trends.

DeepSeek R1: What do we understand till now?

DeepSeek R1 is an affordable, innovative thinking model that equals leading competitors while promoting openness through openly available weights.

DeepSeek R1 is on par with leading reasoning designs. The biggest DeepSeek R1 design (with 685 billion specifications) performance is on par or even much better than some of the leading models by US structure design service providers. Benchmarks show that DeepSeek's R1 design performs on par or better than leading, more familiar designs like OpenAI's o1 and Anthropic's Claude 3.5 Sonnet. DeepSeek was trained at a considerably lower cost-but not to the level that preliminary news recommended. Initial reports suggested that the training expenses were over $5.5 million, but the real value of not just training however developing the model overall has actually been debated because its release. According to semiconductor research and consulting firm SemiAnalysis, the $5.5 million figure is just one component of the expenses, leaving out hardware costs, the wages of the research study and development team, and other elements. DeepSeek's API pricing is over 90% cheaper than OpenAI's. No matter the true expense to develop the design, DeepSeek is using a more affordable proposal for utilizing its API: input and output tokens for DeepSeek R1 cost $0.55 per million and $2.19 per million, respectively, compared to OpenAI's $15 per million and $60 per million for its o1 design. DeepSeek R1 is an ingenious design. The associated scientific paper launched by DeepSeekshows the methods utilized to establish R1 based upon V3: leveraging the mix of professionals (MoE) architecture, reinforcement learning, and extremely innovative hardware optimization to produce models requiring less resources to train and likewise less resources to carry out AI inference, leading to its abovementioned API usage costs. DeepSeek is more open than the majority of its rivals. DeepSeek R1 is available free of charge on platforms like HuggingFace or GitHub. While DeepSeek has actually made its weights available and supplied its training methods in its research paper, the original training code and information have not been made available for a competent individual to develop an equivalent design, consider defining an open-source AI system according to the Open Source Initiative (OSI). Though DeepSeek has been more open than other GenAI business, R1 remains in the open-weight category when considering OSI standards. However, the release triggered interest outdoors source community: Hugging Face has released an Open-R1 effort on Github to develop a complete reproduction of R1 by constructing the "missing pieces of the R1 pipeline," moving the model to completely open source so anybody can recreate and construct on top of it. DeepSeek released effective small models alongside the major R1 release. DeepSeek launched not only the major big design with more than 680 billion parameters but also-as of this article-6 distilled models of DeepSeek R1. The models range from 70B to 1.5 B, the latter fitting on many consumer-grade hardware. As of February 3, 2025, the models were downloaded more than 1 million times on HuggingFace alone. DeepSeek R1 was potentially trained on OpenAI's information. On January 29, 2025, reports shared that Microsoft is examining whether DeepSeek used OpenAI's API to train its designs (an offense of OpenAI's terms of service)- though the hyperscaler also included R1 to its Azure AI Foundry service.
Understanding the generative AI value chain

GenAI spending advantages a broad market worth chain. The graphic above, based on research study for IoT Analytics' Generative AI Market Report 2025-2030 (released January 2025), portrays crucial beneficiaries of GenAI spending throughout the worth chain. Companies along the value chain include:

The end users - End users consist of customers and companies that use a Generative AI application. GenAI applications - Software vendors that include GenAI features in their products or deal standalone GenAI software. This includes business software application companies like Salesforce, with its focus on Agentic AI, and start-ups particularly concentrating on GenAI applications like Perplexity or Lovable. Tier 1 recipients - Providers of foundation models (e.g., OpenAI or Anthropic), design management platforms (e.g., AWS Sagemaker, Google Vertex or AI), information management tools (e.g., MongoDB or Snowflake), cloud computing and information center operations (e.g., Azure, AWS, Equinix or Digital Realty), AI consultants and integration services (e.g., Accenture or Capgemini), and edge computing (e.g., Advantech or HPE). Tier 2 recipients - Those whose product or wiki.lafabriquedelalogistique.fr services regularly support tier 1 services, consisting of companies of chips (e.g., NVIDIA or AMD), network and server devices (e.g., Arista Networks, Huawei or Belden), server cooling technologies (e.g., Vertiv or Schneider Electric). Tier 3 beneficiaries - Those whose items and services frequently support tier 2 services, such as service providers of electronic design automation software service providers for chip style (e.g., Cadence or Synopsis), semiconductor fabrication (e.g., TSMC), heat exchangers for cooling innovations, and electrical grid technology (e.g., Siemens Energy or ABB). Tier 4 beneficiaries and beyond - Companies that continue to support the tier above them, such as lithography systems (tier-4) necessary for semiconductor fabrication machines (e.g., AMSL) or business that offer these providers (tier-5) with lithography optics (e.g., Zeiss).
Winners and losers along the generative AI worth chain

The increase of designs like DeepSeek R1 signals a prospective shift in the generative AI worth chain, challenging existing market characteristics and improving expectations for success and competitive advantage. If more designs with similar capabilities emerge, certain gamers might benefit while others face increasing pressure.

Below, IoT Analytics evaluates the essential winners and most likely losers based upon the developments introduced by DeepSeek R1 and the wider trend towards open, cost-efficient designs. This assessment considers the possible long-term impact of such designs on the worth chain rather than the immediate results of R1 alone.

Clear winners

End users

Why these developments are positive: The availability of more and less expensive models will ultimately decrease costs for the end-users and make AI more available. Why these developments are unfavorable: No clear argument. Our take: DeepSeek represents AI development that eventually benefits the end users of this innovation.
GenAI application providers

Why these developments are positive: Startups building applications on top of structure models will have more choices to select from as more designs come online. As mentioned above, DeepSeek R1 is without a doubt cheaper than OpenAI's o1 model, and though thinking designs are seldom used in an application context, it shows that ongoing advancements and innovation enhance the designs and make them cheaper. Why these innovations are unfavorable: No clear argument. Our take: The availability of more and cheaper models will ultimately lower the expense of including GenAI functions in applications.
Likely winners

Edge AI/edge computing companies

Why these innovations are positive: During Microsoft's recent incomes call, Satya Nadella explained that "AI will be far more ubiquitous," as more workloads will run locally. The distilled smaller models that DeepSeek launched together with the powerful R1 model are little sufficient to operate on many edge devices. While small, the 1.5 B, 7B, and 14B designs are also comparably powerful reasoning models. They can fit on a laptop computer and other less effective devices, e.g., IPCs and smfsimple.com industrial gateways. These distilled designs have already been downloaded from Hugging Face hundreds of thousands of times. Why these developments are negative: No clear argument. Our take: The distilled models of DeepSeek R1 that fit on less powerful hardware (70B and below) were downloaded more than 1 million times on HuggingFace alone. This reveals a strong interest in releasing designs in your area. Edge computing makers with edge AI services like Italy-based Eurotech, and Taiwan-based Advantech will stand to profit. Chip companies that concentrate on edge computing chips such as AMD, ARM, Qualcomm, or perhaps Intel, may also benefit. Nvidia also operates in this market section.
Note: IoT Analytics' SPS 2024 Event Report (published in January 2025) dives into the most recent industrial edge AI patterns, as seen at the SPS 2024 fair in Nuremberg, Germany.

Data management services companies

Why these developments are favorable: There is no AI without data. To establish applications utilizing open models, adopters will need a variety of data for training and during release, needing proper information management. Why these innovations are unfavorable: No clear argument. Our take: Data management is getting more vital as the variety of different AI models boosts. Data management business like MongoDB, Databricks and Snowflake as well as the particular offerings from hyperscalers will stand to earnings.
GenAI providers

Why these innovations are favorable: The sudden development of DeepSeek as a top gamer in the (western) AI ecosystem reveals that the complexity of GenAI will likely grow for some time. The higher availability of various designs can lead to more complexity, driving more demand for services. Why these innovations are negative: When leading designs like DeepSeek R1 are available for complimentary, the ease of experimentation and implementation may restrict the need for combination services. Our take: As brand-new innovations pertain to the market, GenAI services need increases as enterprises try to understand how to best use open models for their business.
Neutral

Cloud computing suppliers

Why these developments are favorable: Cloud players hurried to consist of DeepSeek R1 in their model management platforms. Microsoft included it in their Azure AI Foundry, and AWS enabled it in Amazon Bedrock and Amazon Sagemaker. While the hyperscalers invest greatly in OpenAI and Anthropic (respectively), they are likewise model agnostic and allow hundreds of various designs to be hosted natively in their model zoos. Training and fine-tuning will continue to take place in the cloud. However, online-learning-initiative.org as models end up being more effective, less financial investment (capital investment) will be required, which will increase earnings margins for hyperscalers. Why these developments are unfavorable: More models are anticipated to be deployed at the edge as the edge ends up being more powerful and designs more efficient. Inference is likely to move towards the edge moving forward. The cost of training innovative designs is likewise anticipated to decrease further. Our take: Smaller, more effective models are becoming more crucial. This lowers the need for effective cloud computing both for training and reasoning which may be offset by higher total demand and lower CAPEX requirements.
EDA Software companies

Why these innovations are positive: Demand for new AI chip styles will increase as AI work become more specialized. EDA tools will be important for creating efficient, smaller-scale chips tailored for edge and distributed AI reasoning Why these innovations are unfavorable: The approach smaller sized, less resource-intensive models may decrease the demand for designing advanced, high-complexity chips optimized for enormous data centers, potentially leading to minimized licensing of EDA tools for high-performance GPUs and ASICs. Our take: EDA software companies like Synopsys and Cadence might benefit in the long term as AI specialization grows and drives demand for new chip designs for edge, consumer, and inexpensive AI workloads. However, the market may require to adjust to shifting requirements, focusing less on large data center GPUs and more on smaller, effective AI hardware.
Likely losers

AI chip companies

Why these innovations are positive: The apparently lower training expenses for models like DeepSeek R1 might eventually increase the overall need for AI chips. Some referred to the Jevson paradox, the idea that performance causes more demand for a resource. As the training and inference of AI designs end up being more effective, the demand could increase as greater effectiveness causes decrease expenses. ASML CEO Christophe Fouquet shared a similar line of thinking: "A lower cost of AI could imply more applications, more applications suggests more demand in time. We see that as an opportunity for more chips need." Why these innovations are negative: The allegedly lower expenses for DeepSeek R1 are based mainly on the requirement for less advanced GPUs for training. That puts some doubt on the sustainability of large-scale tasks (such as the recently announced Stargate project) and the capital investment spending of tech business mainly allocated for purchasing AI chips. Our take: IoT Analytics research study for its latest Generative AI Market Report 2025-2030 (released January 2025) discovered that NVIDIA is leading the data center GPU market with a market share of 92%. NVIDIA's monopoly defines that market. However, that likewise demonstrates how strongly NVIDA's faith is linked to the ongoing growth of spending on data center GPUs. If less hardware is required to train and release designs, then this might seriously deteriorate NVIDIA's development story.
Other categories connected to data centers (Networking devices, electrical grid technologies, electrical power companies, and heat exchangers)

Like AI chips, models are most likely to become cheaper to train and more efficient to deploy, so the expectation for further information center facilities build-out (e.g., networking devices, cooling systems, and power supply services) would decrease accordingly. If less high-end GPUs are needed, large-capacity data centers may scale back their investments in associated facilities, possibly impacting need for supporting innovations. This would put pressure on business that offer important parts, most especially networking hardware, power systems, and cooling services.

Clear losers

Proprietary model suppliers

Why these developments are favorable: No clear argument. Why these developments are unfavorable: The GenAI companies that have gathered billions of dollars of financing for their proprietary models, such as OpenAI and Anthropic, stand to lose. Even if they establish and release more open designs, this would still cut into the revenue circulation as it stands today. Further, while some framed DeepSeek as a "side project of some quants" (quantitative analysts), the release of DeepSeek's effective V3 and then R1 designs proved far beyond that sentiment. The concern moving forward: What is the moat of exclusive model providers if cutting-edge designs like DeepSeek's are getting released for free and become fully open and fine-tunable? Our take: DeepSeek launched powerful models for totally free (for regional release) or really low-cost (their API is an order of magnitude more budget friendly than similar models). Companies like OpenAI, Anthropic, and Cohere will deal with significantly strong competition from players that launch totally free and personalized advanced models, like Meta and DeepSeek.
Analyst takeaway and outlook

The emergence of DeepSeek R1 enhances a crucial trend in the GenAI area: open-weight, cost-efficient models are becoming feasible competitors to exclusive alternatives. This shift challenges market presumptions and forces AI suppliers to reconsider their value propositions.

1. End users and GenAI application companies are the greatest winners.

Cheaper, top quality models like R1 lower AI adoption expenses, benefiting both enterprises and customers. Startups such as Perplexity and Lovable, which construct applications on structure designs, now have more options and can significantly reduce API expenses (e.g., R1's API is over 90% cheaper than OpenAI's o1 design).

2. Most professionals agree the stock exchange overreacted, but the development is real.

While major AI stocks dropped dramatically after R1's release (e.g., NVIDIA and Microsoft down 18% and 7.5%, respectively), numerous experts see this as an overreaction. However, DeepSeek R1 does mark a real advancement in expense effectiveness and openness, setting a precedent for future competition.

3. The dish for constructing top-tier AI designs is open, accelerating competition.

DeepSeek R1 has actually shown that launching open weights and a detailed approach is helping success and caters to a growing open-source neighborhood. The AI landscape is continuing to move from a few dominant proprietary players to a more competitive market where brand-new entrants can construct on existing advancements.

4. Proprietary AI providers deal with increasing pressure.

Companies like OpenAI, Anthropic, and Cohere should now differentiate beyond raw model performance. What remains their competitive moat? Some may move towards enterprise-specific services, while others might check out hybrid service models.

5. AI facilities providers deal with blended potential customers.

Cloud computing providers like AWS and Microsoft Azure still gain from design training but face pressure as reasoning relocate to edge devices. Meanwhile, AI chipmakers like NVIDIA could see weaker need for high-end GPUs if more models are trained with fewer resources.

6. The GenAI market remains on a strong development path.

Despite disruptions, AI costs is expected to expand. According to IoT Analytics' Generative AI Market Report 2025-2030, international costs on foundation designs and platforms is forecasted to grow at a CAGR of 52% through 2030, driven by business adoption and continuous effectiveness gains.

Final Thought:

DeepSeek R1 is not simply a technical milestone-it signals a shift in the AI market's economics. The dish for constructing strong AI designs is now more widely available, ensuring greater competitors and faster innovation. While exclusive designs need to adapt, AI application service providers and end-users stand to benefit most.

Disclosure

Companies mentioned in this article-along with their products-are used as examples to showcase market developments. No company paid or got preferential treatment in this short article, and it is at the discretion of the analyst to choose which examples are used. IoT Analytics makes efforts to differ the business and items discussed to help shine attention to the various IoT and associated innovation market gamers.

It is worth noting that IoT Analytics might have commercial relationships with some companies pointed out in its posts, as some business accredit IoT Analytics market research. However, for confidentiality, IoT Analytics can not disclose private relationships. Please contact compliance@iot-analytics.com for any concerns or issues on this front.

More details and more reading

Are you interested in discovering more about Generative AI?

Generative AI Market Report 2025-2030

A 263-page report on the business Generative AI market, incl. market sizing & forecast, competitive landscape, end user adoption, trends, obstacles, and more.

Download the sample to learn more about the report structure, choose definitions, choose data, additional information points, patterns, and more.

Already a customer? View your reports here →

Related posts

You might likewise have an interest in the following articles:

AI 2024 in evaluation: The 10 most significant AI stories of the year What CEOs discussed in Q4 2024: Tariffs, reshoring, and agentic AI The industrial software application market landscape: 7 key stats going into 2025 Who is winning the cloud AI race? Microsoft vs. AWS vs. Google
Related publications

You may likewise have an interest in the following reports:

Industrial Software Landscape 2024-2030 Smart Factory Adoption Report 2024 Global Cloud Projects Report and Database 2024
Sign up for our newsletter and follow us on LinkedIn to remain updated on the most recent patterns forming the IoT markets. For complete business IoT protection with access to all of IoT Analytics' paid material & reports, consisting of dedicated analyst time, check out the Enterprise membership.
Assignee
Assign to
Time tracking