AWS, Microsoft, Google, Others Make DeepSeek-R1 AI Model Available on Their Platforms

Leading cloud service providers are now making the open source DeepSeek-R1 reasoning model available on their platforms. The Chinese startup generated intense interest for its ability to leverage more efficient processing and reduce compute resource consumption, which is a key driver of high AI costs.

Amazon Web Services (AWS), Microsoft, and Google Cloud have all made the model available to their customers, but as of this writing they had yet to implement the per-token pricing structure used for other AI models such as Meta's Llama 3.

Instead, DeepSeek-R1 users on these cloud platforms pay only for the computing resources they consume, rather than for the amount of text the model generates. AWS and Google have reported that this approach aligns with existing pricing models for open-source AI.

DeepSeek launched its latest DeepSeek-V3 model in December 2024. It was followed by the release of DeepSeek-R1, DeepSeek-R1-Zero, and DeepSeek-R1-Distill on Jan. 20, 2025. The DeepSeek-R1-Zero model reportedly features 671 billion parameters, and the DeepSeek-R1-Distill lineup offers models ranging from 1.5 billion to 70 billion parameters. On January 27, 2025, the company expanded its portfolio with Janus-Pro-7B, a vision-based AI model.

DeepSeek-R1 is positioned as a cost-efficient alternative to proprietary AI models, particularly for organizations with large-scale AI deployments. The model was designed to process information more efficiently, reducing the overall compute burden.

However, cloud providers may ultimately profit more from infrastructure rentals than direct model usage fees, industry watchers have observed. And renting cloud servers for AI workloads often costs more than accessing models via APIs. AWS, for example, charges up to $124 per hour for an AI-optimized cloud server, which translates to nearly $90,000 per month for continuous usage. Microsoft Azure customers do not need to rent dedicated servers for DeepSeek, but they still pay for underlying computing power, leading to variable pricing depending on how efficiently they run the model.

In contrast, organizations using Meta's Llama 3.1 through AWS pay $3 per 1 million tokens, a significantly lower upfront cost for those with intermittent AI needs. Tokens represent processed text, with 1,000 tokens equivalent to approximately 750 words, according to AI infrastructure provider Anyscale.

Smaller cloud providers, including Together AI and Fireworks AI, have already implemented fixed per-token pricing for DeepSeek-R1, a structure that could become more common as demand for cost-effective AI models grows.

For organizations seeking the lowest cost, DeepSeek-R1 is available via its parent company's API at $2.19 per million tokens — three to four times cheaper than some Western cloud providers. However, routing AI workloads through Chinese servers raises data privacy and security concerns. Sensitive business information could be subject to Chinese government regulations, including potential data sharing under local laws. And many organizations are cautious about sending proprietary or customer data to servers outside their jurisdiction, especially in regions with less stringent privacy protections.

AWS, Microsoft, and Google have not disclosed how many customers are actively using DeepSeek-R1.

About the Author

John K. Waters is the editor in chief of a number of Converge360.com sites, with a focus on high-end development, AI and future tech. He's been writing about cutting-edge technologies and culture of Silicon Valley for more than two decades, and he's written more than a dozen books. He also co-scripted the documentary film Silicon Valley: A 100 Year Renaissance, which aired on PBS.  He can be reached at [email protected].

Featured

  • stylized AI code and a neural network symbol, paired with glitching code and a red warning triangle

    New Anthropic AI Models Demonstrate Coding Prowess, Behavior Risks

    Anthropic has released Claude Opus 4 and Claude Sonnet 4, its most advanced artificial intelligence models to date, boasting a significant leap in autonomous coding capabilities while simultaneously revealing troubling tendencies toward self-preservation that include attempted blackmail.

  • university building with classical architecture is partially overlaid by a glowing digital brain graphic

    NSF Invests $100 Million in National AI Research Institutes

    The National Science Foundation has announced a $100 million investment in National Artificial Intelligence Research Institutes, part of a broader White House strategy to maintain American leadership as competition with China intensifies.

  • student reading a book with a brain, a protective hand, a computer monitor showing education icons, gears, and leaves

    4 Steps to Responsible AI Implementation

    Researchers at the University of Kansas Center for Innovation, Design & Digital Learning (CIDDL) have published a new framework for the responsible implementation of artificial intelligence at all levels of education.

  • cybersecurity book with a shield and padlock

    NIST Proposes New Cybersecurity Guidelines for AI Systems

    The National Institute of Standards and Technology has unveiled plans to issue a new set of cybersecurity guidelines aimed at safeguarding artificial intelligence systems, citing rising concerns over risks tied to generative models, predictive analytics, and autonomous agents.