AWS, Microsoft, Google, Others Make DeepSeek-R1 AI Model Available on Their Platforms

Leading cloud service providers are now making the open source DeepSeek-R1 reasoning model available on their platforms. The Chinese startup generated intense interest for its ability to leverage more efficient processing and reduce compute resource consumption, which is a key driver of high AI costs.

Amazon Web Services (AWS), Microsoft, and Google Cloud have all made the model available to their customers, but as of this writing they had yet to implement the per-token pricing structure used for other AI models such as Meta's Llama 3.

Instead, DeepSeek-R1 users on these cloud platforms pay only for the computing resources they consume, rather than for the amount of text the model generates. AWS and Google have reported that this approach aligns with existing pricing models for open-source AI.

DeepSeek launched its latest DeepSeek-V3 model in December 2024. It was followed by the release of DeepSeek-R1, DeepSeek-R1-Zero, and DeepSeek-R1-Distill on Jan. 20, 2025. The DeepSeek-R1-Zero model reportedly features 671 billion parameters, and the DeepSeek-R1-Distill lineup offers models ranging from 1.5 billion to 70 billion parameters. On January 27, 2025, the company expanded its portfolio with Janus-Pro-7B, a vision-based AI model.

DeepSeek-R1 is positioned as a cost-efficient alternative to proprietary AI models, particularly for organizations with large-scale AI deployments. The model was designed to process information more efficiently, reducing the overall compute burden.

However, cloud providers may ultimately profit more from infrastructure rentals than direct model usage fees, industry watchers have observed. And renting cloud servers for AI workloads often costs more than accessing models via APIs. AWS, for example, charges up to $124 per hour for an AI-optimized cloud server, which translates to nearly $90,000 per month for continuous usage. Microsoft Azure customers do not need to rent dedicated servers for DeepSeek, but they still pay for underlying computing power, leading to variable pricing depending on how efficiently they run the model.

In contrast, organizations using Meta's Llama 3.1 through AWS pay $3 per 1 million tokens, a significantly lower upfront cost for those with intermittent AI needs. Tokens represent processed text, with 1,000 tokens equivalent to approximately 750 words, according to AI infrastructure provider Anyscale.

Smaller cloud providers, including Together AI and Fireworks AI, have already implemented fixed per-token pricing for DeepSeek-R1, a structure that could become more common as demand for cost-effective AI models grows.

For organizations seeking the lowest cost, DeepSeek-R1 is available via its parent company's API at $2.19 per million tokens — three to four times cheaper than some Western cloud providers. However, routing AI workloads through Chinese servers raises data privacy and security concerns. Sensitive business information could be subject to Chinese government regulations, including potential data sharing under local laws. And many organizations are cautious about sending proprietary or customer data to servers outside their jurisdiction, especially in regions with less stringent privacy protections.

AWS, Microsoft, and Google have not disclosed how many customers are actively using DeepSeek-R1.

About the Author

John K. Waters is the editor in chief of a number of Converge360.com sites, with a focus on high-end development, AI and future tech. He's been writing about cutting-edge technologies and culture of Silicon Valley for more than two decades, and he's written more than a dozen books. He also co-scripted the documentary film Silicon Valley: A 100 Year Renaissance, which aired on PBS.  He can be reached at [email protected].

Featured

  •  laptop on a clean desk with digital padlock icon on the screen

    Study: Data Privacy a Top Concern as Orgs Scale Up AI Agents

    As organizations race to integrate AI agents into their cloud operations and business workflows, they face a crucial reality: while enthusiasm is high, major adoption barriers remain, according to a new Cloudera report. Chief among them is the challenge of safeguarding sensitive data.

  • glowing digital brain interacts with an open book, with stacks of books beside it

    Federal Court Rules AI Training with Copyrighted Books Fair Use

    A federal judge ruled this week that artificial intelligence company Anthropic did not violate copyright law when it used copyrighted books to train its Claude chatbot without author consent, but ordered the company to face trial on allegations it used pirated versions of the books.

  • college student using a laptop alongside an AI robot and academic icons like a graduation cap, lightbulb, and upward arrow

    Nonprofit to Pilot Agentic AI Tool for Student Success Work

    Student success nonprofit InsideTrack has joined Salesforce Accelerator – Agents for Impact, a Salesforce initiative providing technology, funding, and expertise to help nonprofits build and customize AI agents and AI-powered tools to support and scale their missions.

  • three main icons—a cloud, a user profile, and a padlock—connected by circuit lines on a blue abstract background

    Report: Identity Has Become a Critical Security Perimeter for Cloud Services

    A new threat landscape report points to new cloud vulnerabilities. According to the 2025 Global Threat Landscape Report from Fortinet, while misconfigured cloud storage buckets were once a prime vector for cybersecurity exploits, other cloud missteps are gaining focus.