top of page

Running DeepSeek R1 on Denvr Cloud with H100 GPUs for Enterprise-Grade AI

Chandan Kumar

Updated: Feb 11

Artificial Intelligence is advancing at an unprecedented pace, with models like DeepSeek R1 pushing the boundaries of reasoning and computation. Developed by the innovative Chinese startup DeepSeek, the DeepSeek R1 model is gaining global recognition for its exceptional performance in mathematics, coding, and logical reasoning. A unique aspect of DeepSeek R1 is its training methodology, which leverages reinforcement learning without relying on supervised fine-tuning. This novel approach has yielded a model that stands out in the crowded field of generative AI.

In this guide, we will explore how to run the highest-parameter version of DeepSeek R1 on Denvr Cloud, leveraging H100 GPUs for enterprise-grade use cases. We will also discuss when it might make sense to opt for a lower-parameter model on CPUs for cost-efficiency and less demanding scenarios.


Why Choose DeepSeek R1?


DeepSeek R1 differentiates itself through:

  • Advanced Reasoning Capabilities: It excels in tasks requiring complex reasoning, surpassing many proprietary models.

  • Open-Source Accessibility: DeepSeek R1 and its distilled versions are open-source, enabling researchers and developers to explore and deploy the model easily.

  • Scalability: The model offers versions with varying parameter sizes, catering to both lightweight and enterprise-grade applications.


With Denvr Cloud's high-performance infrastructure, deploying DeepSeek R1 becomes a seamless process for businesses looking to harness cutting-edge AI technology.


Why Use H100 GPUs for DeepSeek R1?


DeepSeek R1’s largest parameter model demands immense computational power to deliver real-time, accurate results. H100 GPUs provide the ideal platform for these workloads:


  • High Memory Bandwidth: Supports larger models and batch sizes, accelerating training and inference.

  • Enhanced Precision: Optimized for mixed-precision training and inference, reducing latency while maintaining accuracy.

  • Enterprise-Grade Reliability: Ensures stability and scalability for production environments.


For enterprise use cases such as:


  • Advanced analytics

  • Real-time decision-making

  • Complex reasoning tasks


the combination of DeepSeek R1’s capabilities and H100 GPUs’ power is unparalleled.


When to Choose Lower Parameter Models


While the highest-parameter model is ideal for enterprise-grade tasks, smaller parameter versions of DeepSeek R1 can run effectively on CPUs. This is a cost-effective option for:


  • Prototyping and development.

  • Lightweight applications where latency is not critical.

  • Scenarios with limited computational resources.


For developers or small teams, Denvr Cloud’s CPU offerings provide a budget-friendly way to experiment with DeepSeek R1.


How to Run DeepSeek R1 on Denvr Cloud


Prerequisites

Denvr Cloud Console login

  • Select the Virtual Machine


In today's data-driven world, accessing and analyzing data efficiently is crucial for decision-making. However, interacting with databases often requires knowledge of complex query languages like SQL, creating a barrier for non-technical users. To bridge this gap, we developed an AI-powered tool that enables users to interact with their databases using natural language.

Select the region ( Houston ) and H100 GPU, you can choose A100 ( either 80G or 40G option in our Calgary cluster )


  • Navigate to the GPU instances section.

  • Select an H100-based instance with at least 80GB GPU memory.



  • Select the Option with Nvidia + Docker Pre installed


Denvr Cloud VM option with Pre-installed CUDA + Docker
Denvr Cloud VM option with Pre-installed CUDA + Docker
  • Wait for VM to launch ( takes 5-7 minutes ) and then login to the instance




  • SSH into your instance: ssh username@your-instance-ip


ssh -i <your-ssh-key> ubuntu@<IP of your Denvr VM instance>

  • Update system packages


sudo apt update && sudo apt upgrade -y 

  • Once login check that all H100 GPUs are available



  • Launch Ollama docker container


docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama

  • Download and execute the deepseek model

 

deepseek R1 ollama container
Deepseek Ollama Container
  • Make sure GPUs are utilized for inferencing



  • Now start Inferencing with your choice of front end application or integrate it with any AI Agent


curl -X POST http://localhost/api/v1/completions \
-H "Content-Type: application/json" \
-d '{
"model": "deepseek-r1:671b",
"prompt": "What is AI"
}'

Conclusion


DeepSeek R1 signifies a significant advancement in AI capabilities, providing enhanced reasoning and scalability. Utilizing Denvr Cloud’s H100 GPU instances, businesses can fully harness the potential of the highest-parameter DeepSeek R1 model for challenging use cases. For less extensive applications, the models with fewer parameters offer a budget-friendly option.


Whether you're a startup investigating AI opportunities or a business seeking advanced solutions, DeepSeek R1 on Denvr Cloud offers a robust combination tailored to your requirements. Give it a try today and discover the future of AI-powered reasoning.

 
 
 

Bình luận


Contact Us

bottom of page