DeepSeek and Ollama: Advanced AI on bare-metal servers and private cloud with Stackscale

Deploying DeepSeek on Dedicated Servers and Private Cloud

Generative AI and machine learning have revolutionized how businesses process and analyze data. DeepSeek-R1, an open-source AI model, stands out for its advanced reasoning capabilities, resource optimization, and security, as it runs locally. By combining Ollama with Stackscale’s infrastructure, organizations can deploy DeepSeek-R1 on bare-metal dedicated servers or private cloud environments, ensuring high performance and complete data sovereignty.

Why Use DeepSeek-R1 on Private Infrastructure?

1. Security and Privacy

DeepSeek-R1 enables local data processing, eliminating reliance on external servers and ensuring sensitive information remains protected.

2. Optimized Costs

Running DeepSeek-R1 on dedicated GPU servers from Stackscale reduces the need for cloud-based AI services, eliminating token-based fees or recurring monthly costs.

3. Enhanced Performance with NVIDIA GPUs

Stackscale’s infrastructure offers NVIDIA Tesla T4, L4, and L40S GPUs, designed to accelerate AI workloads, machine learning, and high-performance computing (HPC).

4. Flexibility and Scalability

Private cloud and bare-metal servers allow businesses to scale resources based on project demand, ensuring stability and full control over infrastructure.


Installing DeepSeek-R1 with Ollama

To deploy DeepSeek-R1 on dedicated servers or private cloud, Ollama is the recommended tool for managing local AI models.

Installation Steps:

1️⃣ Install Ollama

curl -fsSL https://ollama.com/install.sh | sh

2️⃣ Download DeepSeek-R1

ollama pull deepseek-r1:8b

(The 8B version provides a balance between performance and resource usage. Other versions are available depending on the GPU.)

3️⃣ Run DeepSeek-R1

ollama run deepseek-r1:8b

Once started, DeepSeek-R1 will be ready to process queries within the private environment, with no need for an internet connection.

Replace 8B with the desired model version:

  • 1.5B parameters: ollama run deepseek-r1:1.5b
  • 7B parameters: ollama run deepseek-r1
  • 70B parameters (Requires 24GB+ VRAM): ollama run deepseek-r1:70b
  • Full-scale 671B model: ollama run deepseek-r1:671b

Optimizing GPU Acceleration

  • Ensure NVIDIA CUDA drivers are installed.
  • Use ollama list to check installed models.
  • Start the service: ollama serve

Enhancing Performance with Open WebUI

For an improved user experience, Open WebUI provides a browser-based interface to interact with AI models running on Ollama. Key features include:
Model switching via @ commands.
Conversation tagging and management.
Easy model download and removal.


Optimizing Performance with NVIDIA GPUs on Stackscale

For maximum DeepSeek-R1 performance, using optimized GPUs is recommended. Stackscale offers:

GPUMemoryTensor CoresShading UnitsTFLOPS (FP32)
Tesla T416 GB GDDR63202,5608.1
L424 GB GDDR62407,02430.3
L40S48 GB GDDR658618,17691.6

These GPUs accelerate AI processing, reduce execution times, and optimize computing resources.


Benefits of Using Stackscale for AI and Machine Learning

100% European Infrastructure: Servers in Madrid and Amsterdam, ensuring full data sovereignty.
High Availability: 99.90% SLA, with redundant power supply and ultra-fast networks.
Complete Isolation: No resource oversubscription, ensuring dedicated high-performance computing.
24/7 Support: Specialized technical assistance in English and Spanish.


Conclusion

Deploying DeepSeek-R1 with Ollama on Stackscale’s private infrastructure enables businesses to access an optimized, secure AI environment with full data control. With high-end GPUs and an infrastructure designed for demanding workloads, Stackscale provides the ideal solution for AI and machine learning projects.

If you need more information about our dedicated GPU solutions and private cloud, contact us, and our team will help you configure the ideal setup for your project.

Share it on Social Media!

Cookies customization
Stackscale, Grupo Aire logo

By allowing cookies, you voluntarily agree to the processing of your data. This also includes, for a limited period of time, your consent in accordance with the Article 49 (1) (a) GDPR in regard to the processing of data outside the EEA, for instead, in the USA. In these countries, despite the careful selection and obligation of service providers, the European high level of data protection cannot be guaranteed.

In case of the data being transferred to the USA, there is, for instance, the risk of USA authorities processing that data for control and supervision purposes without having effective legal resources available or without being able to enforce all the rights of the interested party. You can revoke your consent at any moment.

Necessary Cookies

Necessary cookies help make a web page usable by activating basic functions such as the page navigation and the access to secure areas in the web page. The web page will not be able to work properly without these cookies. We inform you about the possibility to set up your browser in order to block or alert about these cookies, however, it is possible that certain areas of the web page do not work. These cookies do not store any personal data.

- moove_gdpr_popup

 

Analytical cookies

Analytical cookies allow its Editor to track and analyze the websites’ users behavior. The information collected through this type of cookie is used for measuring the activity on websites, applications or platforms, as well as for building user navigation profiles for said websites, application or platform, in order to implement improvements based on the analysis of data on the usage of the service by users.

Google Analytics: It registers a single identification used to generate statistical data about how the visitor uses the website. The data generated by the cookie about the usage of this website is generally transferred to a Google server in the USA and stored there by Google LLC, 1600 Amphitheatre Parkway Mountain View, CA 94043, USA.

- _dc_gtm_UA-XXXXXXXX-X

- _gat_gtag_UA_XXXXXXXX_X

- _ga

- _gcl_au

- _gid