Lock in your current rates now! ⭐ Price update planned for January.
EN
Currency:
EUR – €
Choose a currency
  • Euro EUR – €
  • United States dollar USD – $
VAT:
OT 0%
Choose your country (VAT)
  • OT All others 0%
22.08.2025

New LLM Models Available on HOSTKEY GPU Servers

server one

When ordering a GPU server, you can now select pre-installed LLM models or deploy them later through your client portal.

Qwen3:32B

The flagship model of Alibaba’s third-generation Qwen family.

  • Size: 32 billion parameters, ensuring high accuracy and strong performance on complex tasks.

  • Capabilities: natural language generation, document analysis, advanced reasoning and context-rich responses.

  • Use cases: enterprise-grade chatbots, multilingual support and business process automation.

  • Key advantage: superior reasoning accuracy and extended context handling.

Order

Qwen3-Coder

A specialized Qwen model optimized for software development tasks.

  • Purpose: code generation, explanation, auto-completion and error correction.

  • Supported languages: Python, C++, Java, Go, JavaScript and more.

  • Use cases: accelerating software development, integrating into IDEs, and supporting DevOps workflows.

  • Key advantage: enhanced syntax handling and analysis of large code bases.

Order

GPT-OSS-20B

An open-source LLM with 20 billion parameters designed for versatile applications.

  • Balance: strong performance with moderate compute requirements.

  • Capabilities: text generation, question answering and conversational AI.

  • Use cases: AI assistants, chatbots, information retrieval and content automation.

  • Key advantage: fully open-source, highly customizable and easy to integrate.

Order

In addition to the new models, HOSTKEY also offers GPU servers with pre-installed DeepSeek-r1-14b, Gemma-3-27b-it, Llama-3.3-70B, Phi-4-14b

Advantages of Running LLMs on Your Own GPU Servers

  1. Full control – manage your infrastructure, execution environment and security policies.

  2. Data privacy – your data stays on your servers, without being transferred to third-party providers.

  3. Flexibility – choose the right model for your workload: general-purpose, code-focused or domain-specific.

  4. Scalability – run multiple models in parallel or distribute workloads across GPUs.

  5. No external limits – no request caps or vendor-imposed restrictions.

You can order an LLM model today either when configuring your GPU server or directly through your client portal.

Other news

09.12.2025

Price update planned for January: lock in your current rates now

Prices are going up in January, but you can still lock in your current rates and save up to 30% by paying now for a longer term before the end of December.

27.11.2025

EPYC Black Friday Deals

One day only. Huge Black Friday savings on EPYC and Ryzen servers. Grab top hardware at the best prices of the year.

05.11.2025

Up to 45% OFF on 4th Gen AMD EPYC Dedicated Servers

EPYC Week is here! Save up to 45% on blazing-fast 4th Gen AMD EPYC dedicated servers. Perfect for virtualization, analytics, and demanding workloads — offer ends November 11th!

25.10.2025

Get up to 40% off Ryzen servers this Halloween 2025!

Scary-good savings — up to 40% off popular AMD Ryzen servers!

22.10.2025

HOSTKEY Expands to Frankfurt: New Servers in firstcolo Data Center

We've launched a new server location in Frankfurt to meet growing demand in Germany! Our new setup at the firstcolo Data Center Werkhaus offers powerful and reliable hosting with enhanced connectivity.

Upload