Lock in your current rates now! ⭐ Price update planned for January.
EN
Currency:
EUR – €
Choose a currency
  • Euro EUR – €
  • United States dollar USD – $
VAT:
OT 0%
Choose your country (VAT)
  • OT All others 0%
22.08.2025

New LLM Models Available on HOSTKEY GPU Servers

server one

When ordering a GPU server, you can now select pre-installed LLM models or deploy them later through your client portal.

Qwen3:32B

The flagship model of Alibaba’s third-generation Qwen family.

  • Size: 32 billion parameters, ensuring high accuracy and strong performance on complex tasks.

  • Capabilities: natural language generation, document analysis, advanced reasoning and context-rich responses.

  • Use cases: enterprise-grade chatbots, multilingual support and business process automation.

  • Key advantage: superior reasoning accuracy and extended context handling.

Order

Qwen3-Coder

A specialized Qwen model optimized for software development tasks.

  • Purpose: code generation, explanation, auto-completion and error correction.

  • Supported languages: Python, C++, Java, Go, JavaScript and more.

  • Use cases: accelerating software development, integrating into IDEs, and supporting DevOps workflows.

  • Key advantage: enhanced syntax handling and analysis of large code bases.

Order

GPT-OSS-20B

An open-source LLM with 20 billion parameters designed for versatile applications.

  • Balance: strong performance with moderate compute requirements.

  • Capabilities: text generation, question answering and conversational AI.

  • Use cases: AI assistants, chatbots, information retrieval and content automation.

  • Key advantage: fully open-source, highly customizable and easy to integrate.

Order

In addition to the new models, HOSTKEY also offers GPU servers with pre-installed DeepSeek-r1-14b, Gemma-3-27b-it, Llama-3.3-70B, Phi-4-14b

Advantages of Running LLMs on Your Own GPU Servers

  1. Full control – manage your infrastructure, execution environment and security policies.

  2. Data privacy – your data stays on your servers, without being transferred to third-party providers.

  3. Flexibility – choose the right model for your workload: general-purpose, code-focused or domain-specific.

  4. Scalability – run multiple models in parallel or distribute workloads across GPUs.

  5. No external limits – no request caps or vendor-imposed restrictions.

You can order an LLM model today either when configuring your GPU server or directly through your client portal.

Other news

23.12.2025

Thank You for Being with HOSTKEY in 2025. Wishing You a Merry Christmas and a Happy New Year

As 2025 comes to an end, we want to thank you for your trust and for choosing HOSTKEY. It has been a pleasure supporting your projects this year, and we look forward to your continued success in the year ahead.

15.12.2025

HOSTKEY Establishes Private Network Interconnects (PNIs) with Meta in Amsterdam and Helsinki

HOSTKEY has deployed new 100G Private Network Interconnects with Meta in Amsterdam and Helsinki, delivering faster, more reliable connectivity for customers using our dedicated, GPU and virtual servers.

09.12.2025

Price update planned for January: lock in your current rates now

Prices are going up in January, but you can still lock in your current rates and save up to 30% by paying now for a longer term before the end of December.

27.11.2025

EPYC Black Friday Deals

One day only. Huge Black Friday savings on EPYC and Ryzen servers. Grab top hardware at the best prices of the year.

05.11.2025

Up to 45% OFF on 4th Gen AMD EPYC Dedicated Servers

EPYC Week is here! Save up to 45% on blazing-fast 4th Gen AMD EPYC dedicated servers. Perfect for virtualization, analytics, and demanding workloads — offer ends November 11th!

Upload