When ordering a GPU server, you can now select pre-installed LLM models or deploy them later through your client portal.
The flagship model of Alibaba’s third-generation Qwen family.
Size: 32 billion parameters, ensuring high accuracy and strong performance on complex tasks.
Capabilities: natural language generation, document analysis, advanced reasoning and context-rich responses.
Use cases: enterprise-grade chatbots, multilingual support and business process automation.
Key advantage: superior reasoning accuracy and extended context handling.
A specialized Qwen model optimized for software development tasks.
Purpose: code generation, explanation, auto-completion and error correction.
Supported languages: Python, C++, Java, Go, JavaScript and more.
Use cases: accelerating software development, integrating into IDEs, and supporting DevOps workflows.
Key advantage: enhanced syntax handling and analysis of large code bases.
An open-source LLM with 20 billion parameters designed for versatile applications.
Balance: strong performance with moderate compute requirements.
Capabilities: text generation, question answering and conversational AI.
Use cases: AI assistants, chatbots, information retrieval and content automation.
Key advantage: fully open-source, highly customizable and easy to integrate.
In addition to the new models, HOSTKEY also offers GPU servers with pre-installed DeepSeek-r1-14b, Gemma-3-27b-it, Llama-3.3-70B, Phi-4-14b
Advantages of Running LLMs on Your Own GPU Servers
Full control – manage your infrastructure, execution environment and security policies.
Data privacy – your data stays on your servers, without being transferred to third-party providers.
Flexibility – choose the right model for your workload: general-purpose, code-focused or domain-specific.
Scalability – run multiple models in parallel or distribute workloads across GPUs.
No external limits – no request caps or vendor-imposed restrictions.
You can order an LLM model today either when configuring your GPU server or directly through your client portal.