4x RTX 4090 GPU Servers – Only €903/month with a 1-year rental! 🚀 BM EPYC 7402P, 384GB RAM, 2x3.84TB NVMe ⭐ Best Price on the Market!
Currency:
EUR – €
Choose a currency
  • Euro EUR – €
  • United States dollar USD – $
VAT:
OT 0%
Choose your country (VAT)
  • OT All others 0%
Choose a language
  • Choose a currency
    Choose you country (VAT)
    Dedicated Servers
  • Instant
  • Custom
  • Single CPU servers
  • Dual CPU servers
  • Servers with 4th Gen EPYC
  • Servers with AMD Ryzen and Intel Core i9
  • Storage Servers
  • Servers with 10Gbps ports
  • Hosting virtualization nodes
  • GPU
  • Sale
  • Virtual Servers
    GPU
  • Dedicated GPU server
  • VM with GPU
  • Tesla A100 80GB & H100 Servers
  • Nvidia RTX 5090
  • GPU servers equipped with AMD Radeon
  • Sale
    Apps
    Colocation
  • Colocation in the Netherlands
  • Remote smart hands
  • Services
  • L3-L4 DDoS Protection
  • Network equipment
  • IPv4 and IPv6 address
  • Managed servers
  • SLA packages for technical support
  • Monitoring
  • Software
  • VLAN
  • Announcing your IP or AS (BYOIP)
  • USB flash/key/flash drive
  • Traffic
  • Hardware delivery for EU data centers
  • AI Chatbot Lite
  • AI Platform
  • About
  • Careers at HOSTKEY
  • Server Control Panel & API
  • Data Centers
  • Network
  • Speed test
  • Hot deals
  • Sales contact
  • Reseller program
  • Affiliate Program
  • Grants for winners
  • Grants for scientific projects and startups
  • News
  • Our blog
  • Payment terms and methods
  • Legal
  • Abuse
  • Looking Glass
  • The KYC Verification
  • Hot Deals

    AI Platform

    Pre-installed AI LLM models on high-performance GPU instances

    • Already installed — just start using pre-installed LLM, wasting no time on deployment
    • Optimized server — high performance GPU configurations optimized for LLMs
    • Version Stability — you control the LLM version, having no unexpected changes or updates
    • Security and data privacy — all your data is stored and processed on your server, ensuring it never leaves your environment;
    • Transparent pricing — you only pay for the server rental; the operation and load of the neural network are not charged and are completely free.
    4.3/5
    4.8/5
    SERVERS In action right now 5 000+

    Top LLMs on high-performance GPU instances

    DeepSeek-r1-14b

    DeepSeek-r1-14b

    Open source LLM from China - the first-generation of reasoning models with performance comparable to OpenAI-o1.

    Gemma-2-27b-it

    Gemma-2-27b-it

    Google Gemma 2 is a high-performing and efficient model available in three sizes: 2B, 9B, and 27B.

    Llama-3.3-70B

    Llama-3.3-70B

    New state of the art 70B model. Llama 3.3 70B offers similar performance compared to the Llama 3.1 405B model.

    Phi-4-14b

    Phi-4-14b

    Phi-4 is a 14B parameter, state-of-the-art open model from Microsoft.

    AI & Machine Learning Tools

    PyTorch

    PyTorch

    PyTorch is a fully featured framework for building deep learning models.

    TensorFlow

    TensorFlow

    TensorFlow is a free and open-source software library for machine learning and artificial intelligence.

    Apache Spark

    Apache Spark

    Apache Spark is a multi-language engine for executing data engineering, data science, and machine learning on single-node machines or clusters.

    Anaconda

    Anaconda

    Open ecosystem for Data science and AI development.

    Choose among a wide range of GPU instances

    🚀
    4x RTX 4090 GPU Servers – Only €903/month with a 1-year rental! Best Price on the Market!
    GPU servers are available on both hourly and monthly payment plans. Read about how the hourly server rental works.

    The selected collocation region is applied for all components below.

    Region
    Cores/ GHz
    Performance
    RAM
    Storage
    Control panel
    Delivery ETA
    Price/mo

    Region
    Cores/ GHz
    Performance
    RAM
    Storage
    Control panel
    Delivery ETA
    Price/mo

    Region
    Cores/ GHz
    Performance
    RAM
    Storage
    Control panel
    Delivery ETA
    Price/mo

    Region
    Cores/ GHz
    Performance
    RAM
    Storage
    Control panel
    Delivery ETA
    Price/mo

    Self-hosted AI Chatbot:
    Pre-installed on your VPS or GPU server with full admin rights.

    LLMs and AI Solutions available

    Open-source LLMs

    • gemma-2-27b-it — Google Gemma 2 is a high-performing and efficient model available in three sizes: 2B, 9B, and 27B.
    • DeepSeek-r1-14b — Open source LLM from China - the first-generation of reasoning models with performance comparable to OpenAI-o1.
    • meta-llama/Llama-3.3-70B — New state of the art 70B model. Llama 3.3 70B offers similar performance compared to the Llama 3.1 405B model.
    • Phi-4-14b — Phi-4 is a 14B parameter, state-of-the-art open model from Microsoft.

    Image generation

    • ComfyUI — An open source, node-based program for image generation from a series of text prompts.

    AI Solutions, Frameworks and Tools

    • Self-hosted AI Chatbot — Free and self-hosted AI Chatbot built on Ollama, Lllama3 LLM model and OpenWebUI interface.
    • PyTorch — A fully featured framework for building deep learning models.
    • TensorFlow — A free and open-source software library for machine learning and artificial intelligence.
    • Apache Spark — A multi-language engine for executing data engineering, data science, and machine learning on single-node machines or clusters.
    Already installed
    We provide LLMs as a pre-installed software, saving you time on downloading and installation. Our auto-deployment system handles everything for you—simply place an order and start working in just 15 minutes.
    Optimized servers
    Our high-performance GPU servers are a perfect choice for working with LLMs. Rest assured, every LLM you choose will deliver top-tier performance on recommended servers.
    Version Stability
    If your software product runs on an LLM model, you will be happy to know that there will be no unexpected updates or version renewals. Your choice of LLM version will not change unpredictably.
    Transparent pricing
    At HOSTKEY you pay only for the server rental - no additional fees. All pre-installed LLMs come free with no limits on their usage. You have no restrictions on the number of tokens, the number of requests per unit of time, etc. - the price solely depends on the leased server capacity.
    Independence from IT service providers
    You can choose the most suitable neural network option from hundreds of open source LLMs. You can always install alternative models tailored to your needs. The version of the model used is completely controlled by you.
    Security and data privacy
    The LLM is deployed on our own server infrastructure, your data is completely protected and under your control. It is not shared or processed in the external environment.

    Get Top LLM models on high-performance GPU instances

    FAQ

    Can I rent a virtual server on an hourly basis?

    On an hourly basis you can rent the following servers:

    • preconfigured dedicated servers, including those with GPU cards;
    • virtual GPU servers;
    • VPS.

    Can I rent a virtual GPU server on an hourly basis?

    Yes, all virtual and preconfigured dedicated GPU servers are available for rent on an hourly basis.

    What payment methods are available?

    All our services are paid for in advance. We accept payments via credit card, PayPal, P2P cryptocurrency payments from any wallet, application or exchange through BitPay. We also accept WebMoney, Alipay and wire transfers. Read more about our payment terms and methods. Read more about payment terms and methods.

    What is your refund policy?

    We are very confident in our products and services. We provide fast, reliable and comprehensive service and believe that you will be completely satisfied.

    You can ask for a test server for 3-4 days for free.

    Refund is only possible in case of an accident from our side with your server being offline for 24 hours or more due to that.

    Read more about refund procedure.

    What is a Fair Usage Policy?

    Customers whose servers come with unlimited bandwidth are committed to a fair usage policy.

    That means that servers on the 1 Gbps port cannot use more than 70% of the allocated bandwidth for more than 3 hours a day.

    Upload