Free and self-hosted AI Chatbot built on Ollama, Lllama3 LLM model and OpenWebUI interface.
Personal chatbot powered by Ollama, an open source large language model Lllama3 and OpenWebUI interface running on your own server.
Rent a virtual (VPS) or a dedicated server from HOSTKEY with a pre-installed and ready-to-use Self-hosted AI Chatbot, which can process your data and documents using various LLM models.
You can upload the most recent versions of Phi3, Mistral, Gemma, and Code Llama models.
Servers available in the Netherlands, Finland, Germany and Iceland.
AI Chatbot running on a HOSTKEY-shared GPU server. You get the admin rights to use AI Chatbot privately and manage your team.
AI Chatbot pre-installed on your own VPS or Dedicated GPU server. You get the full admin rights over your personal server.
With every AI Chatbot plan, you will get pre-installed and ready-to-use models:
“Self-hosted AI Chatbot” plan lets you manage and load more models. This feature is not available for the “AI Chatbot Lite” plan.
AI Chatbot Lite - a paid trial tier of the main product
AI Chatbot Lite runs on a HOSTKEY-shared GPU server. This is a paid trial tier of the main product (Self-hosted AI Chatbot). The payment is taken for using a shared HOSTKEY-managed GPU server. AI Chatbot Lite offers basic chatbot functionality with no limits on the number of users or prompt requests. It makes AI Chatbot Lite a perfect trial plan before purchasing the Self-hosted AI Chatbot (main product).
An AI chatbot hosted on your own server ensures security, with all data stored and processed within your environment.
OpenWebUI allows you to use multiple models in a single workspace, speeding up workflow and enabling you to process a single prompt across several models simultaneously without switching windows. Just make one request and visually compare the results to select the best option.
Hosting an AI Chatbot on your own server allows you to conduct code reviews preventing the risk of code leakage and protecting corporate confidentiality. The scalability of a self-hosted solution is very useful for handling increased workloads – you can increase a server capacity as needed.
Connect an AI chatbot to your knowledge base to generate answers to frequently asked questions from your users or employees.
Analyze corporate documents in various formats, such as pdf, csv, rst, xml, md, epub, doc, docx, xls, xlsx, ppt, and txt, while keeping all analytics on your server to ensure confidentiality.
Personal chatbot powered by Ollama, an open source large language model Lllama3 and OpenWebUI interface - a HOSTKEY solution built on officially free and open-source software.
Ollama is an open-source project that serves as a powerful and user-friendly platform for running LLMs on your local machine. Ollama is licensed under the MIT License.
Lllama3 is Meta's latest open-source large language model that has been scaled up to 70 billion parameters, making it one of the largest and most powerful language models in the world.
Open WebUI is an extensible, feature-rich, and user-friendly self-hosted WebUI designed to operate entirely offline. OpenWebUI is licensed under the MIT License.
We guarantee that our servers are running safe and original software.
To install AI Сhatbot, you need to select it in the "App Marketplace" tab (AI and Machine Learning group) while ordering a server on the HOSTKEY website. Our auto-deployment system will install the software on your server.
“Self-hosted AI Chatbot” runs on your personal server rented at HOSTKEY. It provides full chatbot functionality with no limits on the number of users or prompt requests.
“AI Chatbot Lite” is a paid trial tier of the main product, which runs on a shared HOSTKEY-managed GPU server. AI Chatbot Lite offers basic chatbot functionality with no limits on the number of users or prompt requests, making it a perfect trial plan before purchasing the Self-hosted AI Chatbot (main product).
The payment for AI Chatbot Lite covers the use of a shared HOSTKEY-managed GPU server, meaning you don’t need to rent a server to use the software, which is free by itself.
You just need to order a new one and cancel the previous one. Also you can use both of them simultaneously as they operate independently from each other.
Yes, with the “AI-chatbot Lite” plan, you can interact with the chatbot using API requests. To set it up, follow these instructions:
https://docs.openwebui.com/getting-started/advanced-topics/api-endpoints
https://github.com/ollama/ollama/blob/main/docs/api.md
The RAG (Retrieval Augmented Generation) feature allows the chatbot to generate responses based on specific data, such as your documentation. Unfortunately, this feature is not available on the “AI Chatbot Lite” trial plan.
To access this feature, we recommend upgrading to the “AI Chatbot on your own server” plan.
Learn more about adding documents to the knowledge base (RAG):
https://hostkey.com/documentation/marketplace/machine_learning/ai_chatbot/#adding-documents-to-the-knowledge-base-rag
The "AI Chatbot Lite" plan offers 2 models pre-installed by default: gemma2:latest 9.2B and llama3:latest 8.0B. You can use them either separately or simultaneously. The "AI Chatbot Lite" plan does not allow you to install or delete models.
The "AI Chatbot on Your Own Server" plan also includes 2 pre-installed models by default: gemma2:latest 9.2B and llama3:latest 8.0B. However, with this plan, you can install and delete any models available in the Ollama library
A self-hosted AI Chatbot has a number of advantages compared to popular paid services:
Picking between an AI chatbot that you host and a SaaS platform like ChatGPT or Gemini is very important when selecting an AI assistant. By using the Ollama chatbot, you’ll avoid being tied to any vendor, handle your infrastructure yourself and enjoy top-level privacy.
Although ChatGPT is a powerful tool, your data is stored and processed outside your reach. In contrast, running the Ollama chatbot on your own hardware or using an AI chatbot VPS allows you to keep all data local. It is especially important in industries that are regulated such as healthcare and finance.
Google Gemini AI depends a lot on its ecosystem. You may not always be able to see or change how things are done. Using Ollama hosting, you can modify or swap models and set up your stack as you wish, without being limited by the ecosystem.
GitHub Copilot is focused on code generation but lacks flexibility for broader AI interactions.щ With the Ollama chatbot, you can have more conversations and control how it reasons, remembers things and works with other systems.
When you run your Ollama chatbot on your own VPS or server, you gain benefits right away and in the future.
If you’re working on AI apps, the Ollama chatbot is the perfect partner for you.
If you’re working on agents, integrating vector databases or making changes to the UI, a self-hosted AI chatbot allows you to do things that cloud platforms cannot.
The system is designed to provide high performance for ollama hosting.
Latency benchmarks:
The difference between VPS and dedicated server.
Ollama chatbots are being used by many industries to make things more efficient and improve how users interact.
It is very important to have good security. When you use a self-hosted AI chatbot, you don’t have to give up anything.
It takes only a few minutes to deploy any server. You can select from free or discounted AI software that is already installed on our marketplace. Customers can choose to be billed hourly or monthly. Save up to 40% when you choose a long-term plan.
AI-Dedicated-1
AI-Dedicated-2
AI-Dedicated-Pro
AI-VPS-Lite
AI-VPS-Pro
AI-VPS-Max
All of our servers are set up and ready for you to use right away. You can get the Ollama chatbot, Docker stacks and other AI tools directly from our marketplace.
Manage your AI infrastructure entirely on your own. Sign up for Ollama hosting, run your self-hosted AI chatbot on a strong Ollama VPS or Ollama server and start in just a few minutes.