Phi-4-14b¶
In this article
Information
Phi4-14b is a modern language model adapted for local use through the Ollama platform. This solution combines the performance of the Phi-4 model with the convenience of using Open Web UI. The system is deployed on Ubuntu 22.04 and requires a modern graphical accelerator for efficient operation. The combination of Ollama and Open Web UI creates a full-fledged local solution for working with artificial intelligence, providing complete control over data and the request processing process.
Deployment Features¶
-
Powerful Base Architecture: the model has 14 billion parameters and a context window of 16 thousand tokens, allowing it to process long texts and complex tasks with high accuracy;
-
Convenient Open Web UI Interface: provides a modern web interface for interacting with the model, including chat history, parameter settings, and management of different models in a single interface;
-
Advanced RAG Capabilities: built-in support for searching documents in various formats (PDF, PPT, text files) and web scraping, allowing the model to work with current information and user data;
-
Multitasking: the model efficiently handles various tasks, including dialogue management, answering questions, data analysis, text creation, and basic programming;
-
Security and Accuracy: the model underwent multi-stage training considering user preferences under expert supervision, ensuring high accuracy of responses, adherence to ethical norms, and following user instructions;
-
Openness and Flexibility: thanks to Ollama's open source code, the system can be configured and modified for specific needs, including integration with other tools and services;
-
Support for Various Data Types: the model is capable of working with text data, code, structured and semi-structured data, expanding its applicability in various fields.
-
Examples of Use:
-
Customer Support: automation of responses to user questions;
-
Education: creation of educational materials, assistance in solving tasks;
-
Marketing: generation of advertising texts, analysis of reviews;
-
Software Development: creation and documentation of code.
-
Deployment Features¶
ID | Compatible OS | VM | BM | VGPU | GPU | Min CPU (Cores) | Min RAM (Gb) | Min HDD/SDD (Gb) | Active |
---|---|---|---|---|---|---|---|---|---|
256 | Ubuntu 22.04 | - | - | + | + | 4 | 16 | - | Yes |
-
Installation time is 15-30 minutes together with OS;
-
Ollama server loads and runs LLM in memory;
-
Open WebUI is deployed as a web application connected to the Ollama server;
-
Users interact with LLM through the web interface of Open WebUI, sending requests and receiving responses;
-
All computations and data processing occur locally on the server. Administrators can configure LLM for specific tasks using OpenWebUI tools.
System Requirements and Technical Characteristics¶
-
Graphical Accelerator: NVIDIA A4000 (or more powerful) or an AMD equivalent with at least 16 GB of video memory;
-
Disk Space: sufficient for installing the system, drivers, and model;
-
Drivers: NVIDIA drivers and CUDA for correct GPU operation;
-
Video Memory Consumption: 12 GB at a context of 2K tokens;
-
Automatic Restart: automatic container restart is configured in case of failures;
-
GPU Support: full integration with NVIDIA CUDA for maximum performance.
Getting Started After Deploying Phi-4-14b¶
After payment, an order notification will be sent to the email specified during registration indicating that the server is ready for use. It will include the VPS IP address, as well as the login and password for connecting to the server and a link to access the OpenWebUI panel. Our company's clients manage equipment in the server management and API control panel — Invapi.
-
Login Data for Accessing the Server's Operating System (e.g., via SSH) will be sent to you in the received email.
-
Link for Access to Ollama Management Panel with Open WebUI Web Interface: in the tag webpanel in the tab Info >> Tags of the Invapi control panel. The exact link in the format
https://phi<Server_ID_from_Invapi>.hostkey.in
will be sent in the email upon server delivery.
After following the link from the tag webpanel, an OpenWebUI login window will open where you need to create an Administrator account by setting its Name, Login, and Password.
Note
A detailed description of the features of working with the Ollama management panel with Open WebUI can be found in the article AI Chatbot on Your Own Server
Note
It is recommended to use a GPU with more than the minimum requirement of 16 GB of video memory for optimal operation, providing headroom for processing large contexts and parallel requests. Detailed information about Ollama's main settings and Open WebUI can be found in Ollama developers' documentation and in Open WebUI developers' documentation.
Ordering a Server with Phi-4-14b Using API¶
To install this software using the API, follow these instructions.
Some of the content on this page was created or translated using AI.