Ollama icon

Ollama

AI / LLMPopularMIT166k stars

Get up and running with large language models locally

Deploy with GoPanel:One-click deployAuto SSL certificatesScheduled backupsAny VPS provider

About Ollama

Get up and running with large language models locally. Deploy Ollama on your own server with GoPanel.

Alternative to

OpenAI APIOpenAI APIAzure OpenAIAzure OpenAI

Frequently asked questions

What models can I run with Ollama?
Ollama supports hundreds of open-source models including Llama 3, Mistral, Phi-3, Gemma, Code Llama, Qwen, and many more. You can pull models with a single command (ollama pull llama3) and run them locally.
How much RAM does Ollama need?
RAM requirements depend on the model size. 7B parameter models need ~8GB RAM, 13B models need ~16GB, and 70B models need ~64GB. For GPU acceleration, VRAM requirements are similar. Quantized models (Q4_0) use roughly half the RAM of full precision.
Does Ollama support GPU acceleration?
Yes, Ollama supports NVIDIA GPUs (CUDA), Apple Silicon (Metal), and AMD GPUs (ROCm). GPU acceleration dramatically improves inference speed — a 7B model runs 5-10x faster on a GPU compared to CPU-only inference.
Can I use Ollama with other applications?
Yes, Ollama exposes an OpenAI-compatible API. This means any application that works with the OpenAI API can use Ollama as a drop-in replacement. It integrates with Open WebUI, Dify, n8n, LangChain, and many other tools.

Deploy Ollama on your server

GoPanel makes self-hosting effortless. Deploy Ollama on any VPS in seconds with automatic SSL certificates, scheduled backups, and one-click updates — all included free.

App Details

Category
AI / LLM
Language
Ai
License
MIT
Website
ollama.ai

View on GitHub

ollama/ollama

166k stars

Ready to deploy?

Self-host Ollama on your own server in under 60 seconds.

Deploy now