Ollama
Run large language models locally with a single command
Ollama makes it easy to run open-source LLMs locally on your machine. It handles model downloading, quantization, and serving with an OpenAI-compatible API. Supports Llama, Mistral, Gemma, Phi, and many other model families. Popular for local development, testing, and offline AI applications.
Pricing: Free
Ollama Alternatives
Explore 56 products in the Inference APIs category. View all Ollama alternatives.
deepinfra
Run the top AI models using a simple API, pay per use. Low cost, scalable and production ready infrastructure.
Cerebras
Ultra-fast inference on custom wafer-scale hardware with OpenAI-compatible API
AiQu
Swedish GPU infrastructure and LLM hosting platform with API-first deployment, no Kubernetes required
Also listed in
Is your product missing?