Ollama
Run large language models locally with a single command
Ollama makes it easy to run open-source LLMs locally on your machine. It handles model downloading, quantization, and serving with an OpenAI-compatible API. Supports Llama, Mistral, Gemma, Phi, and many other model families. Popular for local development, testing, and offline AI applications.
Pricing: Free
Ollama Alternatives
Explore 51 products in the Inference APIs category. View all Ollama alternatives.
deepinfra
Run the top AI models using a simple API, pay per use. Low cost, scalable and production ready infrastructure.
LLMWise
Multi-LLM API orchestration platform for comparing and blending AI models
novita.ai
APIs, Serverless and GPU Instance In One AI Cloud
Nebius
Full-stack AI cloud with GPU infrastructure for training and inference
Also listed in
Is your product missing?