Proxy requests to publicly accessible Ollama instances discovered across the internet
We scan the internet for open Ollama deployments and automatically proxy your requests to available models. No need to run your own inference hardware - just use our distributed network of public nodes.
Generate text completions using distributed Ollama models
Chat completions with conversation context
List all available models across the network
Show currently running models with resource usage