feat(providers): added openrouter support

This PR is adapted from namanbnsl's original
at https://github.com/namanbnsl/Perplexica/tree/add-openrouter-support

Added minor modifications to bring the openrouter support in line
with the changes in this forked repo
This commit is contained in:
kmac 2025-05-31 21:34:23 -04:00
parent 6422c5b7e0
commit 94315c169a
7 changed files with 108 additions and 0 deletions

View file

@ -79,6 +79,7 @@ There are mainly 2 ways of installing Perplexica - With Docker, Without Docker.
- `OPENAI`: Your OpenAI API key. **You only need to fill this if you wish to use OpenAI's models**.
- `OLLAMA`: Your Ollama API URL. You should enter it as `http://host.docker.internal:PORT_NUMBER`. If you installed Ollama on port 11434, use `http://host.docker.internal:11434`. For other ports, adjust accordingly. **You need to fill this if you wish to use Ollama's models instead of OpenAI's**.
- `GROQ`: Your Groq API key. **You only need to fill this if you wish to use Groq's hosted models**.
- `OPENROUTER`: Your OpenRouter API key. **You only need to fill this if you wish to use models via OpenRouter**.
- `ANTHROPIC`: Your Anthropic API key. **You only need to fill this if you wish to use Anthropic models**.
**Note**: You can change these after starting Perplexica from the settings dialog.