# AI Configuration Guide ## Environment Variables Add these to your `.env` file: ```env # AI Configuration # Choose between 'ollama' or 'openrouter' AI_PROVIDER=openrouter # Ollama Configuration (if AI_PROVIDER=ollama) AI_PORT=11434 AI_MODEL=gpt-oss:20b # OpenRouter Configuration (if AI_PROVIDER=openrouter) OPENROUTER_API_KEY=sk-or-your-api-key-here OPENROUTER_MODEL=gemma OPENROUTER_BASE_URL=openrouter.ai OPENROUTER_REL_PATH=/api OPENROUTER_TEMPERATURE=0.7 ``` ## Available OpenRouter Models Based on your C# implementation, these models are available: - `gemma` - google/gemma-3-12b-it - `dolphin` - cognitivecomputations/dolphin-mixtral-8x22b - `dolphin_free` - cognitivecomputations/dolphin3.0-mistral-24b:free - `gpt-4o-mini` - openai/gpt-4o-mini - `gpt-4.1-nano` - openai/gpt-4.1-nano - `qwen` - qwen/qwen3-30b-a3b - `unslop` - thedrummer/unslopnemo-12b - `euryale` - sao10k/l3.3-euryale-70b - `wizard` - microsoft/wizardlm-2-8x22b - `deepseek` - deepseek/deepseek-chat-v3-0324 - `dobby` - sentientagi/dobby-mini-unhinged-plus-llama-3.1-8b ## Testing 1. Set `AI_PROVIDER=openrouter` in your `.env` 2. Add your OpenRouter API key 3. Test the connection: `GET http://localhost:8083/rest/ai/test-ai` 4. Start an interview to test the full flow ## Switching Back to Ollama To switch back to Ollama: 1. Set `AI_PROVIDER=ollama` in your `.env` 2. Make sure Ollama is running on the specified port 3. Test the connection: `GET http://localhost:8083/rest/ai/test-ai`