AI Configuration Guide

Environment Variables

Add these to your .env file:

# AI Configuration
# Choose between 'ollama' or 'openrouter'
AI_PROVIDER=openrouter

# Ollama Configuration (if AI_PROVIDER=ollama)
AI_PORT=11434
AI_MODEL=gpt-oss:20b

# OpenRouter Configuration (if AI_PROVIDER=openrouter)
OPENROUTER_API_KEY=sk-or-your-api-key-here
OPENROUTER_MODEL=gemma
OPENROUTER_BASE_URL=openrouter.ai
OPENROUTER_REL_PATH=/api
OPENROUTER_TEMPERATURE=0.7

Available OpenRouter Models

Based on your C# implementation, these models are available:

  • gemma - google/gemma-3-12b-it
  • dolphin - cognitivecomputations/dolphin-mixtral-8x22b
  • dolphin_free - cognitivecomputations/dolphin3.0-mistral-24b:free
  • gpt-4o-mini - openai/gpt-4o-mini
  • gpt-4.1-nano - openai/gpt-4.1-nano
  • qwen - qwen/qwen3-30b-a3b
  • unslop - thedrummer/unslopnemo-12b
  • euryale - sao10k/l3.3-euryale-70b
  • wizard - microsoft/wizardlm-2-8x22b
  • deepseek - deepseek/deepseek-chat-v3-0324
  • dobby - sentientagi/dobby-mini-unhinged-plus-llama-3.1-8b

Testing

  1. Set AI_PROVIDER=openrouter in your .env
  2. Add your OpenRouter API key
  3. Test the connection: GET http://localhost:8083/rest/ai/test-ai
  4. Start an interview to test the full flow

Switching Back to Ollama

To switch back to Ollama:

  1. Set AI_PROVIDER=ollama in your .env
  2. Make sure Ollama is running on the specified port
  3. Test the connection: GET http://localhost:8083/rest/ai/test-ai