# Git Issue Agent - Ollama configuration # # Uses a local Ollama instance for LLM inference. # Prerequisite: Ollama must be running with the model pulled: # ollama pull ibm/granite4:latest # LLM configuration TASK_MODEL_ID=ollama_chat/ibm/granite4:latest # Ollama API base URL. Required by litellm (used by crewai >=1.10). # For Docker Desktop / Kind: http://host.docker.internal:11434 # For in-cluster Ollama: http://ollama.ollama.svc:11434 LLM_API_BASE=http://host.docker.internal:11434 OLLAMA_API_BASE=http://host.docker.internal:11434 LLM_API_KEY=ollama MODEL_TEMPERATURE=0 # Agent service SERVICE_PORT=8000 LOG_LEVEL=DEBUG # MCP Tool endpoint MCP_URL=http://github-tool-mcp:9090/mcp