# Quick Start: Using the New Academic Prompting System ## What Changed? RAG Assistant for Zotero now uses **2025 best practices** for academic question answering, including: - Better citation accuracy (every claim is cited) - Clearer distinction between facts and analysis - Improved handling of uncertainty - Reduced repetition in responses - More natural academic writing ## No Action Required The system automatically uses the new prompts. You don't need to: - Reconfigure settings - Re-index your library - Change API keys - Update your queries ## What You'll Notice ### Better Citations **Before**: > "Studies show that neural networks improve performance." **After**: > "Recent research demonstrates that transformer architectures improve accuracy by 15-25% [1][2]." ### Explicit Limitations **Before**: > "This topic isn't well covered." **After**: > "The retrieved documents don't address training efficiency. Your library may benefit from adding papers on optimization techniques." ### Structured Responses Responses now follow a consistent pattern: 1. **Direct Answer** (2-3 sentences) 2. **Key Evidence** (bullet points with citations) 3. **Synthesis** (connections across sources) 4. **Limitations** (what's missing or uncertain) ## Advanced: Customizing Generation If you want to adjust the generation style, you can modify parameters in `backend/academic_prompts.py`: ### Make responses more creative (for literature reviews): ```python # In your code, use: gen_params = AcademicGenerationParams.get_params("creative") ``` ### Make responses more precise (for fact extraction): ```python gen_params = AcademicGenerationParams.get_params("precise") ``` ### Default (balanced): ```python gen_params = AcademicGenerationParams.get_params("standard") ``` ## Troubleshooting ### "Responses are longer than before" - **Expected**: The system now provides more detailed, well-cited answers - **Solution**: Responses are capped at 600 tokens (was 512) for academic depth - **If needed**: Edit `max_tokens` in `AcademicGenerationParams.STANDARD` ### "Model repeating citations" - **Fixed**: New parameters reduce repetition - **If it persists**: Check your model version (update Ollama models: `ollama pull `) ### "Want the old behavior" - **Option 1**: Revert to commit before this change - **Option 2**: Edit `backend/conversation_store.py` and use shorter system prompt - **Recommended**: Try the new system for a few days - most users prefer it ## Performance Notes - **Token Usage**: ~10-15% increase (more detailed responses) - **Speed**: Similar (slightly slower due to more tokens generated) - **Quality**: Significantly better grounding and citation accuracy ## Questions? See full documentation: `docs/ACADEMIC_PROMPTING_2025.md`