- Removed accordion functionality as content became too sparse
- Restored detailed prompt structure from llm_service.py
- Enhanced fallback analysis with comprehensive metrics (VaR, R2, etc)
- Kept popover options for different analysis modes
- Updated max_tokens to 3000 for more detailed responses