Moved experimental LLM integration code to optimization_engine/future/: - llm_optimization_runner.py - Runtime LLM API runner - llm_workflow_analyzer.py - Workflow analysis - inline_code_generator.py - Auto-generate calculations - hook_generator.py - Auto-generate hooks - report_generator.py - LLM report generation - extractor_orchestrator.py - Extractor orchestration Added comprehensive optimization_engine/future/README.md explaining: - MVP LLM strategy (Claude Code skills, not runtime LLM) - Why files were archived - When to revisit post-MVP - Production architecture reference Production runner confirmed: optimization_engine/runner.py is sole active runner. This establishes clear separation between: - Production code (stable, no runtime LLM dependencies) - Experimental code (archived for post-MVP exploration) Part of Phase 1: Core Stabilization & Organization for MVP Generated with Claude Code Co-Authored-By: Claude <noreply@anthropic.com>
106 lines
4.2 KiB
Markdown
106 lines
4.2 KiB
Markdown
# Experimental LLM Features (Archived)
|
|
|
|
**Status**: Archived for post-MVP development
|
|
**Date Archived**: November 24, 2025
|
|
|
|
## Purpose
|
|
|
|
This directory contains experimental LLM integration code that was explored during early development phases. These features are archived (not deleted) for potential future use after the MVP is stable and shipped.
|
|
|
|
## MVP LLM Integration Strategy
|
|
|
|
For the **MVP**, LLM integration is achieved through:
|
|
- **Claude Code Development Assistant**: Interactive development-time assistance
|
|
- **Claude Skills** (`.claude/skills/`):
|
|
- `create-study.md` - Interactive study scaffolding
|
|
- `analyze-workflow.md` - Workflow classification and analysis
|
|
|
|
This approach provides LLM assistance **without adding runtime dependencies** or complexity to the core optimization engine.
|
|
|
|
## Archived Experimental Files
|
|
|
|
### 1. `llm_optimization_runner.py`
|
|
Experimental runner that makes runtime LLM API calls during optimization. This attempted to automate:
|
|
- Extractor generation
|
|
- Inline calculations
|
|
- Post-processing hooks
|
|
|
|
**Why Archived**: Adds runtime dependencies, API costs, and complexity. The centralized extractor library (`optimization_engine/extractors/`) provides better maintainability.
|
|
|
|
### 2. `llm_workflow_analyzer.py`
|
|
LLM-based workflow analysis for automated study setup.
|
|
|
|
**Why Archived**: The `analyze-workflow` Claude skill provides the same functionality through development-time assistance, without runtime overhead.
|
|
|
|
### 3. `inline_code_generator.py`
|
|
Auto-generates inline Python calculations from natural language.
|
|
|
|
**Why Archived**: Manual calculation definition in `optimization_config.json` is clearer and more maintainable for MVP.
|
|
|
|
### 4. `hook_generator.py`
|
|
Auto-generates post-processing hooks from natural language descriptions.
|
|
|
|
**Why Archived**: The plugin system (`optimization_engine/plugins/`) with manual hook definition is more robust and debuggable.
|
|
|
|
### 5. `report_generator.py`
|
|
LLM-based report generation from optimization results.
|
|
|
|
**Why Archived**: Dashboard provides rich visualizations. LLM summaries can be added post-MVP if needed.
|
|
|
|
### 6. `extractor_orchestrator.py`
|
|
Orchestrates LLM-based extractor generation and management.
|
|
|
|
**Why Archived**: Centralized extractor library (`optimization_engine/extractors/`) is the production approach. No code generation needed at runtime.
|
|
|
|
## When to Revisit
|
|
|
|
Consider reviving these experimental features **after MVP** if:
|
|
1. ✅ MVP is stable and well-tested
|
|
2. ✅ Users request more automation
|
|
3. ✅ Core architecture is mature enough to support optional LLM features
|
|
4. ✅ Clear ROI on LLM API costs vs manual configuration time
|
|
|
|
## Production Architecture (MVP)
|
|
|
|
For reference, the **stable production** components are:
|
|
|
|
```
|
|
optimization_engine/
|
|
├── runner.py # Production optimization runner
|
|
├── extractors/ # Centralized extractor library
|
|
│ ├── __init__.py
|
|
│ ├── base.py
|
|
│ ├── displacement.py
|
|
│ ├── stress.py
|
|
│ ├── frequency.py
|
|
│ └── mass.py
|
|
├── plugins/ # Plugin system (hooks)
|
|
│ ├── __init__.py
|
|
│ └── hook_manager.py
|
|
├── nx_solver.py # NX simulation interface
|
|
├── nx_updater.py # NX expression updates
|
|
└── visualizer.py # Result plotting
|
|
|
|
.claude/skills/ # Claude Code skills
|
|
├── create-study.md # Interactive study creation
|
|
└── analyze-workflow.md # Workflow analysis
|
|
```
|
|
|
|
## Migration Notes
|
|
|
|
If you need to use any of these experimental files:
|
|
1. They are functional but not maintained
|
|
2. Update imports to `optimization_engine.future.{module_name}`
|
|
3. Install any additional dependencies (LLM client libraries)
|
|
4. Be aware of API costs for LLM calls
|
|
|
|
## Related Documents
|
|
|
|
- [`docs/07_DEVELOPMENT/Today_Todo.md`](../../docs/07_DEVELOPMENT/Today_Todo.md) - MVP Development Plan
|
|
- [`DEVELOPMENT.md`](../../DEVELOPMENT.md) - Development guide
|
|
- [`.claude/skills/create-study.md`](../../.claude/skills/create-study.md) - Study creation skill
|
|
|
|
## Questions?
|
|
|
|
For MVP development questions, refer to the [DEVELOPMENT.md](../../DEVELOPMENT.md) guide or the MVP plan in `docs/07_DEVELOPMENT/Today_Todo.md`.
|