opencode
opencode copied to clipboard
[FEATURE]: Prompt + file edit logging for benchmarking
Feature hasn't been suggested before.
- [x] I have verified this feature I'm about to request hasn't been suggested before.
Describe the enhancement you want to request
Hi! Would it be possible to add logging for:
- the prompts sent to the model, and
- the file edits/patches produced during execution
This would make it easier to replay the exact same prompt(s) and outputs for benchmarking across different models.
Optionally, we could also support an evaluation mode where an LLM acts as a judge to score/compare results for the same session.