-
Notifications
You must be signed in to change notification settings - Fork 24
[LEADS-182] - Add eval config data to the report #142
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
[LEADS-182] - Add eval config data to the report #142
Conversation
WalkthroughAdds a CSV column for metrics_metadata and a configurable summary_config_sections output option; EvaluationResult now carries metrics_metadata, metrics metadata is extracted during evaluation, and output generation serializes selected system configuration sections into JSON/text summaries. Changes
Sequence Diagram(s)mermaid Estimated code review effort🎯 3 (Moderate) | ⏱️ ~25 minutes Possibly related PRs
Suggested reviewers
🚥 Pre-merge checks | ✅ 3✅ Passed checks (3 passed)
✏️ Tip: You can configure your own custom pre-merge checks in the settings. ✨ Finishing touches
🧹 Recent nitpick comments
📜 Recent review detailsConfiguration used: Path: .coderabbit.yaml Review profile: CHILL Plan: Pro 📒 Files selected for processing (9)
✅ Files skipped from review due to trivial changes (1)
🚧 Files skipped from review as they are similar to previous changes (4)
🧰 Additional context used📓 Path-based instructions (4)**/*.py📄 CodeRabbit inference engine (AGENTS.md)
Files:
src/**/*.py📄 CodeRabbit inference engine (AGENTS.md)
Files:
src/lightspeed_evaluation/**/*.py📄 CodeRabbit inference engine (AGENTS.md)
Files:
tests/**/*.py📄 CodeRabbit inference engine (AGENTS.md)
Files:
🧠 Learnings (2)📚 Learning: 2026-01-15T00:41:27.575ZApplied to files:
📚 Learning: 2025-09-10T15:48:14.671ZApplied to files:
🧬 Code graph analysis (2)src/lightspeed_evaluation/pipeline/evaluation/evaluator.py (2)
tests/unit/pipeline/evaluation/test_evaluator.py (1)
⏰ Context from checks skipped due to timeout of 90000ms. You can increase the timeout in your CodeRabbit configuration to a maximum of 15 minutes (900000ms). (4)
🔇 Additional comments (11)
✏️ Tip: You can disable this entire section by setting Thanks for using CodeRabbit! It's free for OSS, and your support helps us grow. If you like it, consider giving us a shout-out. Comment |
VladimirKadlec
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice work, thank you. Please add this to the documentation.
Other than that LGTM.
asamal4
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you.. LGTM. one nit..
FYI.. Also we will have to refactor metadata mapping to simplify the logic (to remove additional logic in evaluator.py). But not part of this story or PR.
| if self.system_config is not None and hasattr(self.system_config, "output"): | ||
| if hasattr(self.system_config.output, "summary_config_sections"): | ||
| return self.system_config.output.summary_config_sections | ||
| # Default sections if not configured (see system.py:220) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please remove code line number, that may get change.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
in fact remove complete comment itself and use a constant; that will be self-explanatory.
Mock system_config and metric_manager now properly support iteration operations to prevent TypeError when production code iterates over model_fields.keys() or metadata.items(). Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>
25293d8 to
3ffbb1e
Compare
| ) | ||
|
|
||
| summary_config_sections: list[str] = Field( | ||
| default=["llm", "embedding", "api"], |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please use a constant for this.
| default=["llm", "embedding", "api"], | |
| default=DEFAULT_STORED_CONFIGS, |
| "result", | ||
| "score", | ||
| "threshold", | ||
| "metrics_metadata", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Please rename the column name
| "metrics_metadata", | |
| "metric_metadata", |
Description
Add all configuration information to the JSON and TXT file.
Add metric level metadata to CSV file.
Type of change
Tools used to create PR
Assisted-by: Claude
Generated by: Claude-4.5-sonnet
Related Tickets & Documents
Related Issue: RSPEED-2121
Closes: LEADS-182
Checklist before requesting a review
Testing
test_generator.pyandtest_evaluator.pywere updated to support new additions to the codebase.Output Examples
leads-182-json-example.json
leads-182-metrics-metadata-example.csv
leads-182-txt-example.txt
Summary by CodeRabbit
New Features
Documentation
✏️ Tip: You can customize this high-level summary in your review settings.