LLM sometimes returns inconsistent total values. Now we always compute total = sum(criteria.score) for accuracy. |
||
|---|---|---|
| .. | ||
| workflow_templates | ||
| aggregate_final_grade.py | ||
| aggregate_llm_grades.py | ||
| create_minimal_metadata.py | ||
| generate_pdf_report.py | ||
| grade_grouped.py | ||
| grade.py | ||
| llm_grade.py | ||
| objective_grade.py | ||
| post_comment.py | ||
| run_tests.py | ||
| test_objective_grade.sh | ||
| upload_metadata.py | ||