Developed and Evaluated AI Model Rubrics

  • Day: 2025-06-19
  • Time: 00:00 to 23:55
  • Project: Teaching
  • Workspace: WP 2: Operational
  • Status: Completed
  • Priority: MEDIUM
  • Assignee: Matías Nehuen Iglesias
  • Tags: Ai Models, Rubric, Evaluation, Machine Learning, Education

Description

Session Goal

The session aimed to develop and evaluate rubric criteria for AI model assessments, focusing on physics-heavy AI challenges, scientific prompt citations, and grading AI model responses.

Key Activities

  • Created MECE rubric criteria for evaluating AI challenges in physics, ensuring criteria are atomic and self-contained.
  • Established guidelines for citing sources in scientific prompts, distinguishing between common knowledge and niche information.
  • Developed a framework for grading AI models using rubrics, aligning quantitative scores with qualitative assessments.
  • Proposed a correction scheme for evaluating prediction model responses, including evaluation criteria, common errors, and expected scores.
  • Evaluated student performance on bias and variance concepts in machine learning models using a grading scale.

Achievements

  • Successfully developed comprehensive rubric criteria and grading frameworks for AI model evaluation.
  • Clarified the distinction between bias and overfitting in machine learning, enhancing understanding of model evaluation.

Pending Tasks

  • Further refine the rubric criteria for specific AI challenges and model types.
  • Implement the proposed correction scheme in actual model evaluations to validate its effectiveness.

Evidence

  • source_file=2025-06-19.sessions.jsonl, line_number=0, event_count=0, session_id=df9b58261241853dc896c91c22dcc658274ef4478797a7c62c01ea5a47f628a5
  • event_ids: []