TacoSkill LABTacoSkill LAB

The full-lifecycle AI skills platform.

Product

  • SkillHub
  • Playground
  • Skill Create
  • SkillKit

Resources

  • Privacy
  • Terms
  • About

Platforms

  • Claude Code
  • Cursor
  • Codex CLI
  • Gemini CLI
  • OpenCode

© 2026 TacoSkill LAB. All rights reserved.

TacoSkill LAB
TacoSkill LAB
HomeSkillHubCreatePlaygroundSkillKit
  1. Home
  2. /
  3. SkillHub
  4. /
  5. evaluate-model
Improve

evaluate-model

4.9

by majiayu000

145Favorites
57Upvotes
0Downvotes

"Measure model performance on test datasets. Use when assessing accuracy, precision, recall, and other metrics."

model evaluation

4.9

Rating

0

Installs

Machine Learning

Category

Quick Review

The skill provides a well-structured overview of model evaluation with clear workflow steps and appropriate metric coverage. The description adequately conveys the skill's purpose for assessing model performance. Task knowledge is good, including classification and regression evaluation patterns with specific metrics. Structure is clean and logical with distinct sections. However, novelty is limited - model evaluation is a common ML task that CLI agents with standard libraries can handle reasonably well. The skill would benefit from more advanced features like automated metric selection, statistical significance testing, or cross-validation automation to increase its value proposition beyond what standard ML libraries offer.

LLM Signals

Description coverage6
Task knowledge7
Structure8
Novelty4

GitHub Signals

49
7
1
1
Last commit 0 days ago

Publisher

majiayu000

majiayu000

Skill Author

Related Skills

ml-pipelinepyvene-interventionsnnsight-remote-interpretability

Loading SKILL.md…

Try onlineView on GitHub

Publisher

majiayu000 avatar
majiayu000

Skill Author

Related Skills

ml-pipeline

Jeffallan

6.4

pyvene-interventions

zechenzhangAGI

7.6

nnsight-remote-interpretability

zechenzhangAGI

7.0

mlflow

zechenzhangAGI

7.6
Try online