plexus.cli.shared.evaluation_runner module

plexus.cli.shared.evaluation_runner.create_tracker_and_evaluation(*, client: PlexusDashboardClient, account_id: str, scorecard_name: str, number_of_samples: int, sampling_method: str = 'random', score_name: str | None = None) Tuple[TaskProgressTracker, Evaluation]

Create a TaskProgressTracker and an Evaluation record for an accuracy evaluation.

Mirrors the CLI behavior so other callers (e.g., MCP server tools) don’t duplicate logic. Returns the tracker (with its Task created/claimed) and the Evaluation record.

async plexus.cli.shared.evaluation_runner.run_accuracy_evaluation(*, scorecard_name: str, score_name: str | None = None, number_of_samples: int = 10, sampling_method: str = 'random', client: PlexusDashboardClient | None = None, account_id: str | None = None, fresh: bool = True, reload: bool = False, use_yaml: bool = True) dict

Run a complete accuracy evaluation using the same logic as CLI.

This is the shared function that both CLI and MCP should use to avoid code duplication.