clipboard-checkEvaluation

Run evals on your trained model against a test dataset. Access via lr.evals on your LightningRod client.

Methods

create

Create an eval job without waiting:

eval_job = lr.evals.create(
    model_id=job.model_id,
    dataset=test_dataset,
    benchmark_model_id="openai/gpt-5.2",
    temperature=0.0,
)
Parameter
Type
Default
Description

model_id

str

Your trained model ID

dataset

SampleDataset

Test dataset

benchmark_model_id

str | None

None

Optional model to compare against

temperature

float

0.0

Sampling temperature

run

Create an eval job and poll until completion. In notebooks, shows a live progress display:

eval_job = lr.evals.run(
    model_id=job.model_id,
    dataset=test_dataset,
    benchmark_model_id="openai/gpt-5.2",
    temperature=0.0,
    poll_interval=15.0,
)

get

Fetch a single eval job by ID:

list

List eval jobs with pagination:

Pretty-print eval results:

Or with a previously fetched job:

Example

See notebooks/getting_started/05_fine_tuning.ipynbarrow-up-right for the full workflow.

Last updated