Learn how to get started with your first evaluation run in Maxim
Go to `Settings` → `Models`.
Configure model provider
Add New
and fill in the required details.Create prompt
Prompts
tab under the Evaluate
section and click on Single prompts. Click Create prompt
or Try sample
to get started.Write your first prompt
Configure model and parameters.
Iterate
Run
to test your prompt and see the AI’s response. Iterate on your prompt based on the results.Save prompt and publish a version.
Save
to create a new version of your prompt.Create endpoint
HTTP Endpoints
option under the tab Agents
located in the Evaluate
section. Click Create Endpoint
or Try sample
.Configure agent endpoint
URL
field. Configure any necessary headers or parameters. You can use dynamic variables like {input}
to reference static context easily in any part of your endpoint using {}
Test your agent
Run
to test your endpoint in the playground.Configure endpoint for testing
Output Mapping
section, select the part of the response you want to evaluate (e.g., data.response
). Click Save
to create your endpoint.Create dataset
Library
section. Click Create New
or Upload CSV
. We also have a sample dataset created for you. Click on View our sample dataset
to get started.Edit dataset
Save
Save
to create your dataset.Add evaluators from store
Evaluators
tab under the Library
section. Click Add Evaluator
to browse available evaluators.Configure added evaluators
Save
to add the evaluator to your workspace.Select endpoint/prompt to test
Test
in the top right corner.Configure test run
Trigger
Trigger Test Run
to start the evaluation process.View report
Runs
tab in the left navigation menu. Find your recent test run and click on it to view details.Review performance
Iterate