Running Your First Eval
Learn how to get started with your first evaluation run in Maxim
1. Set up your environment
First, configure your AI model providers:
Go to `Settings` → `Models`.
Configure model provider
Add New
and fill in the required details.To learn more about API keys, inviting users, and managing roles, refer to our Workspace and roles guide.
2. Create your first prompt or HTTP endpoint
Create prompts to experiment and evaluate a call to a model with attached context or tools. Use endpoints to easily test your complex AI agents using the HTTP endpoint for your application without any integration.
Prompt
Create prompt
Prompts
tab under the Evaluate
section and click on Single prompts. Click Create prompt
or Try sample
to get started.Write your first prompt
Configure model and parameters.
Iterate
Run
to test your prompt and see the AI’s response. Iterate on your prompt based on the results.Save prompt and publish a version.
Save
to create a new version of your prompt.To learn more about prompts, refer to our detailed guide on prompts.
HTTP Endpoint
Create endpoint
HTTP Endpoints
option under the tab Agents
located in the Evaluate
section. Click Create Endpoint
or Try sample
.Configure agent endpoint
URL
field. Configure any necessary headers or parameters. You can use dynamic variables like {input}
to reference static context easily in any part of your endpoint using {}
Test your agent
Run
to test your endpoint in the playground.Configure endpoint for testing
Output Mapping
section, select the part of the response you want to evaluate (e.g., data.response
). Click Save
to create your endpoint.To learn more about agent endpoints, refer to our detailed guide on Agent Endpoints.
3. Prepare your dataset
Organize and manage the data you’ll use for testing and evaluation:
Create dataset
Library
section. Click Create New
or Upload CSV
. We also have a sample dataset created for you. Click on View our sample dataset
to get started.Edit dataset
Save
Save
to create your dataset.To learn more about datasets, refer to our detailed guide on Datasets.
5. Add evaluators
Set up evaluators to assess your prompt or endpoint’s performance:
Add evaluators from store
Evaluators
tab under the Library
section. Click Add Evaluator
to browse available evaluators.Configure added evaluators
Save
to add the evaluator to your workspace.To learn more about evaluators, refer to our detailed guide on Evaluators.
6. Run your first test
Execute a test run to evaluate your prompt or endpoint:
Select endpoint/prompt to test
Test
in the top right corner.Configure test run
Trigger
Trigger Test Run
to start the evaluation process.7. Analyze test results
Review and analyze the results of your test run:
View report
Runs
tab in the left navigation menu. Find your recent test run and click on it to view details.Review performance
Iterate
Next steps
Now that you’ve completed your first cycle on the Maxim platform, consider exploring these additional capabilities:
- Prompt comparisons: Evaluate different prompts side-by-side to determine which ones produce the best results for a given task.
- Agents via no-code builder: Create complex, multi-step AI workflows. Learn how to connect prompts, code, and APIs to build powerful, real-world AI systems using our intuitive, no-code editor.
- Context sources: Integrate Retrieval-Augmented Generation (RAG) into your agent endpoints.
- Prompt tools: Enhance your prompts with custom functions and agentic behaviors.
- Observability: Use our stateless SDK to monitor real-time production logs and run periodic quality checks.
By following this guide, you’ve learned how to set up your environment, create prompts, prepare datasets, set up endpoints, add evaluators, run tests, and analyze results. This foundational knowledge will help you leverage Maxim’s powerful features to develop and improve your AI applications efficiently.