LLM Playground
The XeroML Playground is an interactive environment for testing prompts against real LLM models directly in the UI. Use it to validate prompt changes before updating the production label.
What You Can Do
- Test prompt variables — fill in variable values and see the compiled prompt
- Switch models — run the same prompt against different models for comparison
- Adjust model parameters — change temperature, max tokens, and other settings
- Compare versions — run the current and previous versions side by side
- Save results — traces from Playground runs appear in your project’s trace list
Using the Playground
- Open a prompt in the XeroML UI
- Click Open in Playground
- Fill in any variable values
- Select a model (requires an LLM Connection configured in your project)
- Click Run
The response appears in the right panel. The Playground trace is automatically linked to the prompt version being tested.
Comparing Versions
To compare two versions:
- Open the Playground for a prompt
- Click Compare to open a split-panel view
- Select the versions to compare on left and right
- Run both — results appear side by side
Use this before deploying a new version to verify it produces better output on your representative test cases.
LLM Connections
The Playground uses LLM Connections configured in your project settings to make actual API calls. You can add connections for:
- OpenAI
- Anthropic
- Azure OpenAI
- Google Vertex AI
- Ollama (for local models)
- Any OpenAI-compatible endpoint
Go to Project Settings → LLM Connections to configure.