Access Model Playground
Overview
The Model Playground is an interactive interface in Armada Bridge that lets you test deployed models, run predictions, and validate model behavior before integrating them into production applications.
Use the playground to:
- Send prompts and inspect model responses in real time
- Verify that a deployed model behaves as expected
- Iterate on prompts and parameters without writing code
- Share the playground URL with your team for quick testing (within tenant isolation)
Prerequisites
Before you can use the Model Playground, ensure that:
- A model has been deployed (as a Tenant Admin or by a user with deployment permissions)
- You have a tenant user account (created by a Tenant Admin)
- You are logged in as that tenant user
Accessing the Model Playground
Step 1: Open AI Studio and Models
-
Log in to Armada Bridge as a tenant user.
-
In the left sidebar, click AI Studio.

-
Under Models, locate the model you want to test.
-
Click Open in Playground for that model.

Step 2: Open the Playground
Click Open in Playground. The playground opens in your browser.
A security warning is shown only when the model endpoint does not have an authorized certificate (for example, when it uses a self-signed or untrusted certificate). If you see this warning, you can proceed as follows:
- Copy the model endpoint URL from the Playground page.
- Open a new browser tab and paste that URL.
- When the browser shows a security warning, click Advanced (or equivalent in your browser).
- Choose Proceed to [site] (unsafe) or Accept the risk and continue to load the endpoint page once.
- You may see a "404 page not found" message which is expected.
- Return to the Playground tab and refresh the page if needed. You can then run queries in the prompting interface.
This one-time step establishes trust for the endpoint in your browser. You do not need to repeat it for the same endpoint.

Step 3: Access the Playground
Return to the Playground tab. You can now run queries in the prompting interface, as shown below.

Tenant user isolation is enforced in Armada Bridge. One tenant user cannot see or use another tenant user's Model Playground. Each tenant user (for example, tenantuser1 and tenantuser2) must complete the steps above with their own login to access their own model playground.
Using the Playground
In the prompting interface you can:
- Enter text prompts and submit them to the deployed model
- View the model’s response in the same screen
- Adjust parameters (if supported by the model) and re-run tests
Use these results to validate behavior before calling the model from applications or APIs.
Playground Interface
The playground provides:
- Input Editor - Define model inputs
- Prediction Viewer - View model outputs
- History - Track prediction calls
- Export - Save test results

Test Models
Single Prediction
Make a single prediction:
- Enter input data in the input editor
- Click Predict or Submit
- View model output
{
"inputs": {
"feature1": 1.0,
"feature2": 2.0,
"feature3": 3.0
}
}

Output:
{
"prediction": 0.95,
"confidence": 0.98
}
Batch Predictions
Test multiple inputs:
- Prepare CSV or JSON file with multiple inputs
- Upload file to playground
- Submit batch request
- Download results
feature1,feature2,feature3
1.0,2.0,3.0
2.0,3.0,4.0
3.0,4.0,5.0

Input Formats
JSON Format
{
"instances": [
[1.0, 2.0, 3.0],
[2.0, 3.0, 4.0]
]
}
CSV Format
col1,col2,col3
1.0,2.0,3.0
2.0,3.0,4.0
Text Format
Raw text input for text models
Validate Models
Check Predictions
Verify model behaves correctly:
- Test with known inputs
- Verify expected outputs
- Check confidence scores
- Test edge cases

Test Different Inputs
Explore model behavior:
{
"test_cases": [
{"normal_input": [1.0, 2.0, 3.0]},
{"edge_case": [0.0, 0.0, 0.0]},
{"extreme": [100.0, 100.0, 100.0]}
]
}
Analyze Predictions
Review results:
- Prediction values
- Confidence scores
- Inference time
- Resource usage
Export Results
Download Predictions
Save test results:
- Run predictions
- Click Download Results
- Choose format (JSON, CSV, Excel)

Share Results
Share findings with team:
- Export results
- Include in reports
- Attach to documentation
- Reference in model validation
Performance Testing
Latency Testing
Measure model response time:
Request: 1.5 seconds
Average: 1.2 seconds (from 10 requests)

Throughput Testing
Test throughput capacity:
Requests per second: 100 RPS
Concurrent connections: 10
Resource Usage
Monitor resource consumption:
- CPU usage
- Memory usage
- GPU usage
- Network bandwidth

Debugging
View Model Logs
Check model serving logs:
[2024-01-09 10:15:32] Model loaded successfully
[2024-01-09 10:15:45] Prediction request: {input: [1,2,3]}
[2024-01-09 10:15:46] Inference complete: 0.95
Error Handling
Debug prediction errors:
- Check input format
- Verify input data types
- Check model requirements
- Review error messages

A/B Testing
Compare Models
Test multiple model versions:
- Deploy Model A and Model B
- Send same inputs to both
- Compare outputs
- Analyze differences

Model Comparison
{
"model_a_output": 0.95,
"model_b_output": 0.92,
"difference": 0.03
}
Best Practices
Before Production
- Test with diverse inputs
- Validate edge cases
- Check performance metrics
- Review resource usage
- Document test results
Continuous Validation
- Regular model testing
- Monitor prediction distribution
- Track confidence scores
- Detect data drift
- Compare against baselines