Skip to main content
Building an action is just the first step. Testing ensures it works as expected across different scenarios and user inputs.

Your Testing Workspace

Screenshot 2026 02 19 At 1 56 47 PM
On the right side of the Action Details page, you’ll find your testing panel. Think of this as a safe sandbox where you can run your action with real data without affecting anything in production.

Test Panel Features

The testing workspace includes:
  • Input box: Enter natural language queries as users would phrase them
  • Example prompts: Pre-configured test cases you can click to run
  • Results display: Shows the agent’s response and any data retrieved
  • Execution trace: Details which steps ran and their outputs

Running Your First Test

Start by typing a natural language query exactly as a user would phrase it. Use natural phrasing without excessive formality.

Example Test Cases

"Show me licenses expiring in the next 30 days"
"What software renewals are coming up?"
"I need to see upcoming license expirations"
Watch what happens as the action executes. You’ll see each step complete in sequence, with the data flowing from one step to the next. If something fails, you’ll see exactly which step caused the problem. Test Execution Results
Test with the exact phrases your users are likely to use, not just technical terminology. The goal is to ensure the action works for real-world usage.

Using Example Prompts

Click any of the example prompts below the input box to automatically run a test with that phrasing. This is a quick way to verify that the action handles different phrasings correctly.

Verification Points

  • Does the action trigger for all the example prompts?
  • Are the results consistent across different phrasings?
  • Does the output format appear clean and professional?
  • Are all required data fields populated correctly?

Testing Edge Cases

Beyond testing expected scenarios, validate the action’s behavior with edge cases and potential failure conditions.

Empty Results

Test what happens when no data matches the query:
"Show me renewals for next week" (when there aren't any)
Expected behavior: Clear message indicating no results, possibly with suggestions for broadening the search.

Invalid Input

Test with queries that don’t make logical sense:
"Show me renewals from tomorrow to yesterday" (date range doesn't make sense)
Expected behavior: User-friendly error message explaining the issue and how to correct it.

Missing Permissions

Test as a user who doesn’t have access to certain data:
Test with a restricted user account
Expected behavior: Appropriate message about permissions, without exposing data the user shouldn’t see.

API Failures

Test how your action handles it when an external service is down. You might need to temporarily break an API endpoint for this test. Expected behavior: Graceful error handling with retry logic and clear user communication.
Always test permission boundaries. Ensure users can’t access data they shouldn’t see by manipulating prompts or parameters.

Interpreting Test Results

Successful Execution

When the test completes successfully and produces expected output, the action is functioning correctly. Verify:
  • Output format matches expectations
  • All required data is present
  • Response time is acceptable
  • User experience feels smooth

Warning Indicators

These signal potential issues that didn’t prevent execution but may cause problems in certain scenarios. Common warnings include:
  • Slow API response times
  • Large data payloads
  • Deprecated API endpoints
  • Missing optional fields
These should be investigated and resolved before deployment.

Error States

Failed tests require immediate attention. Examine error details to identify the failing step and root cause. Common failure scenarios include:
  • API endpoint changed or is down
  • Data format unexpected
  • Missing required fields
  • Permission issues
  • Timeout errors
  • Rate limiting

Issue Resolution Strategies

If the Wrong Action Triggered

Refine your selection criteria to be more specific about when this action should (and shouldn’t) run. Add more descriptive keywords and scenarios.

If a Step Failed

Click on that step to edit its configuration. Common fixes:
  • Update API endpoint URLs
  • Adjust data processing logic
  • Add error handling
  • Include retry mechanisms
  • Update authentication credentials

If the Output is Confusing

Modify your output formatting step to present the data more clearly:
  • Use tables for structured data
  • Add headers and labels
  • Highlight important information
  • Remove unnecessary fields
  • Improve readability with proper spacing

If You’re Missing Data

Add a new data processing step to extract the fields you need:
  • Identify which API response contains the data
  • Add extraction logic
  • Transform format if needed
  • Handle cases where data might be missing
Run the test again after each change to verify your fix worked. Iterative testing is key to building reliable actions.

Testing with Real Users

Before deploying to production, have a few team members test the action. Give them minimal guidance—you want to see if the action is intuitive to use without explanation.

Questions to Ask Testers

  • Could you get the results you needed?
  • Was the output format helpful?
  • Did anything confuse you?
  • What would make this better?
  • How fast did it feel?
  • Would you use this regularly?
This feedback is invaluable for refinement.

Beta Testing Process

  1. Select test users: Choose representatives from different roles
  2. Provide minimal instructions: Just tell them what the action is for
  3. Observe usage: Watch how they phrase requests
  4. Collect feedback: Ask open-ended questions
  5. Iterate: Make improvements based on real usage
  6. Repeat: Test again with refined version

Test Coverage Checklist

Before deploying to production, ensure you’ve tested:
  • Happy path with typical inputs
  • Edge cases (empty results, invalid input)
  • Different user roles and permissions
  • Various phrasings of the same request
  • Error scenarios (API failures, timeouts)
  • Large data sets
  • Concurrent usage
  • Mobile and desktop interfaces
  • Different time zones (if relevant)
  • Boundary values (min/max dates, etc.)

Performance Testing

Beyond functional testing, consider performance:

Response Time

  • Actions should respond within 3-5 seconds for simple queries
  • Complex analyses may take up to 15-20 seconds
  • Anything longer needs optimization or progress indicators

Reliability

  • Actions should succeed 99%+ of the time under normal conditions
  • Transient failures should trigger automatic retries
  • Persistent failures should be logged for investigation

Scalability

  • Test with realistic data volumes
  • Ensure API rate limits aren’t exceeded
  • Verify caching strategies work effectively
Once your action passes testing Deploy to production and manage versions