Skip to main content

Testing using the test harness

Updated this week

The test harness lets you safely simulate how your AI agent responds to various events under the scope of that workflow. By simulating emails, texts, or other events from actual constituents in your system, you can evaluate both the accuracy of the response and the agent’s behavior, all without impacting live communications.

We strongly recommend using the test harness to validate every workflow before you launch it.


What you can test with the test harness

The test harness allows you to:

  • Test a workflow even before launching it live

  • Simulate real constituent events, such as a constituent entering the segment, a document uploaded, or an inbound message

  • See how the agent reasons, what actions it takes, and why

  • Validate knowledge accuracy and behavioral decisions

  • Iterate confidently before real students ever interact with the agent

A workflow does not need to be live or launched to be tested. In fact, it’s best to start testing your workflow before it goes live!


How to access the test harness

  1. Navigate to the Agent —> Workflows tab.

  2. Select the workflow you want to test.

  3. Open the workflow view and click Test.

If this is your first time testing, you will see a message indicating that you are starting a simulation.


Choose a constituent

Each simulation runs in the context of a real constituent, because your agent might behave differently based on the information it knows about the given constituent.

  • Search for a specific constituent by name, or

  • Select a random constituent

These options represent constituents already uploaded to your agent, ensuring realistic context during testing.

Once selected, click Start simulation.


Simulate an event

Next, choose the trigger event you want to simulate. You will only be able to select trigger events that have been configured for that workflow.

Common examples include:

  • Email received

  • SMS received

For most tests, we recommend starting with a common event you expect your agent to experience.

Example event

After entering the details, click Send.


What happens during the simulation

Once the event is simulated:

  • The agent reasons about the event, the context, and the behavioral guidance from the workflow

  • The agent decides what action(s) to take, such as whether to respond, escalate, or set reminders

  • The agent simulates the execution the appropriate action, such as sending an email

You will see a clear record of how the agent reasoned and what actions it would take.


Testing a follow-up scenario

The test harness is especially useful for multi-step scenarios.

Example workflow:

  1. A first email is received from a constituent

  2. The agent responds and sets a follow-up alarm

  3. A second email arrives as a follow-up

  4. The agent ends the original alarm

  5. The agent sends a new response

  6. A new follow-up alarm is created based on the latest message

This multi-step scenario allows you to validate that the workflow handles ongoing conversations and time-based follow-ups correctly.


How to evaluate the results

When reviewing a test run by examining the reasoning blocks and the corresponding actions, focus on two distinct areas.

1. Accuracy

Ask yourself:

  • Did the agent pull the correct information from the knowledge base?

  • Were deadlines, policies, and details accurate?

  • Did the response align with your institution’s official guidance?

Accuracy issues often point to gaps or improvements needed in your knowledge base.

2. Behavior

Consider:

  • Did it escalate when it should have?

  • Did it include unnecessary information?

  • Did it decide to follow up later?

  • What timeframe did it choose for a follow-up?

Behavioral changes can usually be adjusted by reconfiguring the workflow’s instructions, tools, or settings.


Best practices

  • Always test before launching a workflow

  • Review accuracy and behavior separately

  • Iterate based on test results, not assumptions

  • Re-run tests after making configuration or knowledge updates


Next steps

Once you are satisfied with how a workflow performs in the test harness, you can confidently move toward launching it live. If you need help interpreting test results or adjusting behavior, contact your CollegeVine partnerships team for guidance.

Testing early and often ensures your AI recruiter delivers accurate, helpful, and institution-aligned experiences from day one.

Did this answer your question?