Test AI behavior before deployment
Reliability testing for RAG apps, agents, and LLM systems — like pytest for AI.
Catch unsupported claims, retrieval failures, missing abstentions, and output regressions before they reach production. Use the CLI today, and request beta access only if you want to connect to Veritell Cloud.
AI systems fail silently
Traditional testing validates code. Monitoring tools show what happened after the fact. But neither tells you whether the answer relied on retrieved context, whether the model should have refused, or whether a prompt or model change introduced a regression that will hit real users.
Use Veritell CLI today. Request beta access only for Veritell Cloud.
The CLI and public example suites are available now for local and CI workflows. Beta access is only required if you want to connect your tests to the managed Veritell Cloud environment.
1. Open the CLI repo
Start with the Veritell CLI source and installation guidance, then pair it with the public examples repo.
2. Run public examples
Use the examples repo to try grounded-answer, abstention, hallucination, and structured-output workflows.
3. Add Cloud access later
Request beta access only when you want to connect the CLI to the hosted Veritell Cloud experience.
Write tests for AI behavior
Define expected properties with assertions, run them like normal tests, and gate releases on reliability.
1. Define assertions
Describe the behaviors you care about: grounded answers, abstentions, schema validity, retrieval quality, and unsupported claims.
2. Run veritell test
Execute the suite locally, in CI, or against demo fixtures. Review per-test assertion results and suite reliability.
3. Gate deployment
Set a minimum score and fail the build when regressions, unsupported claims, or retrieval misses appear.
What Veritell detects
Unsupported claims
Answers contain information that is not grounded in the provided context.
Retrieval misses
Relevant documents were not returned, or the retrieved context does not support the answer.
Missing abstentions
The model answers when it should refuse or say it lacks enough evidence.
Schema violations
Structured outputs fail JSON or contract expectations.
Behavioral regressions
Prompt, model, or retrieval changes alter expected behavior.
RAG chatbots
Ensure answers stay grounded in retrieved documents and supported context.
AI agents
Validate task behavior, guardrails, and refusal logic before rollout.
Customer support assistants
Catch invented answers, policy violations, and unsafe responses before they reach users.
Structured-output systems
Guarantee JSON and schema stability for APIs, automations, and downstream integrations.
Gate deployments on reliability
Run Veritell in your pipeline and block releases when reliability drops below your team’s threshold.
veritell test --fail-below 80
Observability shows what AI did. Veritell tests what AI should do.
Monitoring helps you inspect production behavior after the fact. Veritell helps you define expected behavior before deployment — and enforce it in a repeatable workflow.
CLI available now, Cloud access in private beta
The Veritell CLI and public example suites are available now so teams can try the workflow, evaluate fit, and understand how reliability gates work in practice. Beta access is only required for teams that want to connect to Veritell Cloud.
Works locally first
Public examples and offline-friendly workflows make it easy to evaluate the CLI without standing up extra infrastructure.
Cloud when you need it
Request beta access when you want managed Veritell Cloud connectivity, shared workflows, or hosted integrations.
FAQ
Is Veritell only for RAG?
Can I use Veritell in CI/CD?
Do I need cloud infrastructure to get value?
Do I need beta access to use the CLI?
Is Veritell only for failures?
Does Veritell replace observability?
Request Veritell Cloud access
The CLI and public examples are available now. Request beta access only if you want to connect to the managed Veritell Cloud environment.