YAML in.
Workbench out.
Promptfoo’s configuration is YAML. Parse `promptfooconfig.yaml`, map prompts and tests into Prompt Assay’s library and eval surface, and land in a workbench with no parent-company tilt and no per-trace tax.
Three steps. Then you ship.
- Station · 01
Parse
Your existing `promptfooconfig.yaml` is the migration source of truth. Parse the file with PyYAML or any YAML library; the structure is well-defined. The fields you care about: `prompts` (the prompt strings or files), `providers` (the model configs), `tests` (the test cases with `vars` and `assert` blocks), and `redteam` (if you used the security probe sets). Keep the YAML around · the parse is non-destructive.
- Station · 02
Map
Each `prompts` entry becomes a new prompt in Prompt Assay's library; preserve variable interpolation as `{{var}}` placeholders. Each `tests` entry becomes a test case in an eval suite, with the `assert` blocks mapping to graders (rubric-based or LLM-as-a-judge depending on the assertion shape). `providers` configs become run-time provider selections; BYOK keys connect directly to Anthropic, OpenAI, Google with no parent-company tilt. `redteam` configurations are not Prompt Assay's lane · pair with a security-testing surface for that workflow.
- Station · 03
Land
Sign up for Prompt Assay. Create a prompt for each parsed entry; paste the body in as version one and re-add variables. Build the eval suite, attach test cases and graders, and run it. The AI pair (Brainstorm, Critique, Improve, Rewrite, Compare) is available from the first save; six-dimension critique runs on every prompt before it ships.
Four rows. Five facts each.
Each row is a real option. Pick the row that matches your eval-CI shape and your sensitivity to OpenAI's roadmap stewardship.
- Platform fee
- Open-source CLI free (MIT-licensed). Enterprise tier custom; pricing not publicly disclosed post-acquisition. Acquired by OpenAI on Mar 9 2026; deal value not disclosed. OpenAI committed to 'continue building out' the open-source offering.
- Provider scope
- Multi-provider via YAML-defined providers in `promptfooconfig.yaml`. Historical scope: 250+ providers including Anthropic, OpenAI, Google, Bedrock, local models. Future non-OpenAI provider commitment is not publicly stated.
- Inference path
- Direct to provider. CLI runs locally; provider keys never leave your machine.
- Promptfoo import
- YAML-driven test cases. Prompt strings, eval assertions, and provider configs all in `promptfooconfig.yaml`. Migration is parsing the YAML and re-creating the prompt + eval suite in the new tool.
- Best fit
- AI security and red-teaming workflows that integrate into OpenAI Frontier post-acquisition. Multi-provider eval users now subject to OpenAI's roadmap.
- Platform fee
- Starter free (1GB processed, 10K scores). Pro $249/mo (5GB, 50K scores, 30-day retention). Enterprise custom. Closed-source, hosted-only on Pro; self-host is Enterprise-only.
- Provider scope
- Multi-provider via AI Gateway accepting OpenAI, Anthropic, and Google SDKs.
- Inference path
- BYOK supported, but inference flows through Braintrust's gateway. Provider keys still pass through their proxy.
- Promptfoo import
- Prompts are TypeScript-defined. Promptfoo migration requires manual transformation: parse YAML, re-author prompts in TS, port assertions to Braintrust's eval format.
- Best fit
- Well-funded eval-driven teams who want a managed gateway plus eval CI surface and accept proxied inference.
- Platform fee
- Hobby free (50K units/mo) · Core $29/mo · Pro $199/mo · Enterprise $2,499/mo. Self-host free under MIT.
- Provider scope
- Provider-neutral via OpenTelemetry. Multi-provider tracing and prompt management.
- Inference path
- Direct to provider. OTel instrumentation; no proxy in the inference path.
- Promptfoo import
- Datasets API accepts re-imported test cases. Promptfoo YAML must be parsed and pushed via Public API.
- Best fit
- Open-source-first teams who want OTel-native tracing alongside dataset and eval primitives, with self-host as a clean compliance fallback.
- Our entry
Prompt Assay
- Platform fee
- Free tier · Solo $49/mo · Team $99/seat/mo · Enterprise contact-sales. BYOK-mandatory at every paid tier with no inference markup.
- Provider scope
- Anthropic, OpenAI, Google with first-class adapters and no parent-company tilt.
- Inference path
- Direct to provider. We never sit in the inference request path. Your bill stays with your provider account.
- Promptfoo import
- No first-class YAML parser. Copy each Promptfoo prompt into a new Prompt Assay prompt; eval suites land in the workbench's evaluation surface (test cases + rubrics + LLM-as-a-judge).
- Best fit
- Multi-provider teams who picked Promptfoo for OSS neutrality and want a workbench that pairs authoring (six-dimension critique, two-version Compare) with the eval surface in a single tool, with no parent-company provider preference.
Verified 2026-05-01 · Read the full breakdown
Frequently asked.
- Is the Promptfoo OSS license going away after the OpenAI acquisition?
- No. The MIT license is preserved per OpenAI's announcement, and OpenAI stated it 'expects to continue building out Promptfoo's open source offering.' The forward-looking framing matters · OpenAI has not committed publicly to maintaining feature parity for non-OpenAI providers (Anthropic, Google, Bedrock, local models) going forward, and the integration target is OpenAI Frontier. The migration question is roadmap stewardship, not platform death.
- Do I have to migrate? Can I keep using Promptfoo alongside Prompt Assay?
- You don't have to. Many teams already run both today · YAML-defined eval CI in Promptfoo for pre-deploy checks, plus a hosted tool for production observability and authoring. Adding Prompt Assay alongside is a low-cost first step: keep Promptfoo for what it does well, and add the workbench surface upstream for authoring, six-dimension critique, two-version Compare, and prompt-level versioning. The decision to fully retire Promptfoo can come later when the roadmap signal is clearer.
- What about the redteam configurations?
- Red-teaming and AI-agent security probing are Promptfoo's strongest surface, and they are not Prompt Assay's lane. If your primary use was red-teaming, the migration calculus is different · pair Prompt Assay (or any other workbench) with a dedicated security-testing tool for the red-team surface. OpenAI Frontier is one option post-acquisition; there are independent security-research tools that occupy this space too.
- Will my assertion library carry over?
- The shape carries over; the syntax does not. Promptfoo's `assert` blocks (icontains, javascript, llm-rubric, model-graded-closedqa) map to Prompt Assay eval-suite graders · rubric checks land as scoring rubrics, LLM-as-a-judge checks land as judge-model graders, deterministic string checks land as expected-output assertions on the test case. The mapping is one-to-one in concept; the YAML-to-UI configuration step is what takes the time.
- Why pick Prompt Assay over Braintrust or Langfuse for this migration?
- Three honest reasons. Authoring depth · Prompt Assay covers the prompt-iteration loop upstream of evals (six-dimension critique, two-version Compare with model-graded diff, AI pair) where Braintrust and Langfuse are eval-and-tracing-first. Provider neutrality · BYOK at every paid tier with no inference markup or proxy in the request path; Braintrust runs a managed gateway, Langfuse is OTel-native but optimizes for self-host or hosted at higher tiers. Pricing shape · Prompt Assay Team is flat at $99 per seat per month with no per-trace tax, no per-GB-processed billing, and no per-score overage.
Land in the workbench.
Free to start. Your keys, your bill, no parent-company tilt.