Graph Inference
After training a graph with ADAS or Graph GEPA, you can run production inference through the Synth API. The graph executes server-side using your optimized prompts and structure.API Endpoint
Request
| Field | Type | Required | Description |
|---|---|---|---|
job_id | string | Yes | ADAS job ID that produced the graph |
input | object | Yes | Input matching your dataset’s input schema |
model | string | No | Override the policy model for this call |
prompt_snapshot_id | string | No | Use a specific snapshot instead of best |
Response
Python SDK
Using ADASJob
Override Model
Run the same graph with a different model:Use Specific Snapshot
Target a particular prompt version from training:cURL Example
Batch Inference
For multiple inputs, call the endpoint in parallel:Verifier Graph Inference
Verifier graphs (custom judges) use a different endpoint that accepts V3 traces and rubrics.API Endpoint
Request
| Field | Type | Required | Description |
|---|---|---|---|
job_id | string | Yes | Verifier graph job ID |
session_trace | object | Yes | V3 SessionTrace to evaluate |
context | object | No | Context including rubric |
context.rubric | object | No | Evaluation criteria |
Response
Python SDK
Using Verifier Output for RL
Verifier graphs integrate with reinforcement learning training. The structured output maps directly to synth-ai reward tables:Input Schema
Your inference inputs must match the schema from your training dataset:Error Handling
Pricing
Graph inference is billed per execution:- Base cost: Per-graph execution fee
- LLM costs: Pass-through from underlying model calls
- Multi-node graphs: Each node’s LLM call is billed separately
Related
- Graphs Overview - What graphs are
- Downloading Graphs - Run graphs locally
- Workflows - Full product documentation