From Idea to Millions of Requests, Simplified
Sutro simplifies the entire batch workflow, letting you start small, test your data generation prompts, and scale up to massive jobs with ease.
import sutro as so
from pydantic import BaseModel
class ReviewClassifier(BaseModel):
sentiment: str
user_reviews = '.
User_reviews.csv
User_reviews-1.csv
User_reviews-2.csv
User_reviews-3.csv
system_prompt = 'Classify the review as positive, neutral, or negative.'
results = so.infer(user_reviews, system_prompt, output_schema=ReviewClassifier)
Progress: 1% | 1/514,879 | Input tokens processed: 0.41m, Tokens generated: 591k
█░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░
Prototype
Start small and iterate fast on your data generation prompts. Accelerate experiments by testing on Sutro with a small batch before committing to a large job, getting feedback in as little as minutes.
Scale
Scale your synthetic data generation to process billions of tokens in hours, not days. Sutro handles the infrastructure so your team can do more in less time without exploding costs.
Integrate
Seamlessly connect Sutro to your existing LLM workflows. Sutro's Python SDK is compatible with popular data orchestration tools like Airflow and Dagster.

Improve Model Performance
Generate high-quality, diverse, and representative synthetic data to fill statistical gaps and improve your LLM or RAG retrieval performance, without the complexity.
Get results faster and dramatically lower your expenses. Sutro parallelizes your LLM calls to generate data at a fraction of the cost of other methods.

Scale Effortlessly
Go from idea to millions of data points without the pain of managing infrastructure. Confidently handle billions of tokens at a time for your most ambitious AI projects.