From Idea to Millions of Documents, Simplified
Sutro takes the pain away from testing and scaling LLM batch jobs, letting you focus on extracting value from your documents.
import sutro as so
from pydantic import BaseModel
class ReviewClassifier(BaseModel):
sentiment: str
user_reviews = '.
User_reviews.csv
User_reviews-1.csv
User_reviews-2.csv
User_reviews-3.csv
system_prompt = 'Classify the review as positive, neutral, or negative.'
results = so.infer(user_reviews, system_prompt, output_schema=ReviewClassifier)
Progress: 1% | 1/514,879 | Input tokens processed: 0.41m, Tokens generated: 591k
█░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░
Prototype
Start small and iterate fast on your document abstraction workflows. Accelerate experiments by testing on Sutro before committing to large jobs.
Scale
Scale your LLM workflows so your team can do more in less time. Process billions of tokens in hours, not days, with no infrastructure headaches or exploding costs.
Integrate
Seamlessly connect Sutro to your existing LLM workflows. Sutro's Python SDK is compatible with popular data orchestration tools, like Airflow and Dagster.

Scale Effortlessly
Confidently handle millions of documents and billions of tokens at a time without the pain of managing infrastructure. Go from a small test to full-scale processing with ease.
Get results from massive document sets faster and reduce costs by 10x or more by parallelizing your LLM calls through Sutro's batch processing.

Shorten Development Cycles
Rapidly prototype your abstraction workflows. Get feedback from large batch jobs in as little as minutes before scaling up to unblock your most ambitious projects.