From Documentation to FAQ in Three Steps
Sutro's Python SDK simplifies generating FAQs from your existing content. Our workflow lets you start small and scale up confidently.
import sutro as so
from pydantic import BaseModel
class ReviewClassifier(BaseModel):
sentiment: str
user_reviews = '.
User_reviews.csv
User_reviews-1.csv
User_reviews-2.csv
User_reviews-3.csv
system_prompt = 'Classify the review as positive, neutral, or negative.'
results = so.infer(user_reviews, system_prompt, output_schema=ReviewClassifier)
Progress: 1% | 1/514,879 | Input tokens processed: 0.41m, Tokens generated: 591k
█░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░
Prototype
Start with a small sample of your content. Iterate fast on your prompts to find the perfect structure and tone for your FAQs before committing to large jobs.
Scale
Scale your workflow to process millions of web pages or your entire document corpus. Process billions of tokens in hours, not days, with no infrastructure headaches.
Integrate
Seamlessly connect Sutro to your existing LLM workflows. Our Python SDK is compatible with popular data orchestration tools, like Airflow and Dagster.

Scale Effortlessly
Go from a single document to millions. Sutro handles the infrastructure, letting you generate FAQs for your entire knowledge base without performance bottlenecks or the pain of managing infrastructure.
By parallelizing your LLM calls through Sutro, you can get results faster and dramatically reduce the cost of generating FAQs at scale.

Shorten Development Cycles
Rapidly prototype your FAQ generation workflow on large batches. Get feedback in minutes before scaling up to accelerate experiments and ship faster.