Streaming SSE
Get transcription results as they arrive via Server-Sent Events. Build real-time UIs, live captions, and voice-driven interfaces with sub-second latency.
Privocio's speech-to-text API is OpenAI-compatible by design. If your app already uses the OpenAI Whisper endpoint, switching to Privocio is a one-line change — same SDK, same request format, better privacy, and fixed pricing that doesn't punish growth.
Most speech APIs require custom audio preprocessing, format conversion, and error-handling boilerplate before you can ship a single feature.
Per-minute billing means costs scale unpredictably. A traffic spike or batch job can blow your budget overnight.
Proprietary SDKs and non-standard response formats mean migrating to another provider requires rewriting your integration from scratch.
Privocio implements the OpenAI audio transcription endpoint spec. Use the official OpenAI SDK in Python or TypeScript — just change the base URL and API key. Your existing code, error handling, and retry logic work without modification.
curl https://api.privocio.com/v1/audio/transcriptions \
-H "Authorization: Bearer pk_your_key" \
-F file=@meeting.mp3 \
-F model=whisper-1 \
-F response_format=jsonfrom openai import OpenAI
client = OpenAI(
base_url="https://api.privocio.com/v1",
api_key="pk_your_key",
)
result = client.audio.transcriptions.create(
model="whisper-1",
file=open("meeting.mp3", "rb"),
)
print(result.text)TypeScript, Go, and Ruby SDKs also supported via the OpenAI client libraries. See full documentation
Get transcription results as they arrive via Server-Sent Events. Build real-time UIs, live captions, and voice-driven interfaces with sub-second latency.
Submit large audio files for asynchronous processing. Get a webhook callback when your transcript is ready — ideal for podcast processing, call center archives, and media pipelines.
Configure webhook endpoints to receive transcription results automatically. No polling, no long-lived connections — just HTTP callbacks to your server.
Choose Raw for full-fidelity transcripts, Clean for human-readable text, or Agent for structured JSON optimized for LLM consumption.
Use the official OpenAI Python, TypeScript, or Go client. Change the base URL and you're done — no new dependencies.
Authenticate with a single API key. No OAuth flows, no token refresh, no session management — just a Bearer header.
Endpoint references, code examples, SDK guides, and webhook setup instructions — everything in one place.
OpenAI-compatible — switch with one line, keep your existing code
Fixed pricing every 4 weeks — no per-minute billing surprises
Private by default — audio never used for model training
Self-hosted option for teams that need full infrastructure control
Three output modes (Raw, Clean, Agent) for every use case
99.9% uptime SLA on hosted plans
Read the API documentation, try the free transcription tool, or pick a plan and start building.