GPT OSS Safeguard 20B
GPT OSS Safeguard 20B is a 20-billion parameter open-source safety model from OpenAI, designed to classify and filter content for harmful or policy-violating material in AI application pipelines.
import { streamText } from 'ai'
const result = streamText({ model: 'openai/gpt-oss-safeguard-20b', prompt: 'Why is the sky blue?'})Frequently Asked Questions
Is GPT OSS Safeguard 20B a general-purpose language model?
No. It's a specialized safety classifier that evaluates content for harmful or policy-violating material. Use it alongside general-purpose models as a guardrail layer.
What types of content does it classify?
It evaluates text for categories of harmful content and policy violations. The specific categories are documented in the model's open-source materials.
Can I customize the safety criteria?
Open weights let teams inspect and adapt the classifier for organization-specific policies and safety requirements.
How do I integrate GPT OSS Safeguard 20B into my pipeline?
Deploy it as a filter layer that evaluates inputs before they reach your generation model and/or evaluates outputs before they reach users. Route requests through AI Gateway.
How does AI Gateway handle authentication for GPT OSS Safeguard 20B?
AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.
What are typical latency characteristics?
This page shows live performance metrics measured across real AI Gateway traffic.