Skip to content

GPT OSS Safeguard 20B

GPT OSS Safeguard 20B is a 20-billion parameter open-source safety model from OpenAI, designed to classify and filter content for harmful or policy-violating material in AI application pipelines.

ReasoningTool Use
index.ts
import { streamText } from 'ai'
const result = streamText({
model: 'openai/gpt-oss-safeguard-20b',
prompt: 'Why is the sky blue?'
})

Frequently Asked Questions

  • Is GPT OSS Safeguard 20B a general-purpose language model?

    No. It's a specialized safety classifier that evaluates content for harmful or policy-violating material. Use it alongside general-purpose models as a guardrail layer.

  • What types of content does it classify?

    It evaluates text for categories of harmful content and policy violations. The specific categories are documented in the model's open-source materials.

  • Can I customize the safety criteria?

    Open weights let teams inspect and adapt the classifier for organization-specific policies and safety requirements.

  • How do I integrate GPT OSS Safeguard 20B into my pipeline?

    Deploy it as a filter layer that evaluates inputs before they reach your generation model and/or evaluates outputs before they reach users. Route requests through AI Gateway.

  • How does AI Gateway handle authentication for GPT OSS Safeguard 20B?

    AI Gateway accepts a single API key or OIDC token for all requests. You don't embed OpenAI credentials in your application; AI Gateway routes and authenticates on your behalf.

  • What are typical latency characteristics?

    This page shows live performance metrics measured across real AI Gateway traffic.