product updateOpenAI

OpenAI Releases Privacy Filter: 1.5B-Parameter On-Premises PII Detection Model with 128K Context

TL;DR

OpenAI has released Privacy Filter, a 1.5B-parameter bidirectional token classification model designed for detecting and masking personally identifiable information in text. The model processes 128,000-token sequences in a single pass and is available under Apache 2.0 license for on-premises deployment.

2 min read
0

OpenAI Releases Privacy Filter: 1.5B-Parameter On-Premises PII Detection Model with 128K Context

OpenAI has released Privacy Filter, a 1.5B-parameter bidirectional token classification model designed for detecting and masking personally identifiable information (PII) in text. The model is available under Apache 2.0 license and can run on-premises, including in web browsers.

Technical Architecture

Privacy Filter uses 1.5 billion total parameters with 50 million active parameters through a sparse mixture-of-experts architecture. According to OpenAI, the model was first pretrained autoregressively similar to GPT-OSS, then converted into a bidirectional token classifier with supervised classification training.

The architecture includes:

  • 8 transformer blocks with grouped-query attention (14 query heads, 2 KV heads)
  • Sparse mixture-of-experts feed-forward blocks (128 experts total, top-4 routing)
  • Bidirectional banded attention with band size 128 (effective 257-token attention window)
  • 640-dimensional residual stream width

Unlike autoregressive models that generate text token-by-token, Privacy Filter labels entire input sequences in one forward pass, then decodes coherent spans using a constrained Viterbi procedure.

Detection Capabilities

The model detects 8 privacy categories:

  • Account numbers
  • Private addresses
  • Private emails
  • Private persons (names)
  • Private phone numbers
  • Private URLs
  • Private dates
  • Secrets

For token-level classification, each category expands into BIOES (Begin, Inside, Outside, End, Single) boundary tags, producing 33 total output classes per token.

Context and Performance

Privacy Filter supports a 128,000-token context window, enabling processing of long documents without chunking. The model includes runtime controls for configuring precision-recall tradeoffs through adjustable operating points that modify span detection aggressiveness.

The sequence decoder uses six transition-bias parameters controlling background persistence, span entry, continuation, closure, and boundary handoff to produce coherent span boundaries rather than per-token independent predictions.

Deployment Options

OpenAI states the model can run in web browsers via WebGPU using Transformers.js with quantization (q4), or on laptops and on-premises infrastructure. The model is available through Hugging Face Transformers with standard pipeline API support.

The Apache 2.0 license permits commercial deployment and fine-tuning on specific data distributions.

Limitations Disclosed

OpenAI explicitly states Privacy Filter is "not an anonymization, compliance, or a safety guarantee" and warns against over-reliance. The model only identifies PII matching its trained taxonomy of 8 categories, which may not cover all privacy use cases or regulatory requirements. OpenAI recommends using it as one layer in a broader privacy-by-design approach rather than a standalone solution.

What This Means

This release addresses a specific enterprise need: fast, on-premises PII detection for data sanitization workflows where cloud APIs are unsuitable due to data residency or throughput requirements. The 128K context window and single-pass labeling design prioritize throughput over the iterative accuracy of larger models. The Apache 2.0 license and small parameter count make it accessible for fine-tuning on domain-specific PII patterns, though organizations must validate it meets their specific privacy requirements rather than treating it as a compliance checkbox.

Related Articles

product update

OpenAI replaces Custom GPTs with Codex-powered workspace agents for team collaboration

OpenAI is replacing Custom GPTs with workspace agents, Codex-powered tools designed for team collaboration that can handle complex tasks and long-running workflows. The feature launches today as a research preview for ChatGPT Business, Enterprise, Edu, and Teachers plans.

model release

OpenAI Releases GPT-5.4 Image 2 with 272K Context Window and Image Generation

OpenAI has released GPT-5.4 Image 2, combining the GPT-5.4 reasoning model with image generation capabilities. The multimodal model features a 272K token context window and is priced at $8 per million input tokens and $15 per million output tokens.

model release

OpenAI releases ChatGPT Images 2.0 with 3840x2160 resolution at $30 per 1M output tokens

OpenAI released ChatGPT Images 2.0, pricing output tokens at $30 per million with maximum resolution of 3840x2160 pixels. CEO Sam Altman claims the improvement from gpt-image-1 to gpt-image-2 equals the jump from GPT-3 to GPT-5.

model release

OpenAI releases ChatGPT Images 2.0 with integrated reasoning and text-image composition

OpenAI has released ChatGPT Images 2.0, which integrates reasoning capabilities to generate complex visual compositions combining text and images. The model supports aspect ratios from 3:1 to 1:3 and outputs up to 2K resolution, with advanced features available to Plus, Pro, Business, and Enterprise users.

Comments

Loading...