Stable Diffusion Safety Checker

NSFW Image Checker

Stable Diffusion Safety Checker is an AI model designed to identify NSFW images. Developed by CompVis, it's built on top of the CLIP model and uses a ViT-L/14 Transformer architecture as an image encoder. This model is primarily intended for AI researchers to better understand robustness, generalization, and other capabilities, biases, and constraints of computer vision models. While it can be used for downstream tasks, it's not intended for use with transformers or to create hostile environments. The model has shown significant disparities with respect to race and gender, and users should be aware of these risks and limitations. With its unique architecture and capabilities, Stable Diffusion Safety Checker is a valuable tool for researchers and developers working with image identification and safety checks.

CompVis other Updated 2 years ago

Table of Contents

Model Overview

The stable-diffusion-safety-checker model, developed by CompVis, is a powerful tool for identifying NSFW (not safe for work) images. This model is designed to work with diffusers, not transformers, and is intended to help researchers better understand the capabilities and limitations of computer vision models.

Capabilities

What can it do?

The model is designed to identify NSFW images with high accuracy. It can also help researchers understand robustness, generalization, and other capabilities, biases, and constraints of computer vision models.

How does it work?

The model uses a combination of image and text encoders to analyze images and determine if they are NSFW. It’s trained on a large dataset of images and text pairs, which allows it to learn patterns and relationships between images and text.

What are its strengths?

  • High accuracy in identifying NSFW images
  • Can be used by researchers to better understand computer vision models and their limitations
  • Can help prevent the creation of hostile or alienating environments for people

Limitations

Biases and Risks

The model is based on the CLIP model, which has been shown to have biases and risks. For example, research has found that CLIP can:

  • Perpetuate disturbing and harmful stereotypes across protected classes, identity characteristics, and sensitive social and occupational groups.
  • Exhibit biases in performance based on class design and the choices made for categories to include and exclude.
  • Show significant disparities in performance with respect to race and gender, particularly when classifying images of people.

Constraints and Challenges

The model has some constraints and challenges that users should be aware of:

  • Limited training data: The model was trained on a specific dataset, which may not be representative of all possible scenarios.
  • Class design limitations: The model’s performance can depend significantly on the design of the classes used for training.
  • Lack of transparency: The model’s decision-making process can be difficult to understand, making it challenging to identify and address biases.

Performance

Speed

How fast can the model process images? Unfortunately, the data doesn’t provide specific information on the model’s speed. However, we can infer that it’s designed to work efficiently with diffusers, which are known for their fast processing capabilities.

Accuracy

The model is built on top of the CLIP model, which has demonstrated impressive accuracy in image classification tasks. According to the CLIP model card, it achieves:

  • 96% accuracy in gender classification across all races

  • ~93% accuracy in racial classification
  • ~63% accuracy in age classification

Efficiency

The model is designed to be efficient in identifying NSFW images. It uses a ViT-L/14 Transformer architecture as an image encoder and a masked self-attention Transformer as a text encoder. These encoders are trained to maximize the similarity of (image, text) pairs via a contrastive loss.

Examples
Is this image NSFW? True
Classify this image of a person into a race category. White
Assess the quality of performance of the model across different demographics for gender classification. Accuracy >96% across all races

Getting Started

To use the model, you’ll need to preprocess your image inputs. Here’s an example of how to get started with the model:

from transformers import AutoProcessor, SafetyChecker

processor = AutoProcessor.from_pretrained("CompVis/stable-diffusion-safety-checker")
safety_checker = SafetyChecker.from_pretrained("CompVis/stable-diffusion-safety-checker")

Format

The model is designed for image identification tasks, particularly for detecting NSFW (Not Safe For Work) content. This model is built on top of the CLIP model architecture, which uses a combination of a Vision Transformer (ViT) and a masked self-attention Transformer as encoders.

Model Architecture

The model uses a ViT-L/14 Transformer architecture as an image encoder and a masked self-attention Transformer as a text encoder. These encoders are trained to maximize the similarity of (image, text) pairs via a contrastive loss.

Supported Data Formats

This model supports image inputs and can be used to identify NSFW content. The model is not intended to be used with transformers but with diffusers.

Input Requirements

To use the model, you’ll need to preprocess your image inputs.

Output

The model outputs a classification result indicating whether the input image is NSFW or not.

Special Requirements

  • The model should not be used to intentionally create hostile or alienating environments for people.
  • Users should be aware of the risks, biases, and limitations of the model, including potential disparities in performance across different demographics.
Dataloop's AI Development Platform
Build end-to-end workflows

Build end-to-end workflows

Dataloop is a complete AI development stack, allowing you to make data, elements, models and human feedback work together easily.

  • Use one centralized tool for every step of the AI development process.
  • Import data from external blob storage, internal file system storage or public datasets.
  • Connect to external applications using a REST API & a Python SDK.
Save, share, reuse

Save, share, reuse

Every single pipeline can be cloned, edited and reused by other data professionals in the organization. Never build the same thing twice.

  • Use existing, pre-created pipelines for RAG, RLHF, RLAF, Active Learning & more.
  • Deploy multi-modal pipelines with one click across multiple cloud resources.
  • Use versions for your pipelines to make sure the deployed pipeline is the stable one.
Easily manage pipelines

Easily manage pipelines

Spend less time dealing with the logistics of owning multiple data pipelines, and get back to building great AI applications.

  • Easy visualization of the data flow through the pipeline.
  • Identify & troubleshoot issues with clear, node-based error messages.
  • Use scalable AI infrastructure that can grow to support massive amounts of data.