top of page

What is Zero Shot Classification?

Zero-shot classification in machine learning: faceless user at a computer where an AI, given the prompt “an animal with black-and-white stripes,” correctly selects Zebra over Cat and Dog.

Imagine teaching a child to recognize animals by showing them pictures of dogs and cats. Then you ask them to identify a zebra—something they've never seen before. Yet somehow, by understanding "striped" and "horse-like," they get it right. That's essentially what zero shot classification does for machines, and it's changing how we build AI systems.




TL;DR

  • Zero shot classification allows AI models to categorize data into classes without seeing any training examples of those classes


  • Models leverage pre-trained knowledge and semantic relationships to classify unseen categories


  • GPT-3 achieved 76% accuracy on LAMBADA dataset in zero-shot mode in 2020 (Brown et al., NeurIPS 2020)


  • Real-world uses include content moderation (98% accuracy), medical diagnosis, customer support automation, and document classification


  • Major platforms like Google, Microsoft, and Hugging Face deploy zero-shot models at scale


  • Challenges include accuracy gaps versus fine-tuned models and potential bias toward seen classes


Zero shot classification is a machine learning technique where a pre-trained model classifies data into categories it has never seen during training. By using semantic embeddings, auxiliary information like text descriptions, and transfer learning from large-scale datasets, models can generalize to entirely new classes without requiring labeled examples, dramatically reducing data annotation costs and deployment time.


Table of Contents

What Exactly Is Zero Shot Classification?

Zero shot classification turns traditional machine learning on its head.


In standard supervised learning, you need hundreds or thousands of labeled examples for every category you want to recognize. Want to classify customer emails into 20 categories? You need labeled examples for all 20.


Zero shot classification breaks free from this constraint.


The technique allows pre-trained models to recognize and categorize objects or concepts without having seen examples of those categories beforehand (IBM, 2025-07-31).


Here's what makes it powerful: the model receives a prompt and a sequence of text describing the task in natural language, along with candidate labels, and can perform classification without any training examples of the desired outcome (Hugging Face, 2024).


Core components:

  • Pre-trained knowledge base - Models trained on massive datasets (millions of images or billions of text tokens)


  • Semantic embeddings - Vector representations that capture meaning and relationships between concepts


  • Auxiliary information - Text descriptions, attributes, or knowledge graphs that describe unseen classes


  • Transfer learning - Applying knowledge from seen classes to predict unseen ones


The breakthrough came from a simple insight: if a model deeply understands concepts like "striped," "four-legged," and "African animal," it can classify a zebra even if it never saw zebra images during training.


The History Behind Zero Shot Learning

Zero shot learning has roots going back over a decade, but recent advances in transformer models catalyzed its mainstream adoption.


2008-2013: Early computer vision work

The idea of zero-data learning dates back over a decade but was mostly studied in computer vision as a way of generalizing to unseen object categories (OpenAI CLIP documentation). Researchers explored using attributes to describe classes—like "has stripes" or "is furry"—so models could recognize new animals by combining known attributes.


2013: Natural language as prediction space

Richard Socher and co-authors at Stanford developed a proof of concept by training a model on CIFAR-10 to make predictions in a word vector embedding space, showing the model could predict two unseen classes (OpenAI CLIP, 2021).


2016: Scaling with Flickr

Ang Li and co-authors at FAIR demonstrated using natural language supervision to enable zero-shot transfer, fine-tuning an ImageNet CNN to predict visual concepts from 30 million Flickr photo descriptions and achieving 11.5% accuracy on ImageNet zero-shot (OpenAI CLIP, 2021).


2019: BART for NLI

BART was released on 2019-10-29 and added to Hugging Face Transformers on 2020-11-16, combining pretraining objectives from BERT and GPT by learning to recover corrupted text (Hugging Face BART documentation).


2020: GPT-3 demonstrates massive scale

GPT-3 was released by OpenAI in 2020 with 175 billion parameters and demonstrated strong zero-shot and few-shot learning abilities on many tasks without fine-tuning (Wikipedia, 2025).


GPT-3 achieved 83.2% accuracy in zero-shot setting on the StoryCloze 2016 dataset and 76% on LAMBADA in zero-shot mode (Springboard, 2023).


2021: CLIP bridges vision and language

OpenAI's CLIP model trained on 400 million image-text pairs from the internet, enabling zero-shot image classification through natural language prompts.


2024-2025: Production deployments

Large language models like GPT-4o-Mini and GPT-3.5-Turbo achieved near human-level accuracy for content moderation across multiple harm categories in few-shot settings (Yang et al., arXiv 2025-01-23).


Microsoft's Community Sift AI Moderator reached over 98% accuracy on Xbox content using AI-powered zero-shot classification in early 2024 (Microsoft Developer, 2024).


How Zero Shot Classification Actually Works

Zero shot classification relies on three interconnected mechanisms: semantic embeddings, transfer learning, and natural language inference.


Semantic Embedding Space

Models transform both inputs (text, images) and class labels into high-dimensional vector spaces where similar concepts cluster together.


For text:

  • Pre-trained language models like BERT or GPT convert words into embeddings

  • The sentence "I love this product" becomes a 768-dimensional vector

  • Labels like "positive," "negative," "neutral" also become vectors

  • The model measures similarity between input and label vectors


For images:

  • Convolutional neural networks extract visual features

  • A photo of a cat becomes a feature vector capturing shapes, textures, colors

  • Class descriptions like "small furry domestic animal" become vectors

  • Similarity scores determine the most likely class


Natural Language Inference (NLI)

The method works by posing the sequence to be classified as the NLI premise and constructing a hypothesis from each candidate label, then converting probabilities for entailment and contradiction into label probabilities (Yin et al., cited by Metatext).


Example:

  • Premise: "This movie made me cry tears of joy"

  • Hypothesis 1: "This example is about positive sentiment" → High entailment

  • Hypothesis 2: "This example is about negative sentiment" → Contradiction

  • Result: Classified as positive


This approach leverages models trained on datasets like MultiNLI, which contains 433,000 sentence pairs annotated with textual entailment information (Hugging Face, 2024).


Transfer Learning from Pre-Training

Transfer learning is used prominently in zero-shot methods that represent classes and samples as semantic embeddings, with models like BERT pre-trained on massive corpora of language data to convert words into vector embeddings (IBM, 2025-07-31).


The model's pre-training phase teaches it:

  • Linguistic patterns and grammar

  • Semantic relationships between concepts

  • World knowledge from billions of text tokens

  • Visual patterns from millions of images


During inference, this accumulated knowledge transfers to new tasks without additional training.


Key Technologies and Models


Transformer Architecture

The transformer is a neural network architecture designed to interpret meaningful representations of sequences using an encoder-decoder structure and self-attention mechanism that computes weights for each token based on relationships to every other token (IBM, 2025-04-17).


Key components:

  • Self-attention - Weighs importance of each word relative to others

  • Multi-head attention - Captures multiple types of relationships simultaneously

  • Position encoding - Preserves word order information

  • Feed-forward layers - Transform representations through learned patterns


BART (Bidirectional and Auto-Regressive Transformers)

BART combines pretraining objectives from BERT and GPT, pretrained by corrupting text in different ways like deleting words, shuffling sentences, or masking tokens and learning to recover the original (Hugging Face BART documentation).


facebook/bart-large-mnli:

  • Pre-trained on 400 million parameters

  • Fine-tuned on MultiNLI dataset for entailment tasks

  • Most popular model for zero-shot text classification

  • Handles multi-label classification

  • Available through Hugging Face Transformers library


GPT-3 and GPT-3.5

GPT-3 has 175 billion parameters with 16-bit precision requiring 350GB storage, with a context window of 2,048 tokens (Wikipedia, 2025).


Performance highlights:

  • 83.2% on StoryCloze 2016 (zero-shot), 87.7% (few-shot with K=70)

  • 76% on LAMBADA zero-shot, gaining 8% over previous state-of-the-art

  • 78.1% on HellaSwag one-shot, 79.3% few-shot


BERT Family Models

The BERT family has witnessed explosive development with variants like RoBERTa, ALBERT, and ELECTRA (ScienceDirect, 2023-06-05).


RoBERTa: RoBERTa classifiers consistently outperform GPT-3 zero-shot and few-shot queries across all levels of domain-specific pre-training and fine-tuning (Bosley et al., MPSA 2023).


Limitations: Prompt-based methods fail in more challenging natural language understanding tasks like GLUE and SuperGLUE benchmarks, though achieving promising results in zero-shot text classification (ScienceDirect, 2023-06-05).


CLIP (Contrastive Language-Image Pre-training)

CLIP was trained on a wide variety of images with natural language supervision from the internet, enabling zero-shot image classification through natural language instructions (OpenAI, 2021).


Performance:

  • CLIP achieves 64.3% accuracy on ImageNet dataset in zero-shot mode (Capa Learning, 2025-03-02)

  • Outperforms best publicly available ImageNet model on 20 out of 26 transfer datasets tested

  • Achieves only 88% accuracy on MNIST handwritten digits, well below 99.75% human accuracy


LlamaGuard 3

LlamaGuard 3 is a Llama-3.1-8B model fine-tuned for content safety classification supporting 14 predefined categories of harmful content, extendable through zero-shot learning for new safety categories (Neural Engineer blog, 2024-10-11).


Real-World Case Studies

Case Study 1: COVID-19 Diagnosis from Chest X-Rays (2020)

Organization: Research collaboration documented in Intelligence-Based Medicine journal


Challenge: Building an AI model to diagnose COVID-19 without providing visual exemplars in the training phase, using only side auxiliary information like medical text descriptions and images of related respiratory diseases (PMC7531283, 2020).


Approach: Used semantic relationships between training data of Asthma, Pneumonia, and SARS with written medical documents and chest X-ray images as auxiliary data, seeking semantic relationships to infer the novel COVID-19 cases (PMC7531283, 2020).


Key technique: Applied zero-shot segmentation to identify white ground-glass opacities in lungs captured by chest X-rays and CT scans when labeled segmented images of COVID-19 cases were scarce (V7 Labs blog).


Outcome: Successfully identified COVID-19 patterns without requiring large labeled datasets of positive COVID-19 cases, enabling faster deployment during the early pandemic when labeled data was extremely limited.


Source: COVID-ChestXRay dataset (Cohen et al., 2020), Intelligence-Based Medicine journal


Case Study 2: Microsoft Xbox Content Moderation (2024)

Organization: Microsoft Gaming / Community Sift


Challenge: Scale content moderation across Xbox Live platform handling millions of user-generated messages, images, and voice communications daily while maintaining 98%+ accuracy.


Solution: Community Sift released AI Moderator (AI Mod) that makes decisions on content based on customer-specific policies, reduces harmful content seen by human moderators, and enables scaling (Microsoft Developer, 2024-03).


Technology:

  • Generative AI-powered classification

  • Custom policy configuration per game/community

  • Real-time proactive moderation before content reaches players

  • Handles "obvious" violations automatically, escalates "gray area" content to humans


Results: In the past month before March 2024, Community Sift's AI Mod solution reached over 98% accuracy on Xbox content (Microsoft Developer, 2024).


Impact:

  • Reduced human moderator exposure to traumatic content

  • Faster moderation response times (pre-publication blocking)

  • Consistent policy enforcement across diverse game titles

  • Compliance with EU Digital Services Act requirements


Source: Microsoft Game Developer Conference 2024


Case Study 3: Google Ads Policy Enforcement (December 2024)

Organization: Google Ads


Challenge: Moderate massive volumes of advertising images across diverse content categories with evolving policies, without retraining models for every new policy.


Approach: Utilized human-curated textual descriptions and cross-modal text-image co-embeddings to enable zero-shot classification of policy violating ads images, bypassing need for extensive supervised training data (Google Research, arXiv 2024-12-18).


Implementation:

  1. Used LLMs to generate candidate text descriptions of policy violations

  2. Refined descriptions through human expertise

  3. Computed similarity between ad images and violation descriptions

  4. Escalated ambiguous cases to fine-tuned LLM review

  5. Final uncertain cases sent to human review


Advantages: Minimal training data required, fast turnaround time with no model training needed, and resource efficiency using same workflow for multiple policies with one scalable search (Google Research, arXiv 2024-12-18).


Outcome: Significantly boosted detection of policy-violating tobacco-related content and other restricted categories while reducing time from policy definition to enforcement.


Source: Google Research paper, arXiv:2412.16215, December 2024


Case Study 4: Healthcare Rare Disease Diagnosis (2024)

Organization: Healthcare organization (documented in XCube Labs case study)


Problem: Early diagnosis of rare diseases is challenging due to limited availability of labeled data, with traditional ML models requiring extensive data to achieve high accuracy (XCube Labs, 2024-09-10).


Solution: Implemented few-shot learning by leveraging a pre-trained model on large dataset of common diseases, then fine-tuning on small dataset of rare diseases (XCube Labs, 2024-09-10).


Results: Few-shot learning models achieved 87% accuracy in diagnosing rare diseases with minimal data in a 2023 study (XCube Labs, 2024-09-10).


Impact: Enabled earlier intervention for patients with rare conditions, reduced diagnostic delays, and lowered costs of collecting extensive training data for uncommon diseases.


Source: XCube Labs blog, September 2024


Step-by-Step Implementation Guide


Using Hugging Face BART for Zero-Shot Text Classification

Prerequisites:

  • Python 3.7+

  • transformers library

  • torch (PyTorch)


Step 1: Install dependencies

pip install transformers torch

Step 2: Initialize the pipeline

The facebook/bart-large-mnli model is a powerful zero-shot text classification model available through Hugging Face Transformers (Hugging Face, 2024).

from transformers import pipeline

classifier = pipeline(
    "zero-shot-classification",
    model="facebook/bart-large-mnli",
    device=0  # Use GPU if available, -1 for CPU
)

Step 3: Define your text and candidate labels

text = "The quarterly earnings report exceeded analyst expectations by 15%."

candidate_labels = [
    "finance",
    "technology",
    "healthcare",
    "sports",
    "politics"
]

Step 4: Run classification

result = classifier(
    text,
    candidate_labels,
    multi_label=False  # Set True for multi-label classification
)

print(f"Labels: {result['labels']}")
print(f"Scores: {result['scores']}")

Output:

Labels: ['finance', 'technology', 'politics', 'sports', 'healthcare']
Scores: [0.892, 0.054, 0.029, 0.015, 0.010]

Step 5: Handle multi-label classification

Set multi_label value to True when the text could reasonably belong to multiple categories simultaneously (Medium, 2023-08-14).

text = "The tech company reported strong earnings amid regulatory scrutiny."

result = classifier(
    text,
    ["finance", "technology", "politics", "legal"],
    multi_label=True
)

# Now each label gets independent probability
for label, score in zip(result['labels'], result['scores']):
    if score > 0.5:  # Threshold for positive classification
        print(f"{label}: {score:.3f}")

Step 6: Production deployment

Load the model once when server starts, then reuse the pipeline without re-initializing for each request to improve performance (Stack Overflow, 2024).

# server_startup.py
from transformers import pipeline

# Initialize once
global_classifier = pipeline(
    "zero-shot-classification",
    model="facebook/bart-large-mnli"
)

# api_endpoint.py
def classify_request(text, labels):
    return global_classifier(text, labels)

Creating Custom Label Descriptions

Tip: More specific, descriptive labels improve accuracy.


Poor labels:

  • "good"

  • "bad"

  • "neutral"


Better labels:

  • "positive customer experience"

  • "negative customer complaint"

  • "neutral factual statement"


Best labels (with context):

  • "customer expressing satisfaction with product quality"

  • "customer reporting product defect or service failure"

  • "customer asking factual question about product features"


Performance Benchmarks and Accuracy


Major Benchmark Datasets

Benchmark datasets for zero-shot learning include aPY, AwA (Animals with Attributes), CUB-200-2011 (Caltech-UCSD Birds), ImageNet, CIFAR-100, and others (Papers with Code).


Common datasets:

Dataset

Classes

Images

Domain

Use Case

ImageNet

1,000+

14M

General objects

Benchmark standard

CUB-200-2011

200

11,788

Bird species

Fine-grained classification

AwA2

50

37,322

Animal species

Attribute-based ZSL

aPY

32

15,339

Multiple categories

General ZSL evaluation

MNIST

10

70,000

Handwritten digits

Baseline comparison

MultiNLI

N/A

433k pairs

Text entailment

NLP zero-shot

Source: Papers with Code, 2024


Model Performance Comparison

Text Classification (MultiNLI-trained models):

Model

Parameters

Zero-Shot Accuracy

Speed (examples/sec)

facebook/bart-large-mnli

400M

High on general text

~100

roberta-large-mnli

355M

Similar to BART

~120

GPT-3

175B

Variable by task

API-dependent

GPT-4o-Mini

Undisclosed

Near human-level

API-dependent

Image Classification:

Model

Training Data

ImageNet Zero-Shot

Notes

CLIP ViT-B/32

400M pairs

64.3%

General purpose

CLIP ViT-L/14

400M pairs

75.5%

Larger variant

ResNet-50 (supervised)

ImageNet-1K

N/A (not zero-shot)

Baseline: 76.2%

Source: OpenAI CLIP paper (2021), Capa Learning (2025)


Task-Specific Performance

GPT-3 Zero-Shot Results (Brown et al., NeurIPS 2020):

Task

Dataset

Zero-Shot

Few-Shot

Previous SOTA

Story completion

StoryCloze

83.2%

87.7%

87.3%

Reading comprehension

LAMBADA

76.0%

86.4%

68.0%

Instruction following

HellaSwag

78.1%

79.3%

75.4%

Source: GPT-3 paper (Brown et al., 2020), Springboard (2023)


Content Moderation Accuracy:

Microsoft Community Sift AI Moderator reached over 98% accuracy on Xbox content in early 2024.


Accuracy Limitations

CLIP struggles on abstract or systematic tasks like counting objects and on complex tasks like predicting nearest car distance, performing only slightly better than random guessing on these datasets (OpenAI, 2021).


CLIP achieves only 88% accuracy on MNIST handwritten digits despite learning capable OCR, well below 99.75% human accuracy (OpenAI, 2021).


Key insight: Zero-shot models trade raw accuracy for flexibility and generalization.


Industry Applications by Sector


Customer Support & Service

Use cases:

  • Ticket routing - Classify support tickets into departments without training on every possible issue

  • Sentiment analysis - Gauge customer emotions in real-time chat

  • Intent detection - Understand what customers want without pre-defining all intents

  • Language detection - Identify customer language dynamically


Benefits:

  • Instant deployment for new product lines

  • Handles emerging issues without retraining

  • Reduces setup time from weeks to hours


Content Moderation

Applications: Platforms like Facebook, Twitter, and Reddit use content moderation to classify content as acceptable or unacceptable, with model-based moderation using statistical models to scale beyond human review (Roboflow, 2024-09-10).


Zero-shot advantages:

  • New policy enforcement without collecting violation examples

  • Handles novel forms of harmful content

  • Adapts to evolving community standards

  • Reduces moderator exposure to traumatic material


Real deployment: LLMs like GPT-4o-Mini and Mistral-7B classify harmful content across six categories including Information, Hate and Harassment, Addictive, Clickbait, Sexual and Physical Harms on YouTube videos (arXiv 2025-01-23).


Healthcare & Medical Imaging

Applications: Generalized zero-shot learning predicts both seen and novel unseen disease classes in multi-label chest X-ray classification, leveraging feature disentanglement and multi-modal information with text embeddings from BioBert (IEEE Transactions on Medical Imaging, 2025-01).


Use cases:

  • Rare disease identification

  • New pathogen detection (COVID-19 early diagnosis)

  • Medical image segmentation without full annotation

  • Cross-institutional deployment without local data


Advantage: Faster clinical deployment when labeled medical data is scarce or requires expert annotation.


E-Commerce & Retail

Applications: Product categorization where e-commerce platforms automatically classify products into relevant categories, enabling better organization and search functionalities (Spot Intelligence, 2024-10-07).


Use cases:

  • Automatic product tagging

  • Review sentiment classification

  • Customer query categorization

  • Dynamic inventory classification


Benefit: New product categories added without retraining classification models.


Finance & Banking

Zero-shot and few-shot learning used in finance to identify fraud, assess risk, and provide personalized financial services, with ability to quickly adapt to new fraud patterns with minimal data (XCube Labs, 2024-09-10).


Applications:

  • Transaction classification

  • Fraud pattern detection

  • Document processing (invoices, contracts)

  • Regulatory compliance monitoring


Advantage: Rapid adaptation to emerging fraud tactics without waiting for large labeled datasets.


Autonomous Vehicles

Detecting novel objects and knowing how to respond to them is essential in autonomous navigation, where seeing a car/truck/bus means avoiding them, and a red traffic light means stopping (V7 Labs).


Requirements:

  • Real-time object classification

  • Handling rare or unusual objects

  • Safety-critical decisions

  • Continuous learning from environment


Multilingual NLP

Language identification where zero-shot classification can identify the language of given text, allowing multilingual applications to adapt to different languages dynamically (Spot Intelligence, 2024-10-07).


Applications:

  • Cross-lingual document classification

  • Multilingual customer support

  • Translation quality assessment

  • Content localization


Case reference: Google's use of zero-shot learning for multilingual text classification demonstrates practical application in production systems (Lyzr AI, 2025-03-15).


Comparison: Zero-Shot vs Few-Shot vs Fine-Tuning

Aspect

Zero-Shot

Few-Shot

Fine-Tuning

Training examples

0

1-100

1,000-1,000,000+

Setup time

Minutes

Hours

Days to weeks

Labeled data cost

$0

$10-$1,000

$10,000-$1,000,000+

Accuracy

Lower

Medium

Highest

Flexibility

Highest

High

Low

Computational cost

Low

Low-Medium

High

Best for

Rapid prototyping, new categories

Limited data scenarios

Production systems

Model updates

Instant

Quick

Requires retraining

When to Use Each Approach

Choose Zero-Shot when:

  • You need instant deployment

  • Categories change frequently

  • Labeled data is unavailable or expensive

  • You're prototyping or exploring feasibility

  • Handling long-tail or rare categories


Choose Few-Shot when:

  • You have 5-100 examples per class

  • Need better accuracy than zero-shot

  • Data labeling budget is limited

  • Classes are somewhat similar to pre-training data

  • A 2023 study found few-shot learning models can reduce time to detect new fraud patterns by 50% compared to traditional methods (XCube Labs, 2024)


Choose Fine-Tuning when:

  • Accuracy is paramount

  • You have 1,000+ labeled examples per class

  • Task is business-critical

  • Categories are stable

  • Computational resources available


Advantages and Limitations


Advantages

  1. Dramatic reduction in data requirements

    Zero-shot learning significantly reduces need for extensive data annotation, which can be expensive and time-consuming, making it valuable for scenarios where labeled data is scarce (Spot Intelligence, 2024-10-07).


    Cost comparison:

    • Traditional supervised learning: $50,000-$500,000 for data labeling

    • Zero-shot classification: $0 for training data (pre-trained model costs only)


  2. Rapid deployment

    Fast turnaround time with no model training needed, allowing textual description design for faster iteration from definition to launch (Google Research, 2024-12-18).


    Deployment timeline:

    • Traditional ML: 3-6 months

    • Zero-shot: Hours to days


  3. Enhanced flexibility

    Models can effectively generalize to unseen classes, improving adaptability and enabling broader use cases across different domains (Lyzr AI, 2025-03-15).


    New categories added by simply defining them in natural language—no retraining required.


  4. Handles long-tail scenarios

    Perfect for rare categories where collecting training data is impractical:

    • Rare diseases (100-1,000 cases worldwide)

    • Emerging fraud patterns

    • New product types

    • Novel threats or risks


  5. Domain transfer

    Models trained on general data work across specialized domains with minimal adaptation.


Limitations

  1. 1. Accuracy gap versus fine-tuned models

    Zero-shot typically achieves 60-85% of fine-tuned model accuracy on same task.

    Example:

    • Fine-tuned sentiment classifier: 95% accuracy

    • Zero-shot sentiment classifier: 75-85% accuracy


  2. Bias toward seen classes

    Generalized zero-shot learning must overcome the tendency for classifiers to bias predictions towards classes seen in training over unseen classes not yet exposed to (IBM, 2025-07-31).


  3. Requires well-formulated prompts

    Zero-shot classification models perform well on generalized tasks but accuracy might be limited because there is no fine-tuning on specific tasks, requiring well-formulated prompts (IBM, 2025-04-17).


    Poor prompt: "good or bad"Better prompt: "Does this express customer satisfaction or dissatisfaction with product quality?"


  4. Struggles with abstract or complex tasks

    CLIP struggles on more abstract or systematic tasks such as counting number of objects and on complex tasks like predicting how close the nearest car is (OpenAI, 2021).


  5. Fine-grained classification challenges

    CLIP has poor performance on very fine-grained classification such as telling difference between car models, variants of aircraft, or flower species (OpenAI, 2021).


  6. Limited to model's pre-training knowledge

    Models cannot classify concepts completely outside their training distribution.


  7. Computational costs at inference

    Large models (GPT-3, GPT-4) require significant compute for each prediction:

    • API costs: $0.0001-$0.06 per 1,000 tokens

    • Latency: 100-2,000ms per request


Risk Matrix

Risk

Impact

Mitigation

Lower accuracy

Medium

Test on validation set; use few-shot if needed

Bias toward seen classes

High

Monitor prediction distribution; threshold tuning

Poor prompt design

Medium

A/B test multiple prompt formulations

Computational cost

Low-Medium

Cache common queries; batch processing

Domain shift

Medium

Evaluate on target domain; consider domain adaptation

Common Myths vs Facts


Myth 1: Zero-shot models don't need any training

Fact: Zero-shot models require extensive pre-training on large-scale general datasets before they can perform zero-shot classification on unseen classes (IBM, 2025-07-31).

Pre-training GPT-3 cost OpenAI an estimated $4-12 million in compute. The "zero" refers to zero task-specific training examples, not zero training overall.


Myth 2: Zero-shot is always more cost-effective

Fact: For high-volume production systems with stable categories, fine-tuned models often have lower total cost of ownership.


Break-even analysis:

  • If processing >10 million predictions/month

  • If categories remain stable >6 months

  • If accuracy improvement >10% with fine-tuning → Fine-tuning may be more cost-effective long-term


Myth 3: Zero-shot models can classify anything

Fact: CLIP still has poor generalization to images not covered in its pre-training dataset (OpenAI, 2021).


Models are constrained by their pre-training data. A model trained only on English text cannot classify Chinese documents. A model trained on natural images struggles with medical scans or satellite imagery.


Myth 4: Bigger models always perform better at zero-shot

Fact: While larger models generally improve zero-shot performance, RoBERTa classifiers consistently outperform GPT-3 zero-shot queries on domain-specific tasks (Bosley et al., 2023).


Model architecture, pre-training data quality, and alignment with downstream task matter as much as parameter count.


Myth 5: You can't improve zero-shot performance without labeled data

Fact: Multiple techniques improve zero-shot performance:

  • Prompt engineering and optimization

  • Calibration techniques

  • Ensemble methods

  • Using multiple candidate label formulations

  • Chain-of-thought prompting for complex tasks


Simple strategies like Multi-Null Prompt that concatenates multiple [MASK] tokens can outperform manual prompts in text classification tasks (ScienceDirect, 2023-06-05).


Myth 6: Zero-shot eliminates the need for human review

Fact: AI Mod enables human moderators to focus on complex gray area content while AI handles obvious violations (Microsoft Developer, 2024).

Best practice: Use zero-shot for initial triage and confidence scoring, escalate uncertain cases to humans.


Practical Pitfalls to Avoid


Pitfall 1: Unclear or Ambiguous Labels

Problem: "positive" and "negative" are context-dependent


Example:

  • Text: "The surgery was negative"

  • "negative" could mean:

    • Sentiment: Bad outcome

    • Medical: No disease detected (good outcome)


Solution: Use descriptive, unambiguous labels

  • "patient outcome was unfavorable"

  • "diagnostic test showed no disease"


Pitfall 2: Too Many Candidate Labels

Problem: Performance degrades with 20+ labels

Why: Model must compute similarity for each label, increasing noise

Solution:

  • Hierarchical classification (broad categories first, then narrow)

  • Group related labels

  • Pre-filter to likely categories


Example approach:

# Step 1: Broad classification
broad_result = classifier(text, ["medical", "legal", "financial"])

# Step 2: Narrow classification based on broad result
if broad_result['labels'][0] == "medical":
    narrow_result = classifier(text, [
        "symptoms", "diagnosis", "treatment", "prescription"
    ])

Pitfall 3: Not Testing on Representative Data

Problem: Model performs well on general examples but fails on real data

Solution:

  1. Create evaluation set from actual production data

  2. Test on edge cases and difficult examples

  3. Measure confusion matrix, not just overall accuracy

  4. Analyze failure modes before deployment


Pitfall 4: Ignoring Confidence Scores

Problem: Treating all predictions equally regardless of confidence

Solution: Implement confidence-based routing

result = classifier(text, labels)
top_score = result['scores'][0]

if top_score > 0.9:
    # High confidence - auto-process
    process_automatically(result)
elif top_score > 0.6:
    # Medium confidence - flag for review
    flag_for_human_review(result)
else:
    # Low confidence - immediate escalation
    escalate_to_expert(result)

Pitfall 5: Using Generic Pre-trained Models for Specialized Domains

Problem: General-purpose models struggle with technical jargon

Example: Medical text classification

Solutions:

  • Use domain-specific models (BioBERT, BlueBERT for medical)

  • Provide domain context in prompts

  • Consider domain-adaptive pre-training

  • Use few-shot examples from domain


Pitfall 6: Not Monitoring Prediction Distribution

Problem: Model defaults to most common class

Warning sign: One label represents >60% of predictions

Solution:

  • Monitor label distribution weekly

  • Compare to expected distribution

  • Investigate if bias detected

  • Adjust confidence thresholds per label


Pitfall 7: Overlooking Latency Requirements

Problem: API-based models too slow for real-time applications


Latency comparison:

  • Local BART model: 10-50ms

  • OpenAI API: 200-2,000ms


Solutions:

  • Deploy smaller models locally for real-time needs

  • Batch requests when possible

  • Cache frequent predictions

  • Use async processing for non-critical paths


Future Outlook 2025-2027


Emerging Trends

  1. 1. Multimodal zero-shot classification

    Models combining text, image, audio, and video for unified classification. Zero-shot classification capability opens new possibilities for innovation across industries from healthcare to finance and e-commerce (PingCAP, 2024-12-12).


    Prediction: By 2026, 40% of zero-shot deployments will use multimodal models versus 15% in 2024.


  2. Smaller, more efficient models

    Foundation models use transformer architecture enabling them to classify labels without specific training data through self-supervised learning and transfer learning (IBM, 2025-04-17).


    Trend: Distillation techniques creating models with 90% of performance at 10% of size:

    • Current: 400M-175B parameters

    • 2026 target: 100M-1B parameters with comparable accuracy

    • Benefit: Local deployment on edge devices


  3. Prompt optimization automation

    Manual prompt engineering being replaced by:

    • Automated prompt search algorithms

    • Reinforcement learning for prompt optimization

    • Meta-learning approaches


  4. Regulatory compliance features

    EU Digital Services Act requires companies to manage player complaints against content moderation decisions, disclose policies and tools, and detect illegal content with escalation to authorities (Microsoft Developer, 2024).


    Expected features by 2026:

    • Built-in explainability for predictions

    • Audit trails and provenance tracking

    • Bias detection and mitigation tools

    • Compliance reporting APIs


  5. Domain-specific zero-shot models


    Specialized models emerging:

    • LegalBERT for legal classification

    • FinBERT for financial documents

    • SciBERT for scientific literature

    • ClinicalBERT for healthcare


    Advantage: Better accuracy on domain-specific tasks while maintaining zero-shot flexibility.


Market Growth Projections

The global machine learning market including zero-shot capabilities:

  • 2024: $21 billion

  • 2027: $47 billion (projected)

  • CAGR: 30.4%


Source: Various industry analyst reports, 2024


Zero-shot specific growth drivers:

  1. Increasing cost of data labeling (up 15% annually)

  2. Demand for rapid AI deployment

  3. Regulatory pressure for explainable AI

  4. Expansion into emerging markets with limited labeled data


Research Directions

Active areas:

  1. Improved calibration - Better confidence estimates

  2. Compositional zero-shot - Combining attributes for novel concepts

  3. Continual zero-shot learning - Updating without catastrophic forgetting

  4. Cross-lingual zero-shot - Training on one language, deploying on 100+

  5. Zero-shot with retrieval - Combining zero-shot with knowledge bases


Challenges Ahead

  1. Scaling laws and diminishing returns

    As models grow larger, zero-shot improvements plateau. Research needed on efficient architectures.


  2. Ethical considerations

    Content moderation raises concerns about identifying types of harm accurately while minimizing false positives that could restrict legitimate speech (arXiv, 2025-01-23).


  3. Energy consumption

    Large model inference produces significant carbon footprint. Sustainable AI practices required.


  4. Copyright and training data

    Ongoing legal questions about using web-scraped data for pre-training.


FAQ


1. What is the difference between zero-shot and few-shot learning?

Zero-shot learning classifies data into categories without any training examples of those categories. Few-shot learning uses 1-100 labeled examples per category. Zero-shot learning is the extreme case of few-shot learning where K equals 0, meaning devoid of any visual examples of target classes in training phase (PMC7531283, 2020). Few-shot typically achieves 5-15% higher accuracy but requires some labeled data and additional training time.


2. Can zero-shot classification work with images?

Yes. CLIP enables zero-shot image classification through natural language prompts, achieving 64.3% accuracy on ImageNet by training on 400 million image-text pairs (OpenAI, 2021; Capa Learning, 2025). Models like CLIP bridge computer vision and natural language processing, allowing you to classify images using text descriptions of categories.


3. What accuracy can I expect from zero-shot classification?

Accuracy varies by task and domain. Text classification typically achieves 70-85% of fine-tuned model accuracy. Microsoft's production system reached over 98% accuracy on content moderation (Microsoft Developer, 2024), while GPT-3 achieved 76-83% on various NLP benchmarks in zero-shot mode (Springboard, 2023). Expect 60-90% accuracy depending on task complexity and label clarity.


4. What are the best models for zero-shot text classification?

facebook/bart-large-mnli is the most popular zero-shot classification model, trained on the MultiNLI dataset with 433,000 sentence pairs (Hugging Face, 2024). Other strong options include RoBERTa-large-mnli, DeBERTa-v3-large-mnli, and for production at scale, GPT-3.5 or GPT-4 via API. Choose based on accuracy needs, latency requirements, and budget.


5. Do I need to retrain the model for new categories?

No. That's the core advantage. Zero-shot classification allows predicting classes that weren't seen during model training by providing candidate labels in natural language (Hugging Face, 2024). Simply add new category labels as text descriptions. The model uses its pre-trained knowledge to classify into new categories immediately.


6. How do I improve zero-shot classification accuracy?

Six proven techniques: (1) Use more descriptive, specific labels instead of generic terms. (2) Provide context in prompts. (3) Use multi-label mode when appropriate. (4) Ensemble multiple models. (5) Implement simple strategies like Multi-Null Prompt that can outperform manual prompts (ScienceDirect, 2023). (6) Set confidence thresholds and escalate uncertain cases to human review.


7. What's the cost of using zero-shot classification?

Open-source models like BART are free to use but require compute infrastructure ($50-500/month for CPU/GPU instances). API-based models like GPT-3.5 cost $0.0005-$0.002 per 1,000 tokens (roughly $0.50-$2.00 per 1,000 classifications). Zero-shot approaches offer resource efficiency using the same workflow for multiple policies (Google Research, 2024), eliminating $10,000-$100,000 data labeling costs.


8. Can zero-shot classification handle multiple languages?

Yes, if the pre-training data included multiple languages. Models like XLM-RoBERTa and mBERT support 100+ languages. Zero-shot classification can identify language of given text, allowing multilingual applications to adapt dynamically (Spot Intelligence, 2024). Performance varies by language based on pre-training data representation.


9. Is zero-shot classification secure for sensitive data?

Consider these factors: (1) Open-source models (BART, RoBERTa) can run on-premise with full data control. (2) API services may log prompts—check provider privacy policies. (3) Deploy locally for HIPAA/GDPR compliance. (4) Use encryption for data in transit. (5) Audit model outputs for unintended information disclosure. Most enterprises prefer self-hosted models for sensitive data.


10. How does zero-shot classification handle bias?

Zero-shot models inherit biases from pre-training data. Models tend to bias predictions toward classes seen in training over unseen classes (IBM, 2025). Mitigation strategies: (1) Test on diverse demographic groups. (2) Monitor prediction distributions. (3) Use debiasing techniques during inference. (4) Regularly audit for fairness. (5) Combine with human review for high-stakes decisions. (6) Document known limitations transparently.


11. What's the difference between zero-shot learning and transfer learning?

Zero-shot learning is a subfield of transfer learning where the model extends knowledge from training instances to classify testing instances of completely different classes (V7 Labs). Transfer learning typically fine-tunes a pre-trained model on target task data. Zero-shot applies pre-trained knowledge directly without any task-specific training data. All zero-shot learning uses transfer learning, but not all transfer learning is zero-shot.


12. Can I use zero-shot for regression tasks?

Zero-shot primarily designed for classification. For regression (predicting continuous values), use few-shot learning with example values or fine-tune on target task. Some recent research explores zero-shot regression through prompt engineering ("rate this from 1-10"), but accuracy is limited. Classification tasks are more suitable for zero-shot approaches.


Key Takeaways

  • Zero-shot classification enables AI models to categorize data into classes they've never encountered during training, eliminating the need for labeled examples of every category


  • Models leverage semantic embeddings, transfer learning, and natural language inference to understand relationships between concepts and generalize to unseen classes


  • GPT-3 achieved 76-83% accuracy on various benchmarks in zero-shot mode, while production systems like Microsoft's content moderator reach 98% accuracy (Brown et al., 2020; Microsoft, 2024)


  • Real-world deployments include COVID-19 diagnosis from chest X-rays, Xbox content moderation processing millions of messages, and Google Ads policy enforcement


  • Primary advantage: Reduces data labeling costs from $10,000-$1,000,000 to near-zero and cuts deployment time from months to days


  • Best use cases: Rapid prototyping, frequently changing categories, rare/long-tail classifications, and scenarios where labeled data is unavailable or expensive


  • Accuracy trade-off: Zero-shot typically achieves 60-85% of fine-tuned model accuracy, requiring well-formulated prompts and confidence-based routing


  • Popular models: facebook/bart-large-mnli for text (400M parameters), CLIP for images (trained on 400M image-text pairs), GPT-3/4 for general-purpose classification


  • Implementation: Available through Hugging Face Transformers library, OpenAI API, and open-source models deployable on-premise for data security


  • Future outlook: Expected growth in multimodal models, smaller efficient architectures, and domain-specific variants with 30%+ annual market growth through 2027


Actionable Next Steps

  1. Evaluate your classification needs

    • List all categories you need to classify

    • Estimate how often categories change

    • Calculate current data labeling costs

    • Identify critical vs non-critical classification tasks


  2. Run a pilot test

    • Install Hugging Face Transformers: pip install transformers torch

    • Load facebook/bart-large-mnli model

    • Test on 100 real examples from your domain

    • Measure accuracy against ground truth

    • Document failure modes and edge cases


  3. Compare approaches

    • Benchmark zero-shot vs current method (if any)

    • Test few-shot with 10 examples per class

    • Calculate total cost: data labeling + compute + maintenance

    • Determine accuracy requirements for your use case


  4. Optimize prompts

    • Write 3-5 label formulations per category

    • A/B test different descriptions

    • Add domain context to prompts

    • Use multi-label mode if categories overlap


  5. Implement confidence-based routing

    • Set thresholds: >0.9 (auto-process), 0.6-0.9 (flag), <0.6 (escalate)

    • Route uncertain predictions to human review

    • Monitor prediction distribution weekly

    • Adjust thresholds based on precision/recall needs


  6. Deploy to production

    • Start with non-critical workflow

    • Load model once at server startup for efficiency

    • Implement caching for common queries

    • Log all predictions and confidence scores

    • Set up monitoring and alerting


  7. Measure and iterate

    • Track accuracy on production data weekly

    • Collect feedback on misclassifications

    • Refine label descriptions based on errors

    • Consider few-shot or fine-tuning if accuracy insufficient


Glossary

  1. Attention Mechanism - Neural network component that weighs importance of different input elements, allowing models to focus on relevant information when making predictions.


  2. Auxiliary Information - Additional data used in zero-shot learning such as text descriptions, attributes, or knowledge graphs that describe unseen classes without providing labeled examples.


  3. BART (Bidirectional and Auto-Regressive Transformers) - Sequence-to-sequence model combining BERT and GPT pretraining objectives, popular for zero-shot text classification when fine-tuned on NLI datasets.


  4. BERT (Bidirectional Encoder Representations from Transformers) - Transformer-based language model that uses bidirectional context to create word embeddings, enabling better language understanding.


  5. CLIP (Contrastive Language-Image Pre-training) - OpenAI model trained on 400 million image-text pairs enabling zero-shot image classification through natural language prompts.


  6. Embedding - Vector representation of data (text, images) in high-dimensional space where similar concepts cluster together, enabling semantic comparison.


  7. Few-Shot Learning - Machine learning approach using 1-100 labeled examples per class to achieve better accuracy than zero-shot while requiring minimal data.


  8. Fine-Tuning - Process of additional training on task-specific data to adapt a pre-trained model to particular use case, typically requiring thousands of labeled examples.


  9. Generalized Zero-Shot Learning (GZSL) - Variant where model classifies data that might belong to either seen or unseen classes, more challenging than standard zero-shot which only contains unseen classes at test time.


  10. GPT (Generative Pre-trained Transformer) - Family of large language models by OpenAI using decoder-only transformer architecture, capable of zero-shot and few-shot task performance.


  11. MultiNLI (Multi-Genre Natural Language Inference) - Dataset of 433,000 sentence pairs annotated for textual entailment, widely used to train zero-shot classification models like BART-large-mnli.


  12. Natural Language Inference (NLI) - Task of determining whether a hypothesis is true (entailment), false (contradiction), or undetermined (neutral) given a premise, foundational technique for zero-shot classification.


  13. Pre-training - Initial training phase where models learn from massive general-purpose datasets (billions of tokens, millions of images) before being applied to specific tasks.


  14. Prompt Engineering - Craft of designing input text to elicit desired behavior from language models, critical for zero-shot classification accuracy.


  15. Semantic Embeddings - Vector representations capturing meaning and relationships between concepts, enabling models to understand similarity between seen and unseen classes.


  16. Transfer Learning - Machine learning technique applying knowledge gained from one task to different but related tasks, foundational to all zero-shot approaches.


  17. Transformer - Neural network architecture using self-attention mechanisms to process sequential data, basis for modern language models like BERT, GPT, and BART.


  18. Zero-Shot Learning (ZSL) - Machine learning paradigm where models classify data into categories without seeing any training examples of those categories, relying on semantic knowledge and transfer learning.


Sources and References

  1. Brown, T., Mann, B., Ryder, N., et al. (2020). "Language Models are Few-Shot Learners." Advances in Neural Information Processing Systems, 33, 1877-1901. https://proceedings.neurips.cc/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html


  2. IBM Think Topics. (2025-07-31). "What Is Zero-Shot Learning?" IBM. https://www.ibm.com/think/topics/zero-shot-learning


  3. Hugging Face. (2024-01-30). "Zero-Shot Classification Task." Hugging Face. https://huggingface.co/tasks/zero-shot-classification


  4. Chae, Y., Davidson, T. (2025). "Large Language Models for Text Classification: From Zero-Shot Learning to Instruction-Tuning." SAGE Journals. https://journals.sagepub.com/doi/10.1177/00491241251325243


  5. V7 Labs. "What Is Zero Shot Learning in Image Classification?" V7 Labs Blog. https://www.v7labs.com/blog/zero-shot-learning-guide


  6. OpenAI. (2021). "CLIP: Connecting text and images." OpenAI Blog. https://openai.com/index/clip/


  7. Redfield AI. (2022-08-27). "Zero Shot Learning - Complete Guide 2025." Redfield. https://redfield.ai/zero-shot-learning/


  8. GeeksforGeeks. (2025-07-23). "Zero Shot Classification." GeeksforGeeks Machine Learning. https://www.geeksforgeeks.org/machine-learning/zero-shot-classification/


  9. PingCAP. (2024-12-12). "How Zero-Shot Classification Enhances AI Models." PingCAP Blog. https://www.pingcap.com/article/how-zero-shot-classification-enhances-ai-models/


  10. Spot Intelligence. (2024-10-07). "Zero-Shot Classification: Top 6 Models, How To Tutorial." Spot Intelligence. https://spotintelligence.com/2023/08/01/zero-shot-classification/


  11. Yang, X., et al. (2025-01-23). "Towards Safer Social Media Platforms: Scalable and Performant Few-Shot Harmful Content Moderation Using Large Language Models." arXiv. https://arxiv.org/html/2501.13976v1


  12. Microsoft Game Developer. (2024-03). "GDC 2024: Community Sift & the Future of Content Moderation." Microsoft Developer. https://developer.microsoft.com/en-us/games/articles/2024/03/community-sift-and-the-future-of-content-moderation/


  13. Google Research. (2024-12-18). "Zero-Shot Image Moderation in Google Ads with LLM-Assisted Textual Descriptions and Cross-modal Co-embeddings." arXiv:2412.16215. https://arxiv.org/html/2412.16215


  14. XCube Labs. (2024-09-10). "Exploring Zero-Shot and Few-Shot Learning in Generative AI." XCube Labs Blog. https://www.xcubelabs.com/blog/exploring-zero-shot-and-few-shot-learning-in-generative-ai/


  15. PMC (PubMed Central). (2020-10-02). "Zero-shot learning and its applications from autonomous vehicles to COVID-19 diagnosis: A review." PMC7531283. https://pmc.ncbi.nlm.nih.gov/articles/PMC7531283/


  16. IEEE Transactions on Medical Imaging. (2025-01). "Multi-Label Generalized Zero Shot Chest X-Ray Classification by Combining Image-Text Information With Feature Disentanglement." PubMed:39018216. https://pubmed.ncbi.nlm.nih.gov/39018216/


  17. Cao, W., Yao, X., Xu, Z., et al. (2025-04-04). "A Survey of Zero-Shot Object Detection." Big Data Mining and Analytics, 8(3), 726-750. https://www.sciopen.com/article/10.26599/BDMA.2024.9020098


  18. IBM Think Tutorials. (2025-04-17). "Zero-shot Classification Tutorial with Granite." IBM. https://www.ibm.com/think/tutorials/zero-shot-classification


  19. Springboard. (2023-10-06). "OpenAI GPT-3: Everything You Need to Know." Springboard Blog. https://www.springboard.com/blog/data-science/machine-learning-gpt-3-open-ai/


  20. Bosley, M., Harukawa, T., Licht, A., Hoyle, R. (2023). "Do we still need BERT in the age of GPT? Comparing Transformers for Political Text Classification." MPSA 2023. https://mbosley.github.io/papers/bosley_harukawa_licht_hoyle_mpsa2023.pdf


  21. ScienceDirect. (2023-06-05). "Are the BERT family zero-shot learners? A study on their potential and limitations." Artificial Intelligence Journal. https://www.sciencedirect.com/science/article/abs/pii/S0004370223000991


  22. Medium / Neural Engineer. (2024-10-11). "Moderation Classifier: LlamaGuard Zero-Shot Learning." Medium. https://medium.com/neural-engineer/moderation-classifier-llamaguard-zero-shot-learning-dddabbbcb00c


  23. Medium / Sganesh. (2023-08-14). "Content moderation to Zero Shot classification." Medium. https://medium.com/@sganesh.7/content-moderation-to-zero-shot-classification-295805008e83


  24. Roboflow. (2024-09-10). "Zero-Shot Content Moderation with OpenAI's New CLIP Model." Roboflow Blog. https://blog.roboflow.com/zero-shot-content-moderation-openai-new-clip-model/


  25. Lyzr AI. (2025-03-15). "What Is Zero-Shot Learning." Lyzr AI Glossary. https://www.lyzr.ai/glossaries/zero-shot-learning/


  26. Papers with Code. "Zero-Shot Learning." Papers with Code. https://paperswithcode.com/task/zero-shot-learning/codeless


  27. Capa Learning. (2025-03-02). "N-Shot Learning: Zero vs. Single vs. Two vs. Few (2025)." Capa Learning. https://capalearning.com/2025/03/02/n-shot-learning-zero-vs-single-vs-two-vs-few-2025/


  28. Hugging Face Documentation. "BART Model Documentation." Hugging Face Transformers. https://huggingface.co/docs/transformers/en/model_doc/bart


  29. GeeksforGeeks. (2025-07-23). "Zero-Shot Text Classification using HuggingFace Model." GeeksforGeeks. https://www.geeksforgeeks.org/nlp/zero-shot-text-classification-using-huggingface-model/


  30. Wikipedia. (2025). "GPT-3." Wikipedia. https://en.wikipedia.org/wiki/GPT-3




$50

Product Title

Product Details goes here with the simple product description and more information can be seen by clicking the see more button. Product Details goes here with the simple product description and more information can be seen by clicking the see more button

$50

Product Title

Product Details goes here with the simple product description and more information can be seen by clicking the see more button. Product Details goes here with the simple product description and more information can be seen by clicking the see more button.

$50

Product Title

Product Details goes here with the simple product description and more information can be seen by clicking the see more button. Product Details goes here with the simple product description and more information can be seen by clicking the see more button.

Recommended Products For This Post
 
 
 

Comments


bottom of page