🚀 Try Zilliz Cloud, the fully managed Milvus, for free—experience 10x faster performance! Try Now>>

Milvus
Zilliz

How does zero-shot learning handle unseen classes?

Zero-shot learning (ZSL) enables machine learning models to recognize classes they haven’t encountered during training by leveraging semantic or auxiliary information about those classes. Instead of relying solely on labeled examples, ZSL uses shared attributes, textual descriptions, or relationships between known and unknown classes to generalize to new categories. For example, a model trained to recognize animals like lions and tigers might infer the existence of a “zebra” by combining attributes like “stripes,” “four legs,” and “savannah habitat,” even if it never saw a zebra in training data. This approach bypasses the need for retraining on new data, making it efficient for scenarios where labeled examples for unseen classes are scarce or unavailable.

The core mechanism involves mapping input data (e.g., images or text) to a semantic space that encodes relationships between classes. During training, the model learns to associate features of inputs (like pixel patterns in images) with semantic attributes or embeddings (e.g., word vectors from descriptions). For instance, a vision model might use a pre-trained ResNet to extract image features, which are then projected into a space aligned with Word2Vec embeddings of class names. At inference time, the model compares the projected features of an unseen input to the embeddings of all possible classes—including unseen ones—to predict the most likely match. A practical example is classifying a “horse” versus a "zebra": if the model knows “stripes” are a key attribute of zebras, it can distinguish them even without zebra-specific training images.

ZSL faces challenges like domain shift (differences between training and test data distributions) and the “hubness” problem, where some embeddings dominate predictions. To address these, methods like generative models (e.g., GANs) synthesize features for unseen classes based on their semantic descriptions, improving alignment. Other techniques use graph-based relationships to model class hierarchies. For example, in the Animals with Attributes (AwA) dataset, models infer unseen animal classes by leveraging shared attributes (e.g., “has hooves”) between known and unknown species. While ZSL isn’t perfect, it’s particularly useful in domains like rare species identification, niche product categorization, or handling emerging concepts in NLP tasks, where labeling new data is impractical. Developers can implement ZSL using frameworks like PyTorch or TensorFlow, combining pre-trained feature extractors with semantic embeddings for efficient adaptation to new classes.

Like the article? Spread the word