TY - JOUR
T1 - Dual prototype contrastive network for generalized zero-shot learning
AU - Jiang, Huajie
AU - Li, Zhengxian
AU - Hu, Yongli
AU - Yin, Baocai
AU - Yang, Jian
AU - Hengel, Anton van den
AU - Yang, Ming-Hsuan
AU - Qi, Yuankai
PY - 2025/2
Y1 - 2025/2
N2 - Generalized zero-shot learning (GZSL) requires that models are able to recognize classes they were trained on, and new classes they haven’t seen before. Feature-generation approaches are popular due to their effectiveness in mitigating overfitting to the training classes. Existing generative approaches usually adopt simple discriminators for distribution or classification supervision, however, thus limiting their ability to generate visual features that are discriminative of and transferable to novel categories. To overcome this limitation and improve the quality of generated features, we propose a dual prototype contrastive augmented discriminator for the generative adversarial network. Specifically, we design a Dual Prototype Contrastive Network (DPCN), which leverages complementary information between visual space and semantic space through multi-task prototype contrastive learning. Contrastive learning of the visual prototypes enhances the ability of the generated features to distinguish between classes, while the contrastive learning of the semantic prototypes improves their transferability. Furthermore, we introduce margins into the contrastive learning process to ensure both intra-class compactness and inter-class separation. To demonstrate the effectiveness of the proposed approach, we conduct experiments on three widely-used zero-shot learning benchmark datasets, where DPCN achieves state-of-the-art performance for GZSL.
AB - Generalized zero-shot learning (GZSL) requires that models are able to recognize classes they were trained on, and new classes they haven’t seen before. Feature-generation approaches are popular due to their effectiveness in mitigating overfitting to the training classes. Existing generative approaches usually adopt simple discriminators for distribution or classification supervision, however, thus limiting their ability to generate visual features that are discriminative of and transferable to novel categories. To overcome this limitation and improve the quality of generated features, we propose a dual prototype contrastive augmented discriminator for the generative adversarial network. Specifically, we design a Dual Prototype Contrastive Network (DPCN), which leverages complementary information between visual space and semantic space through multi-task prototype contrastive learning. Contrastive learning of the visual prototypes enhances the ability of the generated features to distinguish between classes, while the contrastive learning of the semantic prototypes improves their transferability. Furthermore, we introduce margins into the contrastive learning process to ensure both intra-class compactness and inter-class separation. To demonstrate the effectiveness of the proposed approach, we conduct experiments on three widely-used zero-shot learning benchmark datasets, where DPCN achieves state-of-the-art performance for GZSL.
UR - http://www.scopus.com/inward/record.url?scp=85207139108&partnerID=8YFLogxK
U2 - 10.1109/TCSVT.2024.3474910
DO - 10.1109/TCSVT.2024.3474910
M3 - Article
SN - 1051-8215
VL - 35
SP - 1111
EP - 1122
JO - IEEE Transactions on Circuits and Systems for Video Technology
JF - IEEE Transactions on Circuits and Systems for Video Technology
IS - 2
ER -