Feature Generating Networks for Zero-Shot Learning

Authors
Publication date 2018
Book title 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition
Book subtitle proceedings : 18-22 June 2018, Salt Lake City, Utah
ISBN
  • 9781538664216
ISBN (electronic)
  • 9781538664209
Event 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition
Pages (from-to) 5542-5551
Publisher Los Alamitos, California: IEEE Computer Society
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Suffering from the extreme training data imbalance between seen and unseen classes, most of existing state-of-the-art approaches fail to achieve satisfactory results for the challenging generalized zero-shot learning task. To circumvent the need for labeled examples of unseen classes, we propose a novel generative adversarial network (GAN) that synthesizes CNN features conditioned on class-level semantic information, offering a shortcut directly from a semantic descriptor of a class to a class-conditional feature distribution. Our proposed approach, pairing a Wasserstein GAN with a classification loss, is able to generate sufficiently discriminative CNN features to train softmax classifiers or any multimodal embedding method. Our experimental results demonstrate a significant boost in accuracy over the state of the art on five challenging datasets – CUB, FLO, SUN, AWA and ImageNet – in both the zero-shot learning and generalized zero-shot learning settings.
Document type Conference contribution
Language English
Published at https://doi.org/10.1109/CVPR.2018.00581
Permalink to this page
Back