Recent progress towards designing models that can generalize to unseen domains (i.e domain generalization) or unseen classes (i.e zero-shot learning) has embarked interest towards building models that can tackle both domain-shift and semantic shift simultaneously (i.e zero-shot domain generalization). For models to generalize to unseen classes in unseen domains, it is crucial to learn feature representation that preserves class-level (domain-invariant) as well as domain-specific information. Motivated from the success of generative zero-shot approaches, we propose a feature generative framework integrated with a COntext COnditional Adaptive (COCOA) Batch-Normalization to seamlessly integrate class-level semantic and domain-specific information. The generated visual features better capture the underlying data distribution enabling us to generalize to unseen classes and domains at test-time. We thoroughly evaluate and analyse our approach on established large-scale benchmark - DomainNet and demonstrate promising performance over baselines and state-of-art methods.
Publications
2022
COCOA: Context-Conditional Adaptation for Recognizing Unseen Classes in Unseen Domains [Abs] [Poster] [PPT]
Puneet Mangla, Shivam Chandhok, Vineeth N Balasubramanian, Fahad Shahbaz Khan
Published in Proceedings of IEEE Winter Conference on Applications of Computer Vision (WACV) 2022
(Also presented at IJCAI 2021 Weakly Supervised Representation Learning Workshop)
Data InStance Prior (DISP) in Generative Adversarial Networks [Abs] [Poster] [PPT]
Puneet Mangla, Nupur Kumari, Mayank Singh, Vineeth N Balasubramanian, Balaji Krishnamurthy
Published in Proceedings of IEEE Winter Conference on Applications of Computer Vision (WACV) 2022
We propose a novel transfer learning method for GANs in the limited data domain by leveraging informative data prior derived from self-supervised/supervised pre-trained networks trained on a diverse source domain. We perform experiments on several standard vision datasets using various GAN architectures (BigGAN, SNGAN, StyleGAN2) to demonstrate that the proposed method effectively transfers knowledge to domains with few target images, outperforming existing state-of-the-art techniques in terms of image quality and diversity. We also show the utility of data instance prior in large-scale unconditional image generation and image editing tasks.
2021
On the Benefits of Defining Vicinal Distributions in Latent Space [Abs] [Poster] [PPT]
Puneet Mangla, Vedant Singh, Shreyas Jayant Havaldar, Vineeth N Balasubramanian
Extended version accepted at Elsevier Journal - Pattern Recognition Letters (Impact Factor: 3.756)
Best Paper Award at Workshop on Adversarial Machine Learning, CVPR 2021
(Also accepted at RobustML Workshop, ICLR 2021 (Oral) and Generalization Workshop, ICLR 2021)
In this work, we investigate the benefits of defining these vicinal distributions like mixup in latent space of generative models rather than in input space itself. We propose a new approach - VarMixup (Variational Mixup) - to better sample mixup images by using the latent manifold underlying the data. Our empirical studies on CIFAR-10, CIFAR-100, and Tiny-ImageNet demonstrate that models trained by performing mixup in the latent manifold learned by VAEs are inherently more robust to various input corruptions/perturbations, are significantly better calibrated, and exhibit more local-linear loss landscapes.
2020
Attributional Robustness Training using Input-Gradient Spatial Alignment [Abs] [PPT]
Mayank Singh, Nupur Kumari, Puneet Mangla, Abhishek Sinha, Vineeth N Balasubramanian, Balaji Krishnamurthy
Published in Proceedings of European Conference on Computer Vision (ECCV) 2020
In this work, we study the problem of attributional robustness (ie models having robust explanations) by showing an upper bound for attributional vulnerability in terms of spatial correlation between the input image and its explanation map. We propose a training methodology that learns robust features by minimizing this upper bound using soft-margin triplet loss. Our methodology of robust attribution training (\textit {ART}) achieves the new state-of-the-art attributional robustness measure by a margin of 6-18 on several standard datasets, ie. SVHN, CIFAR-10 and GTSRB. We further show the utility of the proposed robust training technique (\textit {ART}) in the downstream task of weakly supervised object localization by achieving the new state-of-the-art performance on CUB-200 dataset.
On Saliency Maps and Adversarial Robustness [Abs] [PPT]
Puneet Mangla, Vedant Singh, Vineeth N Balasubramanian
Published in Proceedings of European Conference of Machine Learning and Principles and Practice of Knowledge Discovery in Databases (ECML-PKDD) 2020
In this work, we provide a different perspective to this coupling, and provide a method, Saliency based Adversarial training (SAT), to use saliency maps to improve adversarial robustness of a model. In particular, we show that using annotations such as bounding boxes and segmentation masks, already provided with a dataset, as weak saliency maps, suffices to improve adversarial robustness with no additional effort to generate the perturbations themselves. Our empirical results on CIFAR-10, CIFAR-100, Tiny ImageNet and Flower-17 datasets consistently corroborate our claim, by showing improved adversarial robustness using our method. saliency maps. We also show how using finer and stronger saliency maps leads to more robust models, and how integrating SAT with existing adversarial training methods, further boosts performance of these existing methods.
Charting the Right Manifold: Manifold Mixup for Few-shot Learning [Abs] [Poster]
Puneet Mangla, Mayank Singh, Abhishek Sinha, Nupur Kumari, Vineeth N Balasubramanian, Balaji Krishnamurthy
Published in Proceedings of IEEE Winter Conference on Applications of Computer Vision (WACV) 2020
(Also accepted as spotlight at NeurIPS 2020 - MetaLearn Workshop)
This work investigates the role of learning relevant feature manifold for few-shot tasks using self-supervision and regularization techniques. We observe that regularizing the feature manifold, enriched via self-supervised techniques, with Manifold Mixup significantly improves few-shot learning performance. We show that our proposed method S2M2 beats the current state-of-the-art accuracy on standard few-shot learning datasets like CIFAR-FS, CUB, mini-ImageNet and tiered-ImageNet by 3-8%. Through extensive experimentation, we show that the features learned using our approach generalize to complex few-shot evaluation tasks, cross-domain scenarios and are robust against slight changes to data distribution.
2019
AdvGAN++ : Harnessing latent layers for adversary generation [Abs]
Puneet Mangla, Surgan Jandial, Sakshi Varshney, Vineeth N Balasubramanian
Published in Proceedings of Neural Architects Workshop, ICCV 2019
In this work, we show how latent features can serve as better priors than input images for adversary generation by proposing AdvGAN++, a version of AdvGAN that achieves higher attack rates than AdvGAN and at the same time generates perceptually realistic images on MNIST and CIFAR-10 datasets.