IIIT-Delhi Institutional Repository

Deep learning for attribute prediction and small sample size problems

Show simple item record

dc.contributor.author Sethi, Akshay
dc.contributor.author Vatsa, Mayank (Advisor)
dc.contributor.author Singh, Richa (Advisor)
dc.date.accessioned 2018-09-24T09:46:51Z
dc.date.available 2018-09-24T09:46:51Z
dc.date.issued 2018-04-17
dc.identifier.uri http://repository.iiitd.edu.in/xmlui/handle/123456789/659
dc.description.abstract The first part of this work focuses on facial attribute prediction using a novel deep learning formulation, termed as R-Codean autoencoder. The work presents Cosine similarity based loss function in an autoencoder which is then incorporated into the Euclidean distance based autoencoder to formulate R-Codean. The proposed loss function thus aims to incorporate both magnitude and direction of image vectors during feature learning. Inspired by the utility of shortcut connections in deep models to facilitate learning of optimal parameters, without incurring the problem of vanishing gradient, the proposed formulation is extended to incorporate shortcut connections in the architecture. The proposed R-Codean autoencoder is utilized in facial attribute prediction framework which incorporates patch-based weighting mechanism for assigning higher weights to relevant patches for each attribute. The experimental results on publicly available CelebA and LFWA datasets demonstrate the efficacy of the proposed approach in addressing this challenging problem.The second part of this work focuses on understanding the limits of data augmentation by synthesized images by Generative Adversarial Networks. Using extensive experiments, we investigate the question of 'how much augmentation is good augmentation', in terms of the ultimate aim of classification. Experiments with multiple GAN architectures and classi_ers across datasets further substantiate our findings.In the final part,we present new architectural and training procedures for Generative Adversarial Networks(GAN)[16] .We first determine whether the probability distribution of data generated by GANs is equal(or close enough) to the probability distribution of the dataset. For this purpose we train Deep Convolutional Network Classifiers[32, 37] on the generated images and compare them with those trained on the dataset. We show that the probability distributions in both the cases differ.We conduct this analysis on the MNIST[36], CIFAR10[31] and Adience[12] datasets.Having shown the above, we suggest changes in the current GAN training algorithm using the subclass information present in datasets. We show improved results using the CIFAR100[31] and Adience(each of these datasets have the subclass information present) and follow the standard protocol defined on these datasets for training the GAN en_US
dc.language.iso en_US en_US
dc.publisher IIIT-Delhi en_US
dc.subject Autoencoders en_US
dc.subject Skip Connections en_US
dc.subject Facial Attributes en_US
dc.subject Generative Adversarial Networks en_US
dc.subject Convolutional Neural Network en_US
dc.subject Deep Neural Network en_US
dc.subject Sub-Class en_US
dc.subject Data Augmentation en_US
dc.subject Generative Modelling en_US
dc.title Deep learning for attribute prediction and small sample size problems en_US
dc.type Other en_US


Files in this item

This item appears in the following Collection(s)

Show simple item record

Search Repository


Advanced Search

Browse

My Account