+ All Categories
Home > Documents > Zero Shot Recognition Using GCN

Zero Shot Recognition Using GCN

Date post: 02-Jun-2022
Category:
Upload: others
View: 5 times
Download: 0 times
Share this document with a friend
16
Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details M E0-270 : Machine Learning Indian Institute of Science, Bangalore. Zero Shot Recognition Using GCN Ameenudeen PE, Danish Shaikh, Deepa TM, Najath Pathiyil April 27, 2019 1 / 16
Transcript
Page 1: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

E0-270 : Machine LearningIndian Institute of Science,Bangalore.

Zero Shot Recognition Using GCN

Ameenudeen PE, Danish Shaikh,Deepa TM, Najath Pathiyil

April 27, 2019

1 / 16

Page 2: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Overview

1 Introduction

2 Problem Statement

3 ConSE

4 GCN for Zero-shot Learning

5 Approach

6 Training and Testing

7 Implementation Details

8 Modified approach: Dense Graph Propagation (DGP)

9 Results and Discussions

2 / 16

Page 3: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Introduction

Zero-Shot Learning: A method to predict the labels of unseenexamples without the burden of collecting its training samples.

Lets elaborate in more detail, consider the below image.

Okapi : ”zebra-striped four-legged animal with a brown torsoand a deer-like face.”

3 / 16

Page 4: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Problem Statement

Motivation: Can we build a classifier given the abovementioned constraints?Solution: Transfer knowledge that is obtained from seenclasses to describe the unseen classes.

1 Learn a vector representation of different categories using textdata and then learn a mapping between the vectorrepresentation to visual classifier directly.

2 Use explicit knowledge bases or knowledge graphs.

Our Method

Combine both word embeddings and knowledge graphs.

Each node corresponds to a class and relationships among them areshown via edges.

Word embeddings of each category is given as the node input.

Use GCN to transfer information between the layers.

4 / 16

Page 5: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

ConSE

Convex Combination of Semantic Embeddings (ConSE) usesMethod 1 approach.

We estimate the conditional probability distribution of eachtraining class given the samples.

For test samples, we compute a weighted combination of thelabel embeddings of top ’T’ training classes in the semanticspace where, T is hyperparameter.

Using cosine similarity to calculate the nearest class in thesemantic space.

5 / 16

Page 6: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

GCN for Zero-shot Learning

C denotes the set of all classes

Cte : Testing classes.Ctr : Training classes.

Given a graph G = (V, E), it takes the feature matrix andAdjacency matrix of the graph as the input.

Cte ∩ Ctr = φ

Training data points:Dtr = (Xi , ci ), i = 1, ...N. ci ∈ Ctr is theclass-label.

We give 300-D semantic representation vector of all classes tothe graph.

Using GCN, we predict the class labels of Cte in a semisupervised manner.

6 / 16

Page 7: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Approach

Given: A graph with N nodes and S input features per node,X ∈ RNXS denotes the feature matrix.

Each node represents one distinct class.

Adjacency matrix A ∈ RNXN , shows connections between theclasses in the knowledge graph.

Propagation rule : H i+1 = f (D−1AH iW i )

H i : Activations in the i th layerW i : Trainable weight matrix for layer i.D: Degree matrix.

H0 = X

The activation function f used is Leaky Relu.

7 / 16

Page 8: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Figure: Block diagram of the approach

8 / 16

Page 9: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Training and Testing

Training :

We use the mean square error between the predicted andground truth weights as loss function and use this to estimatethe parameters of GCN.

L =1

2M

M∑i=1

P∑j=1

(Wij − Wij)2

,M: no. of training classes, W : predicted weights, P:dimensionality of weights.Using these parameters, classifier weights for zero shotcategories are estimated.

Testing :

Extract the features of testing images using pre-trained CNN.

Cosine similarity of those features with generated classifiers.9 / 16

Page 10: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Implementation Details

Training : Imagenet 2012 1K dataset

Testing.1 300 classes from ”2-hops”.2 100 random classes.3 AWA dataset consisting of 50 classes.

Knowledge graph: sub-graph of the WordNet.

Feature extraction using ResNet-50.

Word embedding by GloVe text model.

Average of class attributes of the word embeddings.

6 Convolution layers.

Output feature dimension: 2048

During inference phase, cosine-similarity is used betweenlearned GCN weights and test set features from CNN.

Softmax score.

10 / 16

Page 11: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Modified approach: Dense Graph Propagation (DGP)

Normal graph: Dilution of knowledge because of heavysmoothing in each layer.Solution: Dense Graph and convert it into one layer.

Figure: GP vs. DGP for Node ’Cat’

DGP for zero-shot learning aims to use a two phasehierarchical graph structure: namely descendant propagationand ancestor propagation.

Figure: GP vs. DGP for Node ’Cat’

11 / 16

Page 12: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Results and Discussions

Figure: Top K accuracy for different models in different settings.

Figure: Test Accuracy with AWA dataset.

12 / 16

Page 13: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Conclusions

Our work shows that a knowledge graph provides supervisionto learn meaningful classifiers on top of semantic embeddings.

We also compared our results with current state-of-the-artConSE and got significant improvements.

A modified approach using a DGP module is alsoimplemented and results were improved significantly.

We also observed that the DGP model has overcome theproblem of knowledge dilution.

13 / 16

Page 14: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Future Scope

There are many hyperparameters that we did not tune due tothe lack of time, like, number of layers in GCN, the CNNmodel to extract features, etc.

We have given input as averaging of the class-attributes. Itcan be replaced with another weighted graph that givesrelation of these attributes.

We can use weighted edge between classes in the graphinstead of direct connection.

Instead of pre-trained word-embedding space, one can usesome other space to get the relation between test and trainingsamples.

14 / 16

Page 15: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

References

[1] Yongqin Xian, Christoph H. Lampert, Bernt Schiele,Zeynep Akata: ”Zero-Shot Learning - A ComprehensiveEvaluation of the Good, the Bad and the Ugly”. CoRRabs/1707.00600 (2017).

[2] Xiaolong Wang, Yufei Ye, Abhinav Gupta: ”Zero-ShotRecognition via Semantic Embeddings and KnowledgeGraphs”. CVPR 2018: 6857-6866.

[3] Michael Kampffmeyer, Yinbo Chen et al.: ”RethinkingKnowledge Graph Propagation for Zero-Shot Learning” toappear in CVPR 2019.

15 / 16

Page 16: Zero Shot Recognition Using GCN

Introduction Problem Statement ConSE GCN for Zero-shot Learning Approach Training and Testing Implementation Details Modified approach: Dense Graph Propagation (DGP) Results and Discussions

Thank You!

16 / 16


Recommended