+ All Categories
Home > Documents > TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire...

TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire...

Date post: 04-Aug-2020
Category:
Upload: others
View: 0 times
Download: 0 times
Share this document with a friend
22
TRECVID-2015 Semantic Indexing task: Overview Georges Quénot Laboratoire d'Informatique de Grenoble George Awad Dakota Consulting - NIST
Transcript
Page 1: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

TRECVID-2015 Semantic Indexing task:

Overview

Georges QuénotLaboratoire d'Informatique de Grenoble

George AwadDakota Consulting - NIST

Page 2: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Outline

•Task summary (Goals, Data, Run types, Concepts, Metrics)

•Evaluation details•Inferred average precision•Participants

•Evaluation results•Hits per concept•Results per run•Results per concept•Significance tests

•Progress task results•Global Observations

2

Page 3: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Semantic Indexing task

•Goal: Automatic assignment of semantic tags to video segments (shots)

•Secondary goals: •Encourage generic (scalable) methods for detector development.•Semantic annotation is important for filtering, categorization, searching and browsing.

•Task: Find shots that contain a certain concept, rank them according to confidence measure, submit the top 2000.

•Participants submitted one type of runs: •Main run Includes results for 60 concepts, from which NIST evaluated 30.

3

Page 4: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Semantic Indexing task (data)

•SIN testing dataset • Main test set (IACC.2.C): 200 hours, with durations between

10 seconds and 6 minutes.

•SIN development dataset• (IACC.1.A, IACC.1.B, IACC.1.C & IACC.1.tv10.training): 800 hours, used from 2010 – 2012 with durations between 10 seconds to just longer than 3.5 minutes.

•Total shots: •Development: 549,434•Test: IACC.2.C (113,046 shots)

• Common annotation for 346 concepts coordinated by LIG/LIF/Quaerofrom 2007-2013 made available.

4

Page 5: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Semantic Indexing task (Concepts)

• Selection of the 60 target concepts Were drawn from 500concepts chosen from the TRECVID “high level features” from 2005 to 2010 to favor cross-collection experiments Plus a selection of LSCOM concepts.

• Generic-Specific relations among concepts for promoting research on methods for indexing many concepts and using ontology relations between them.

• we cover a number of potential subtasks, e.g. “persons” or “actions” (not really formalized).

• These concepts are expected to be useful for the content-based (instance) search task.

•Set of relations provided:•427 “implies” relations, e.g. “Actor implies Person”

•559 “excludes” relations, e.g. “Daytime_Outdoor excludes Nighttime”

5

Page 6: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Semantic Indexing task (training types)

•Six training types were allowed:•A – used only IACC training data (30 runs)

•B – used only non-IACC training data (0 runs)

•C – used both IACC and non-IACC TRECVID (S&V and/or

Broadcast news) training data (2 runs)

•D – used both IACC and non-IACC non-TRECVID training

data(54 runs)

•E – used only training data collected automatically using only the

concepts’ name and definition (0 runs)

•F – used only training data collected automatically using a query

built manually from the concepts’ name and definition (0 runs)

6

Page 7: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

30 Single concepts evaluated(1)

3 Airplane*5 Anchorperson9 Basketball*13 Bicycling*15 Boat_Ship*17 Bridges*19 Bus*22 Car_Racing27 Cheering*31 Computers*38 Dancing41 Demonstration_Or_Protest49 Explosion_fire56 Government_leaders71 Instrumental_Musician*

-The 14 marked with “*” are a subset of those tested in 2014

8

72 Kitchen80 Motorcycle*85 Office86 Old_people95 Press_conference100 Running*117 Telephones*120 Throwing261 Flags*297 Hill321 Lakes392 Quadruped*440 Soldiers454 Studio_With_Anchorperson478 Traffic

Page 8: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Evaluation

•The 30 evaluated single concepts were chosen after examining TRECVid 2013 60 evaluated concept scores across all runs and choosing the top 45 concepts with maximum score variation.

•Each feature assumed to be binary: absent or present for each master reference shot

•NIST sampled ranked pools and judged top results from all submissions

•Metrics: inferred average precision per concept

•Compared runs in terms of mean inferred average precision across the 30 concept results for main runs.

9

Page 9: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

2015: mean extended Inferred average precision (xinfAP)

• 2 pools were created for each concept and sampled as:•Top pool (ranks 1-200) sampled at 100%

•Bottom pool (ranks 201-2000) sampled at 11.1%

•Judgment process: one assessor per concept, watched

complete shot while listening to the audio.

•infAP was calculated using the judged and unjudged pool by

sample_eval

30 concepts

195,500 total judgments

11,636 total hits

7489 Hits at ranks (1-100)

2970 Hits at ranks (101-200)

1177 Hits at ranks (201-2000)

11

Page 10: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

2015 : 15 Finishers

PicSOM Aalto U., U. of Helsinki

ITI_CERTH Information Technologies Institute, Centre for Research and

Technology Hellas

CMU Carnegie Mellon U.; CMU-Affiliates

Insightdcu Dublin City Un.; U. Polytechnica Barcelona

EURECOM EURECOM

FIU_UM Florida International U., U. of Miami

IRIM CEA-LIST, ETIS, EURECOM, INRIA-TEXMEX, LABRI, LIF, LIG, LIMSI-

TLP, LIP6, LIRIS, LISTIC

LIG Laboratoire d'Informatique de Grenoble

NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

TokyoTech Tokyo Institute of Technology

MediaMill U. of Amsterdam Qualcomm

siegen_kobe_nict U. of Siegen; Kobe U.; Natl. Inst. of Info. and Comm. Tech.

UCF_CRCV U. of Central Florida

UEC U. of Electro-Communications

Waseda Waseda U.

12

Page 11: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Inferred frequency of hits varies by concept

0

500

1000

1500

2000

2500

3000

3500

Airp

lane

An

cho

rpe

rso

n

Ba

sketb

all

Bic

yclin

g

Bo

at_

Sh

ip

Brid

ge

s

Bu

s

Ca

r_R

acin

g

Ch

ee

ring

Co

mp

ute

rs

Da

ncin

g

De

mo

nstra

tion

_O

r_P

rote

st

Exp

losio

n_

fire

Go

vern

me

nt_

lead

ers

Instru

me

nta

l_M

usic

ian

Kitc

hen

Mo

torc

ycle

Offic

e

Old

_p

eo

ple

Pre

ss_

con

fere

nce

Ru

nn

ing

Tele

ph

on

es

Thro

win

g

Fla

gs

Hill

Lake

s

Qu

ad

rup

ed

So

ldie

rs

Stu

dio

_W

ith_A

ncho

rpe

rso

n

Tra

ffic

Inf. Hits

1%**

**from total test shots

13

Page 12: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Total true shots contributed uniquely by team

Team No. of

Shots

Team No. of

shots

Insightdcu 27 Mediamill 8

NII 19 NHKSTRL 7

UEC 17 ITI_CERTH 6

siegen_kobe_nict 13 HFUT 4

EURECOM 10 CMU 3

FIU 10 LIG 2

UCF 10 IRIM 1

Fewer unique shots

compared to TV2014,

TV2013 & TV2012

14

Page 13: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

0

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

D_

Me

dia

Mill

.15

D_

Me

dia

Mill

.15

D_

Me

dia

Mill

.15

D_

Me

dia

Mill

.15

D_

Wa

sed

a.1

5D

_W

ased

a.1

5D

_W

ased

a.1

5D

_W

ased

a.1

5D

_T

okyoT

ech

.15

D_

To

kyoT

ech

.15

D_

To

kyoT

ech

.15

D_

IRIM

.15

D_

LIG

.15

D_

LIG

.15

D_

IRIM

.15

D_

IRIM

.15

D_

To

kyoT

ech

.15

D_

Pic

SO

M.1

5D

_P

icS

OM

.15

D_

UC

F_C

RC

V.1

5D

_LIG

.15

D_

Pic

SO

M.1

5D

_IR

IM.1

5D

_U

CF

_C

RC

V.1

5D

_LIG

.15

D_

Pic

SO

M.1

5D

_E

UR

EC

OM

.15

D_

EU

RE

CO

M.1

5D

_U

CF

_C

RC

V.1

5D

_E

UR

EC

OM

.15

D_

EU

RE

CO

M.1

5C

_C

MU

.15

D_

UC

F_C

RC

V.1

5C

_C

MU

.15

D_

ITI_

CE

RT

H.1

5D

_IT

I_C

ER

TH

.15

D_

ITI_

CE

RT

H.1

5D

_IT

I_C

ER

TH

.15

D_

UE

C.1

5D

_U

EC

.15

A_

NII_

Hita

chi_

UIT

.15

A_

NII_

Hita

chi_

UIT

.15

A_

NII_

Hita

chi_

UIT

.15

A_

NII_

Hita

chi_

UIT

.15

D_

insig

htd

cu.1

5D

_in

sig

htd

cu.1

5D

_in

sig

htd

cu.1

5D

_in

sig

htd

cu.1

5D

_sie

ge

n_

kob

e_

nic

t.1

5D

_sie

ge

n_

kob

e_

nic

t.1

5D

_U

EC

.15

A_

FIU

_U

M.1

5D

_sie

ge

n_

kob

e_

nic

t.1

5A

_F

IU_

UM

.15

A_

FIU

_U

M.1

5A

_F

IU_

UM

.15

Main runs scores – 2015 submissions

Median = 0.239

Me

an

In

fAP.

Type D runs (both IACC and non-IACC non-TRECVID )

Type A runs (only IACC for training)

Type C runs (both IACC and non-IACC TRECVID )

Higher median

and max scores

than 2014

15

Page 14: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

0

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

D_M

ed

iaM

ill.1

5D

_M

ed

iaM

ill.1

5D

_M

ed

iaM

ill.1

5D

_M

ed

iaM

ill.1

5D

_W

ase

da

.15

D_W

ase

da

.15

D_W

ase

da

.15

D_W

ase

da

.15

D_T

okyo

Te

ch

.15

D_T

okyo

Te

ch

.15

D_T

okyo

Te

ch

.15

D_IR

IM.1

5D

_L

IG.1

5D

_L

IG.1

5D

_IR

IM.1

5D

_IR

IM.1

5D

_T

okyo

Te

ch

.15

D_n

ist.b

ase

line

.15

D_

Pic

SO

M.1

5D

_P

icS

OM

.15

D_

UC

F_

CR

CV

.15

D_L

IG.1

5D

_P

icS

OM

.15

D_IR

IM.1

5D

_U

CF

_C

RC

V.1

5D

_L

IG.1

5D

_P

icS

OM

.15

D_E

UR

EC

OM

.15

D_E

UR

EC

OM

.15

D_

UC

F_

CR

CV

.15

D_L

IG.1

4D

_IR

IM.1

4D

_IR

IM.1

4D

_L

IG.1

4D

_E

UR

EC

OM

.15

D_E

UR

EC

OM

.15

A_

LIG

.13

A_

LIG

.13

A_

Vid

eo

Sen

se.1

3A

_IR

IM.1

3A

_in

ria.le

ar.

13

A_

inri

a.le

ar.

13

A_

axe

s.1

3A

_axe

s.1

3D

_E

UR

EC

OM

.14

A_

inri

a.le

ar.

13

A_

axe

s.1

3A

_IR

IM.1

3D

_E

UR

EC

OM

.14

C_C

MU

.15

D_

UC

F_

CR

CV

.15

C_C

MU

.15

D_IT

I_C

ER

TH

.15

D_IT

I_C

ER

TH

.15

D_IT

I_C

ER

TH

.15

A_

NII.1

3A

_N

II.1

3D

_U

EC

.14

A_

ITI-

CE

RT

H.1

3A

_in

sig

htd

cu.1

3D

_IT

I_C

ER

TH

.15

A_

ITI-

CE

RT

H.1

3A

_N

HK

ST

RL

.13

D_U

EC

.15

D_U

EC

.14

D_U

EC

.15

A_

NII

_H

ita

ch

i_U

IT.1

5A

_N

II_H

ita

ch

i_U

IT.1

5A

_N

II_H

ita

ch

i_U

IT.1

5A

_N

II_H

ita

ch

i_U

IT.1

5D

_in

sig

htd

cu

.15

D_in

sig

htd

cu

.15

D_in

sig

htd

cu

.15

D_in

sig

htd

cu

.15

A_

insig

htd

cu.1

4D

_sie

ge

n_

ko

be_

nic

t.1

5A

_H

FU

T.1

3D

_sie

ge

n_

ko

be_

nic

t.1

5D

_U

EC

.15

A_

EU

RE

CO

M.1

3A

_E

UR

EC

OM

.13

A_

FIU

_U

M.1

5D

_sie

ge

n_

ko

be_

nic

t.1

5A

_F

IU_U

M.1

5A

_F

IU_U

M.1

5A

_F

IU_U

M.1

5A

_U

EC

.13

Median = 0.188

Me

an

In

fAP.

NIST median baseline run

* Submitted runs in 2013 against 2015 testing data (Progress runs)

Main runs scores – Including progress

16

* Submitted runs in 2014 against 2015 testing data (Progress runs)

Page 15: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1A

irpla

ne*

Anch

orp

ers

on

Ba

ske

tba

ll*

Bic

yclin

g*

Boa

t_S

hip

*

Bri

dges*

Bus*

Ca

r_R

acin

g

Che

eri

ng*

Com

pu

ters

*

Dan

cin

g

De

mon

str

ation_O

r_P

rote

st

Explo

sio

n_fire

Go

vern

men

t_le

aders

Instr

um

en

tal_

Music

ian*

Kitche

n

Moto

rcycle

*

Offic

e

Old

_pe

ople

Pre

ss_co

nfe

rence

Run

nin

g*

Te

lep

hon

es*

Thro

win

g

Fla

gs*

Hill

La

kes

Qu

adru

pe

d*

Sold

iers

Stu

dio

_W

ith_

Anchorp

ers

on

Tra

ffic

Median

Top 10 InfAP scores by concept

InfA

P.

* Common concept in TV201417

Most common

concept’s has

higher max scores

than TV14

Page 16: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Statistical significant differences among top 10 Main

runs (using randomization test, p < 0.05)

•Run name (mean infAP)

D_MediaMill.15_4 0.362

D_MediaMill.15_2 0.359

D_MediaMill.15_1 0.359

D_MediaMill.15_3 0.349

D_Waseda.15_1 0.309

D_Waseda.15_4 0.307

D_Waseda.15_3 0.307

D_Waseda.15_2 0.307

D_TokyoTech.15_1 0.299

D_TokyoTech.15_2 0.298

D_MediaMill.15_4

D_MediaMill.15_3

D_TokyoTech.15_1

D_TokyoTech.15_2

D_Waseda.15_1

D_Waseda.15_3

D_Waseda.15_4

D_Waseda.15_2

D_MediaMill.15_1

D_MediaMill.15_3

D_Waseda.15_1

D_Waseda.15_3

D_Waseda.15_4

D_Waseda.15_2

D_TokyoTech.15_1

D_TokyoTech.15_2

D_MediaMill.15_2

D_MediaMill.15_3

D_Waseda.15_1

D_Waseda.15_3

D_Waseda.15_4

D_Waseda.15_2

D_TokyoTech.15_1

D_TokyoTech.15_2

18

Page 17: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Progress subtask

•Measuring progress of 2013, 2014, & 2015 systems

on IACC.2.C dataset.

•2015 systems used same training data and

annotations as in 2013 & 2014.

•Total 6 teams submitted progress runs against

IACC.2.C dataset.

19

Page 18: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

0

0.05

0.1

0.15

0.2

0.25

0.3

0.35

EURECOM IRIM ITI_CERTH LIG UEC insightdcu

Mea

n I

nfA

P

2013_system

2014_system

2015_system

Randomization tests show that 2015 systems are better than

2013 & 2014 systems (except for UEC, 2014 is better)

Progress subtask: Comparing best

runs in 2013, 2014 & 2015 by team

20

Page 19: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

Progress subtask: Concepts

improved vs weaken by team

21

EURECOM IRIM insightdcu LIG UEC ITI_CERTH

better than 2014 23 24 19 25 6

better than 2013 30 25 14 25 30 21

worse than 2013 0 5 16 5 0 9

worse than 2014 6 6 10 5 21

same as 2014 1 0 1 0 3

same as 2013 0 0 0 0 0 0

0

5

10

15

20

25

30

35

No. of C

on

cep

ts

Most 2015

concepts

improved

Page 20: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

2015 Observations

• 2015 main task was harder than 2014 main task that was itself

harder than 2013 main task (different data and different set of target

concepts)

• Raw system scores have higher Max and Median compared to

TV2014 and TV2103, still relatively low but regularly improving

• Most common concepts with TV2015 have higher median scores.

• Most Progress systems improved significantly from 2014 to 2015 as

this was also the case from 2013 to 2014.

• Stable participation (15 teams) between 2014 and 2015 (but was 26

teams for TV2013).

22

Page 21: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

2015 Observations - methods• Further moves toward deep learning

• More “deep-only” submissions

• Retraining of networks trained on ImageNet

• Use of many deep networks in parallel

• Data augmentation for training

• Use of multiple frames per shot for predicting

• Feeding of DCNNs with gradient and motion features

• Use of “deep features” (either final or hidden) with “classical” learning

• Hybrid DCNN-based/classical systems

• Engineered features still used as a complement (mostly Fisher

Vectors, SuperVectors, improved BoW, and similar) but no new

development

• Use of re-ranking or equivalent methods

23

Page 22: TRECVID-2015 Semantic Indexing task: Overview...TLP, LIP6, LIRIS, LISTIC LIG Laboratoire d'Informatique de Grenoble NII_Hitachi_UIT Natl.Inst. Of Info.; Hitachi Ltd; U. of Inf. Tech.(HCM-UIT)

SIN 2016 ?

• No SIN task is planned for 2016

• Resuming the ad hoc video retrieval task is

considered instead

24


Recommended