36
TRECVID 2011 Content Based Copy Detection task overview Wessel Kraaij TNO, Radboud University Nijmegen George Awad NIST

TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

TRECVID 2011

Content Based Copy Detection

task overview

Wessel Kraaij

TNO, Radboud University Nijmegen

George Awad

NIST

Page 2: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Background

• Copy detection is applied in several real-word tasks: • television advertisement monitoring

• detection of copyright infringement

• detection of known (illegal) content

• Initial framework developed by NoE MUSCLE/INRIA

• Evaluation procedure re-defined at TV08, consolidated at TV09 (FA/Miss rate; actual vs optimal NDCR), renewed in TV10 (used internet videos (IACC) of much shorter videos with variable frame rates, Camcorder feature back, just AV runs, and adjusted „balanced‟ profile).

• TV11: • Same dataset as in TV10 to be able to compare performance.

Page 3: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

CBCD task overview

• Goal:

• Build a benchmark collection for video copy detection methods

• Task:

• Given a set of reference (test) video collection and a set of 11256 queries,

• determine for each query if it contains a copy, with possible transformations, of video from the reference collection,

• and if so, from where in the reference collection the copy comes

• As in 2010, only one task type:

• Copy detection of video + audio (11256) queries

• At least 2 runs are required representing two application profiles (“no false alarms”, “balanced”).

Page 4: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Datasets and queries

• Dataset:

• Reference video collection:

• Testing data: IACC.1.A (~8000 videos, 200 hr, < 3.5min)

• Development data : IACC.1.tv10.training(~3200 videos, 200 hr, 3.6 - 4.1min)

• Non-reference video collection :

• Internet Archives videos (~12480 videos, ~4000 hr, 10 – 30min)

• Queries: (Developed by INRIA-IMEDIA software run at NIST)

• Types:

• Type 1: composed of a reference video only. (1/3)

• Type 2: composed of a reference video embedded in a non-reference video. (1/3)

• Type 3: composed of a non-reference video only. (1/3)

• 201 total original queries. 67 queries for each type.

• Type 1 & 2 durations (average of 27 sec)

• Type 3 durations (average of 89 sec)

Copies

Page 5: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Datasets and queries

• After creating the queries, each was transformed by NIST • 8 video transformations using tools developed by Laurent Joyeaux

(independent agent at INRIA)

• 7 audio transformations using tools developed by Dan Ellis (Columbia University)

• Yielding… • 8 * 201 = 1608 video queries

• 7 * 201 = 1407 audio queries

• 8 * 7 * 201 = 11256 audio+video queries

• 8 groups of duplicate videos were found: • Systems were asked to submit all found duplicate result items.

• Before evaluation, runs were filtered by removing duplicate result items that doesn‟t match the G.T reference video.

Page 6: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Some important task details/ assumptions

• Detection systems submit a run threshold.

• Systems are asked to output a list of possible copies (each associated with a desicion score).

• The run threshold is used to determine the asserted copies.

• A query can yield just one true positive

• A query can give rise to many false alarms

• Consequence: • Type I error modeled as false alarm rate

• Type II error modeled as Pmiss

Page 7: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Video transformations

• 8 Transformations were selected:

• Simulated camcording (T1) – by perspective transform, automatic gain control, and blurring effects.

• Picture in picture (T2)

• Insertions of pattern (T3)

• Strong re-encoding (T4)

• Change of gamma (T5)

• Decrease in quality (T6) - by introducing 3 randomly selected combination of Blur, Gamma, Frame dropping, Contrast, Compression, Ratio, White noise

• Post production (T8) – by introducing 3 randomly selected combination of Crop, Shift, Contrast, Text insertion, Vertical mirroring, Insertion of pattern, Picture in picture,

• Combination of 3 randomly selected transformations (T10) chosen from T2-T5, T6 and T8.

Page 8: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Audio transformations

• T1: nothing

• T2: mp3 compression

• T3: mp3 compression and multiband companding

• T4: bandwidth limit and single-band companding

• T5: mix with speech

• T6: mix with speech, then multiband compress

• T7: bandpass filter, mix with speech, compress

Page 9: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Evaluation metrics

Three main metrics were adopted:

1. Normalized Detection Cost Rate (NDCR) • measures error rates/probabilities on the test set:

• Pmiss (probability of a missed copy)

• Rfa (false alarm rate)

• combines them using assumptions about two possible realistic scenarios:

1 - No False Alarm profile:

• Copy target rate (Rtarget) = 0.005/hr

• Cost of a miss (CMiss) = 1

• Cost of a false alarm (CFA) = 1000

2 – Balanced profile:

• Copy target rate (Rtarget) = 0.005/hr

• Cost of a miss (CMiss) = 1

• Cost of a false alarm (CFA) = 1

2. F1 (how accurately the copy is located, harmonic mean of P and R)

3. Mean processing time per query

[Kraaij, Over, Fiscus, Joly,2009] Final CBCD Evaluation plan TRECVID 2009 v1.3

Page 10: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Evaluation metrics (2)

General rules:

• No two query result items for a given video can overlap.

• For multiple result items per query, one mapping of submitted

extents to ref extents is determined based on a combination of F1-

score and the decision score (using the Hungarian solution to the

Bipartite Graph matching problem).

• The reference data has been found if and only if: the asserted test

video ID is correct AND asserted copy and ref. video overlap.

Page 11: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Actual vs. Optimal

• Actual NDCR:

• Each transformation is evaluated on the single submiitted threshold

value. This models the real situation where systems do not know

the transformation in advance

• Optimal NDCR:

• The minimal NDCR is computed by sweeping through the result

list, for each combination of transformations

• This way, teams can see where there is still room for improvement

Page 12: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

AT&T Labs - Research CCD INS --- --- --- ---

Beijing University of Posts and Telecom.-MCPRL CCD INS KIS --- SED SIN

Brno University of Technology CCD --- --- *** SED SIN

CRIM - Vision & Imaging team CCD --- --- --- SED ---

France Telecom Orange Labs (Beijing) CCD --- --- --- --- SIN

Osaka Prefecture University CCD *** --- --- --- ---

INRIA-LEAR CCD --- *** MED SED ***

INRIA-TEXMEX CCD *** --- --- --- ---

Istanbul Technical University CCD --- --- --- --- ---

University of Kaiserslautern CCD --- --- --- --- SIN

KDDILabs CCD *** --- --- --- ---

NTT Communication Science Laboratories-CSL CCD --- --- --- --- ---

Peking University-IDM CCD --- --- --- SED ---

PRISMA-University of Chile CCD --- --- --- --- ---

RMIT University School of CS&IT CCD --- --- --- --- ---

Sun Yat-sen University – GITL CCD --- --- --- SED ---

Telephonica Research CCD --- --- --- --- ---

Tokushima University CCD INS --- --- --- ---

University of Queensland CCD --- --- --- --- SIN

USC Viterbi School of Engineering CCD *** --- --- --- ***

Xi'an Jiaotong University CCD *** *** *** SED ***

Zhejiang University CCD *** --- --- SED ---

--- : group didn‟t participate

** : group applied but didn‟t submit

22 Participants (finishers)

Page 13: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Submission types and counts Run type 2008 2009 2010 2011

V (video only) 48 53 - -

A (audio only) 1 12 - -

M (video + audio) 6 42 78 73

Total runs 55 107 78 73

Type M

(Balanced)

Type M

(NoFa)

Type M

(Balanced)

Type M

(NoFa)

41 37 41 32

Balanced number of submission types

2010 2011

Page 14: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Top “video + audio” runs

00.10.20.30.40.50.60.70.80.9

1

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

PKU

-ID

M.m

.bal

ance

d.ca

scad

eCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BPK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BPK

U-I

DM

.m.b

alan

ced.

casc

ade

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BPK

U-I

DM

.m.b

alan

ced.

casc

ade

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T5

8BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T58B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BCR

IM-V

ISI.m

.bal

ance

d.V4

8A66

T65B

CRIM

-VIS

I.m.b

alan

ced.

V48A

66T6

5BPK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

PKU

-ID

M.m

.bal

ance

d.ca

scad

ePK

U-I

DM

.m.b

alan

ced.

casc

ade

NTT

-CSL

.m.b

alan

ced.

2N

TT-C

SL.m

.bal

ance

d.2

PKU

-ID

M.m

.bal

ance

d.ca

scad

e

Act

. ND

CR

Actual Balanced

T1

T2

T3

T4

T5

T6

T7

T8

T8

T9

T9

T10

T1

1

T12

T1

3

T14

T1

5

T15

T1

6

T16

T1

7

T18

T1

8

T19

T1

9

T20

T2

0

T21

T2

1

T22

T2

3

T24

T2

5

T26

T2

7

T28

T2

9

T29

T3

0

T30

T3

1

T31

T3

2

T32

T3

3

T33

T3

4

T34

T3

5

T35

T3

6

T37

T3

8

T39

T4

0

T40

T4

0

T41

T4

1

T42

T4

2

T50

T5

1

T51

T5

2

T52

T5

3

T54

T5

4

T55

T5

5

T56

T5

6

T64

T6

5

T66

T6

7

T68

T6

9

T70

T7

0

VT

1

VT

2

VT

3

VT

4

VT

5

VT

6

VT

8

VT

10

Video

transformation 1

7 Audio

transformations

Page 15: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Top “video + audio” runs

00.10.20.30.40.50.60.70.80.9

1

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

INR

IA-L

EA

R.m

.ba

lan

ce

d.d

od

oP

KU

-ID

M.m

.ba

lan

ce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

CR

IM-V

ISI.

m.b

ala

nce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BP

KU

-ID

M.m

.ba

lan

ce

d.c

asca

de

CR

IM-V

ISI.

m.b

ala

nce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BP

KU

-ID

M.m

.ba

lan

ce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

CR

IM-V

ISI.

m.b

ala

nce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BP

KU

-ID

M.m

.ba

lan

ce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

INR

IA-T

EX

ME

X.m

.ba

lan

ce

d.z

ozo

INR

IA-L

EA

R.m

.ba

lan

ce

d.d

od

oIN

RIA

-TE

XM

EX

.m.b

ala

nce

d.z

ozo

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

CR

IM-V

ISI.

m.b

ala

nce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

BP

KU

-ID

M.m

.ba

lan

ce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

INR

IA-L

EA

R.m

.ba

lan

ce

d.d

od

oP

KU

-ID

M.m

.ba

lan

ce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

PK

U-I

DM

.m.b

ala

nce

d.c

asca

de

NT

T-C

SL.m

.ba

lan

ce

d.2

CR

IM-V

ISI.

m.b

ala

nce

d.V

48

A6

6T

58

BC

RIM

-VIS

I.m

.ba

lan

ce

d.V

48

A6

6T

65

B

Min

. N

DC

R

Optimal Balanced

T1

T2

T3

T4

T5

T6

T7

T8

T9

T10

T1

1

T12

T1

3

T14

T1

5

T15

T1

6

T16

T1

6

T17

T1

7

T18

T1

8

T19

T1

9

T20

T2

0

T21

T2

1

T22

T2

3

T24

T2

5

T26

T2

7

T28

T2

9

T29

T3

0

T30

T3

1

T31

T3

2

T32

T3

3

T33

T3

4

T34

T3

5

T35

T3

6

T37

T3

8

T39

T4

0

T41

T4

2

T50

T5

1

T51

T5

2

T52

T5

3

T53

T5

4

T54

T5

5

T55

T5

6

T56

T6

4

T65

T6

6

T66

T6

7

T68

T6

9

T70

T7

0

VT

1

VT

2

VT

3

VT

4

VT

5

VT

6

VT

8

VT

10

Page 16: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Top “video+audio” runs

00.10.20.30.40.50.60.70.80.9

1

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

NT

T-C

SL.m

.no

fa.0

CR

IM-…

NT

T-C

SL.m

.no

fa.0

PK

U-I

DM

.m.n

ofa

.ca

sca

de

CR

IM-…

CR

IM-…

CR

IM-…

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

CR

IM-…

PK

U-I

DM

.m.n

ofa

.ca

sca

de

KD

DIL

ab

s.m

.no

fa.4

sy

sN

TT

-CS

L.m

.no

fa.0

KD

DIL

ab

s.m

.no

fa.4

sy

sN

TT

-CS

L.m

.no

fa.0

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

NT

T-C

SL.m

.no

fa.0

NT

T-C

SL.m

.no

fa.0

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

CR

IM-…

CR

IM-…

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

NT

T-C

SL.m

.no

fa.0

NT

T-C

SL.m

.no

fa.0

PK

U-I

DM

.m.n

ofa

.ca

sca

de

CR

IM-…

CR

IM-…

PK

U-I

DM

.m.n

ofa

.ca

sca

de

CR

IM-…

CR

IM-…

CR

IM-…

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

NT

T-C

SL.m

.no

fa.0

NT

T-C

SL.m

.no

fa.0

Act.

ND

CR

Actual Nofa

T1

T2

T3

T4

T5

T6

T7

T8

T9

T10

T1

1

T12

T1

3

T14

T1

5

T16

T1

7

T18

T1

9

T20

T2

1

T21

T2

2

T23

T2

3

T24

T2

4

T25

T2

5

T26

T2

6

T26

T2

7

T28

T2

8

T29

T2

9

T30

T3

0

T31

T3

1

T32

T3

2

T33

T3

3

T34

T3

5

T36

T3

7

T38

T3

9

T39

T4

0

T40

T4

1

T42

T5

0

T51

T5

2

T53

T5

4

T55

T5

6

T64

T6

5

T66

T6

7

T68

T6

9

T70

VT

1

VT

2

VT

3

VT

4

VT

5

VT

6

VT

8

VT

10

Page 17: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Top “video+audio” runs

00.10.20.30.40.50.60.70.80.9

1

INR

IA-L

EA

R.m

.no

fa.d

od

oIN

RIA

-LE

AR

.m.n

ofa

.do

do

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

PK

U-I

DM

.m.n

ofa

.ca

sca

de

INR

IA-L

EA

R.m

.no

fa.d

od

oIN

RIA

-LE

AR

.m.n

ofa

.do

do

INR

IA-L

EA

R.m

.no

fa.d

od

oP

KU

-ID

M.m

.no

fa.c

asc

ad

eP

KU

-ID

M.m

.no

fa.c

asc

ad

eP

KU

-ID

M.m

.no

fa.c

asc

ad

eP

KU

-ID

M.m

.no

fa.c

asc

ad

eIN

RIA

-LE

AR

.m.n

ofa

.do

do

PK

U-I

DM

.m.n

ofa

.ca

sca

de

INR

IA-L

EA

R.m

.no

fa.d

od

oC

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…N

TT

-CS

L.m

.no

fa.0

CR

IM-…

CR

IM-…

NT

T-C

SL.

m.n

ofa

.0IN

RIA

-LE

AR

.m.n

ofa

.do

do

INR

IA-L

EA

R.m

.no

fa.d

od

oC

RIM

-…C

RIM

-…IN

RIA

-LE

AR

.m.n

ofa

.do

do

PK

U-I

DM

.m.n

ofa

.ca

sca

de

INR

IA-L

EA

R.m

.no

fa.d

od

oIN

RIA

-LE

AR

.m.n

ofa

.do

do

NT

T-C

SL.

m.n

ofa

.0IN

RIA

-LE

AR

.m.n

ofa

.do

do

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

CR

IM-…

NT

T-C

SL.

m.n

ofa

.0N

TT

-CS

L.m

.no

fa.0

INR

IA-L

EA

R.m

.no

fa.d

od

oIN

RIA

-LE

AR

.m.n

ofa

.do

do

PK

U-I

DM

.m.n

ofa

.ca

sca

de

INR

IA-L

EA

R.m

.no

fa.d

od

oP

KU

-ID

M.m

.no

fa.c

asc

ad

eIN

RIA

-LE

AR

.m.n

ofa

.do

do

INR

IA-L

EA

R.m

.no

fa.d

od

oN

TT

-CS

L.m

.no

fa.0

INR

IA-L

EA

R.m

.no

fa.d

od

oP

KU

-ID

M.m

.no

fa.c

asc

ad

eC

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…C

RIM

-…P

KU

-ID

M.m

.no

fa.c

asc

ad

eIN

RIA

-LE

AR

.m.n

ofa

.do

do

INR

IA-L

EA

R.m

.no

fa.d

od

oIN

RIA

-LE

AR

.m.n

ofa

.do

do

NT

T-C

SL.

m.n

ofa

.0P

KU

-ID

M.m

.no

fa.c

asc

ad

eN

TT

-CS

L.m

.no

fa.0

INR

IA-L

EA

R.m

.no

fa.d

od

o

Min

. N

DC

R

Optimal Nofa

T1

T2

T2

T3

T4

T5

T6

T7

T8

T9

T10

T1

1

T12

T1

3

T14

T1

5

T15

T1

6

T16

T1

7

T17

T1

8

T18

T1

9

T19

T2

0

T21

T2

1

T21

T2

2

T23

T2

4

T24

T2

4

T24

T2

5

T26

T2

7

T28

T2

9

T29

T3

0

T30

T3

1

T31

T3

2

T32

T3

3

T33

T3

4

T35

T3

6

T37

T3

7

T38

T3

8

T39

T4

0

T41

T4

2

T50

T5

1

T51

T5

2

T52

T5

3

T53

T5

4

T54

T5

5

T55

T5

6

T56

T6

4

T65

T6

6

T67

T6

8

T68

T6

9

T70

VT

1

VT

2

VT

3

VT

4

VT

5

VT

6

VT

8

VT

10

Page 18: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

0.01

0.1

1

10

100

1 11 21 31 41 51 61

Min

ND

CR

Transformations

Opt. Median

1

2

3

4

5

6

7

8

9

10

Act. Median

1

2

3

4

5

6

7

8

9

10

Actual median values

are better than 2010

Balanced detection(Top 10 performance)

Page 19: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

0.001

0.01

0.1

1

10

100

1000

1 11 21 31 41 51 61

Min

ND

CR

Transformations

Opt. Median

1

2

3

4

5

6

7

8

9

10

Act. Median

1

2

3

4

5

6

7

8

9

10

Big gap between actual and optimal

medians! But better than 2010

Nofa detection (Top 10 performance)

Page 20: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

1 11 21 31 41 51 61

F1

Transformations

Opt. Median

10

9

8

7

6

5

4

3

2

1

Act. Median

10

9

8

7

6

5

4

3

2

1

Balanced localization (Top 10 performance)

Better precision than 2010

Page 21: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

1 11 21 31 41 51 61

F1

Transformations

Opt. Median

10

9

8

7

6

5

4

3

2

1

Act. Median

10

9

8

7

6

5

4

3

2

1

Nofa localization (Top 10 performance)

Better precision than 2010

Page 22: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

1

10

100

1000

1 11 21 31 41 51 61

Pro

cess

. Tim

e

Median

1

2

3

4

5

6

7

8

9

10

1

10

100

1000

1 11 21 31 41 51 61

Pro

cess

. Tim

e

Median

1

2

3

4

5

6

7

8

9

10

Balanced efficiency (Top 10 performance)

Very few fast systems!

Nofa efficiency (Top 10 performance)

Systems

are slower

than 2010

Page 23: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

Comparing best runs (detection)

0

0.05

0.1

0.15

0.2

0.25

0.3

1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49 51 53 55 57 59 61 63 65 67 69

Balanced Act

Nofa Act.

Balanced Opt.

Nofa Opt.

Audio transformations

5,6,7

Page 24: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

0

0.2

0.4

0.6

0.8

1

1 10 100 1000 10000

F1

Process. Time

T1

0

0.2

0.4

0.6

0.8

1

1 10 100 1000 10000

F1

Process. Time

T2

0

0.2

0.4

0.6

0.8

1

1 10 100 1000 10000

F1

Process. Time

T3

0

0.2

0.4

0.6

0.8

1

1 10 100 1000 10000

F1

Process. Time

T4

0

0.2

0.4

0.6

0.8

1

1 10 100 1000 10000

F1

Process. Time

T5

0

0.2

0.4

0.6

0.8

1

1 10 100 1000 10000

F1

Process. Time

T6

0

0.2

0.4

0.6

0.8

1

1 10 100 1000 10000

F1

Process. Time

T10

Actual Balanced runs by video transformations

(across all audio transformations)

Increasing proc. time did not significantly enhance localization. Very few systems achieved high localization in small proc. time.

0

0.2

0.4

0.6

0.8

1

1 10 100 1000 10000

F1

Process. Time

T8

On average,

systems are

slower than in

2010

Page 25: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

0.0010.010.1

110

1001000

10000

1 10 100 1000 10000

nd

cr

Process. Time

T8

0.01

0.1

1

10

100

1000

10000

1 10 100 1000 10000 100000

nd

cr

Process. Time

T1

0.0010.010.1

110

1001000

10000

1 10 100 1000 10000

nd

crProcess. Time

T2

0.0010.01

0.11

10100

100010000

1 10 100 1000 10000

nd

cr

Process. Time

T3

0.010.1

110

1001000

10000

1 10 100 1000 10000

nd

cr

Process. Time

T4

0.0010.010.1

110

1001000

10000

1 10 100 1000 10000

nd

cr

Process. Time

T5

0.0010.010.1

110

1001000

10000

1 10 100 1000 10000

nd

cr

Process. Time

T6

0.0010.010.1

110

1001000

1 10 100 1000 10000

nd

cr

Process. Time

T10

Actual Balanced runs by video transformations

(across all audio transformations)

In general, more processing time does not significantly improve detection. Very few strong & fast systems!

Page 26: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

0

0.2

0.4

0.6

0.8

1

0.001 0.01 0.1 1 10 100 1000 10000

F1

ndcr

T8

0

0.2

0.4

0.6

0.8

1

0.001 0.01 0.1 1 10 100 1000 10000

F1

ndcr

T5

0

0.2

0.4

0.6

0.8

1

0.01 0.1 1 10 100 1000 10000

F1

ndcr

T4

0

0.2

0.4

0.6

0.8

1

0.001 0.01 0.1 1 10 100 1000 10000

F1

ndcr

T10

0

0.2

0.4

0.6

0.8

1

0.001 0.01 0.1 1 10 100 1000 10000

F1

ndcr

T6

0

0.2

0.4

0.6

0.8

1

0.01 0.1 1 10 100 1000 10000

F1

ndcr

T1

0

0.2

0.4

0.6

0.8

1

0.001 0.01 0.1 1 10 100 1000 10000

F1ndcr

T2

0

0.2

0.4

0.6

0.8

1

0.001 0.01 0.1 1 10 100 100010000

F1

ndcr

T3

Actual Balanced runs by video transformations

(across all audio transformations)

Most of the systems that are good in separating copies from non-copies (low NDCR) are also good in localization.

Page 27: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

TV2009

TV2010

TV2011

Caution!...different dataset

Detection progress across 3 years

Page 28: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

TV2009

TV2010

TV2011

Caution!...different dataset

Localization progress across 3 years

Page 29: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

-AT&T: speed enhancements, new audio features, PiP detector

-BUPT: fusion of SIFT, color corellogram, LBP, audo, PiP detector

-Brno: BoW based on SIFT and SURF descriptors

-CRIM: added video processing, based on audio KNN approach, using visual features proposed by NTT@tv10

-FT Orange: 50000 visual words, spatial consistency filtering, EDF/CEPS audio features, audio has preference in results merging

29

Focus of participating teams

Page 30: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

-Osaka prefecture u.: study performance of existing image

recognition. Hashes of PCA reduced SIFT

-INRIA tex-mex: - presentation follows-

-Univ. Kaiserslautern: BoW approach plus Reciprocal geometry

verification

-KDDI: Focus on non-geometric transformations. DCT-sign based

feature representations. IDF weighting, temporal burstiness aware

scoring

-NTT: TV10 system: Coarsely-quantized area matching (CAM)

video, Divide and Locate (DAL) audio,improved to detect very short

copied segments. Simple merging (audio OR video) match

30

Focus of participating teams

Page 31: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

-PRISMA (Univ Chile+Telefonica): early fusion of audio and video

features – talk follows-

-RMIT: apply techniques from image recognition based on global

features: divide frame in border (intensity histogram) and central

region (auto-correlogram)

-Telefonica: video system from TV10 (DART), new audio system,

exploiting peak structure preservation under transformations

(MASK). Median bases score normalization and fusion experiments

– talk follows –

-University of Queensland: Combination of local binary patterns

and global HSV Color histogram. No audio analysis

31

Focus of participating teams

Page 32: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

-Top actual balanced detection scores are very near to top optimal balanced detection scores

-In general top balanced detection results are better than no false alarm.

- TV11 top balanced detection results are better than 2009 and 2010. Median scores for 2011 (<= 1) are lower than 2010.

-Bigger gap between actual and optimal medians for no false alarm detection results. However, TV11 looks better than TV10

32

Observations (NDCR)

Page 33: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

-TV11 top localization scores for both profiles are better

than TV10

- On average, TV11 systems are slower than TV10!

- Good detecting systems are also good in localization.

- Audio transformations 5,6 & 7 still seems to be the

hardest. while video transformations 3,4,5 & 6 seems to

be the easiest.

33

Observations

Page 34: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

-CCD task community is stable in size

-CCD attracts new TrecVID participants

-CCD platform is a good starting point for the INS task

-NDCR and F1 results do approach a ceiling

34

Observations

Page 35: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

-Did any one cross check their TV10 & TV11 system on

2010 & 2011 queries?

- It appears that efficiency is not an important goal for

participating groups

-Did any one run comparison between a+v vs v-only or

a-only?

-Do teams feel that their systems are getting better?

Why?

-What did systems learn over the past 4 years in this

task?

-Are there any remaining challenges?

TRECVID 2009 @ NIST 35

Questions

Page 36: TRECVID 2011 Content Based Copy Detection task overview · 2012. 1. 18. · CBCD task overview • Goal: • Build a benchmark collection for video copy detection methods • Task:

CfP Special Issue IEEE Multimedia

• Web-Scale Near-Duplicate Search: Techniques and

Applications

• Important Dates:

• Submission Deadline: 29 June 2012

• Publication Issue: July–September 2013

• More information at:

• http://www.computer.org/portal/web/computingnow/2013/mmcfp3