34
INEX 2002 - 2006: Understanding XML Retrieval Evaluation Mounia Lalmas and Anastasios Tombros Queen Mary, University of London Norbert Fuhr University of Duisburg-Essen

INEX 2002 - 2006: Understanding XML Retrieval Evaluation

  • Upload
    karik

  • View
    36

  • Download
    1

Embed Size (px)

DESCRIPTION

INEX 2002 - 2006: Understanding XML Retrieval Evaluation. Mounia Lalmas and Anastasios Tombros Queen Mary, University of London Norbert Fuhr University of Duisburg-Essen. Book. Chapters. Sections. Subsections. - PowerPoint PPT Presentation

Citation preview

Page 1: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

INEX 2002 - 2006: Understanding XML Retrieval Evaluation

Mounia Lalmas and Anastasios Tombros

Queen Mary, University of London

Norbert Fuhr

University of Duisburg-Essen

Page 2: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

XML retrieval vs. document retrieval(Retrieval of structured vs. unstructured documents)

• No predefined unit of retrieval

• Dependency of retrieval units

• Aims of XML retrieval:– Not only to find relevant

elements– But those at the appropriate

level of granularity

Book

Chapters

Sections

SubsectionsXML retrieval allows users to retrieve document components that are more focused, e.g. a subsection of a book instead of an entire book.

Page 3: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Outline

• Collections

• Topics

• Retrieval tasks

• Relevance and assessment procedures

• Metrics

Page 4: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Evaluation of XML retrieval: INEXPromote research and stimulate development of XML information access and retrieval, through

Creation of evaluation infrastructure and organisation of regular evaluation campaigns for system testing

Building of an XML information access and retrieval research community

Construction of test-suites

Collaborative effort participants contribute to the development of the collection

End with a yearly workshop, in December, in Dagstuhl, Germany

INEX has allowed a new community in XML information access to emerge

Page 5: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

INEX: Background

University of Amsterdam, NLUniversity of Otago, NZUniversity of WaterlooCWI, NLCarnegie Mellon University, USAIBM Research Lab, ILUniversity of Minnesota Duluth, USAUniversity of Paris 6, FR

Queensland University of Technology, AUSUniversity of California, Berkeley, USARoyal School of LIS, DKQueen Mary, University of London, UKUniversity of Duisburg-Essen, DEINRIA-Rocquencourt, FRYahoo! ResearchMicrosoft Research Cambridge, UKMax-Planck-Institut fur Informatik, DE

Since 2002 Sponsored by DELOS Network of Excellence for Digital

Libraries under FP5 and FP6 – IST programme Mainly dependent on voluntary efforts Coordination is distributed for tasks and tracks 64 participants in 2005; 80+ in 2006

Main Institutions involved in Coordination for 2006

Page 6: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Document collections

Yearnumber

documentsnumber

elementssize

average number

elements

average element depth

corpus

2002-2004-

12,107 8M 494MB 1,532 6.9IEEE

Journal Articles

2005 16,819 11M 764MB ‘’ ‘’

2006 659,388 30M

60GB 4.6GB

w/o images

161.35 6.72 Wikipedia

Page 7: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Topics

Page 8: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Two types of topics in INEX

• Content-only (CO) topics– ignore document structure– simulates users, who do not have any knowledge of

the document structure or who choose not to use such knowledge

• Content-and-structure (CAS) topics – contain conditions referring both to content and

structure of the sought elements– simulate users who do have some knowledge of the

structure of the searched collection

Page 9: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

CO topics 2003-2004<title>

"Information Exchange", +"XML", "Information Integration"</title><description> How to use XML to solve the information exchange (information integration)

problem, especially in heterogeneous data sources? </description><narrative> Relevant documents/components must talk about techniques of using XML to solve information exchange (information integration) among heterogeneous data sources where the structures of participating data sources are different although they might use the same ontologies about the same content. </narrative>

Page 10: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

CAS topics 2003-2004<title> //article[(./fm//yr = '2000' OR ./fm//yr = '1999') AND about(., '"intelligent

transportation system"')]//sec[about(.,'automation +vehicle')]</title><description> Automated vehicle applications in articles from 1999 or 2000 about

intelligent transportation systems.</description><narrative> To be relevant, the target component must be from an article on

intelligent transportation systems published in 1999 or 2000 and must include a section which discusses automated vehicle applications, proposed or implemented, in an intelligent transportation system.

</narrative>

Page 11: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

CO+S topics 2005-2006 <title>markov chains in graph related algorithms</title> <castitle>//article//sec[about(.,+"markov chains" +algorithm +graphs)]

</castitle> <description>Retrieve information about the use of markov chains in graph theory and in graphs-related algorithms. </description> <narrative>I have just finished my Msc. in mathematics, in the field of stochastic processes. My research was in a subject related to Markov chains. My aim is to find possible implementations of my knowledge in current research. I'm mainly interested in applications in graph theory, that is, algorithms related to graphs that use the theory of markov chains. I'm interested in at least a short specification of the nature of implementation (e.g. what is the exact theory used, and to which purpose), hence the relevant elements should be sections, paragraphs or even abstracts of documents, but in any case, should be part of the content of the document (as opposed to, say, vt, or bib). </narrative>

Page 12: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Expressing structural constraints: NEXI

• Narrowed Extended XPath I• INEX Content-and-Structure (CAS) Queries• Specifically targeted for content-oriented XML

search (i.e. “aboutness”)

//article[about(.//title, apple) and about(.//sec, computer)]

Page 13: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Retrieval Tasks

Page 14: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Retrieval tasks

• Ad hoc retrieval:“a simulation of how a library might be used and involves the searching of a static set of XML documents using a new set of topics”

– Ad hoc retrieval for CO topics– Ad hoc retrieval for CAS (+S) topics

• Core task:– “identify the most appropriate granularity XML

elements to return to the user, with or without structural constraints”

Page 15: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

CO retrieval task (2002 - )

• Specification:– make use of the CO topics– retrieves the most specific elements and only those,

which are relevant to the topic– no structural constraints regarding the appropriate

granularity – must identify the most appropriate XML elements to

return to the user

• Two main strategies• Focused strategy• Thorough strategy

Page 16: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Focused strategy (2005 - )

• Specification:“find the most exhaustive and specific element on a path within a given document containing relevant information and return to the user only this most appropriate unit of retrieval”

– no overlapping elements– preference for specificity over exhaustivity

Page 17: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Thorough strategy (2002 - )

• Specification:– “core system's task underlying most XML retrieval

strategies, which is to estimate the relevance of potentially retrievable elements in the collection”

– overlap problem viewed as an interface and presentation issues

– challenge is to rank elements appropriately

• Task that most XML approaches performed up to 2004 in INEX.

Page 18: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Fetch & Browse - 2005• Document ranking, and in each document, element

ranking

• Query: wordnet information retrieval

Page 19: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Fetch & Browse - 2006• Document ranking, and in each document

– All in context task: rank relevant elements, no overlap allowed (actual refinement of fetch & Browse)

– Best in context task: identify the one element from where to start reading in the document

• Likely to be the two tasks in INEX 2007

Page 20: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Retrieval strategies - to recap

• Focussed: assume that user prefers a single element that is the most relevant.

• Thorough: assume that user prefers all highly relevant elements.

• All In Context: assume that user interested in highly relevant elements that are contained only within highly relevant articles.

• Best In Context: assume that user interested in the best entry points, one per article, of highly relevant articles

Page 21: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Relevance and assessment procedures

Page 22: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Relevance in XML retrievalsmallest component (specificity) that is highly relevant (exhaustivity)

specificityspecificity: extent to which a document component is focused on the information need, while being an informative unitexhaustivityexhaustivity: extent to which the information contained in a document component satisfies the information need.

XML retrieval evaluation

XML retrieval

article

ss1 ss2

s1 s2 s3

XML evaluation

Query: ‘XML retrieval evaluation’

Page 23: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Relevance in XML retrieval: INEX 2003 - 2004

• Relevance = (0,0) (1,1) (1,2) (1,3) (2,1) (2,2) (2,3) (3,1) (3,2) (3,3)

exhaustivity = how much the section discusses the query: 0, 1, 2, 3

specificity = how focused the section is on the query: 0, 1, 2, 3

• If a subsection is relevant so must be its enclosing section, ...

XML retrieval evaluation

XML retrieval

article

ss1 ss2

s1 s2 s3

XML evaluation

Page 24: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Relevance assessment task

• Pooling technique

• CompletenessCompleteness– Rules that force assessors to assess related elements – E.g. element assessed relevant its parent element and

children elements must also be assessed– …

• ConsistencyConsistency– Rules to enforce consistent assessments– E.g. Parent of a relevant element must also be relevant,

although to a different extent– E.g. Exhaustivity increases going up; specificity increases

going down– …

Page 25: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Quality of assessments• Very laborious assessment task, eventually impacting on

the quality of assessments

• Interactive study shows that assessors agreement levels are high only at extreme ends of the relevance scale (very vs. not relevant)

• Statistical analysis of 2004 data showed that comparisons of approaches would lead to same outcomes using a reduced scale

• A simplified assessment procedure based on highlighting

Page 26: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Relevance in XML - 2005specificity defined continuous scale defined as ratio (in characters) of the highlighted text to element size.

Exhaustivity

–Highly exhaustive (2):

–Partly exhaustive (1)

–Not exhaustive (0)

–Too Small (?)

New assessment procedure led to better quality assessments

Page 27: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Latest analysis• Statistical analysis on the INEX 2005 data:

– The exhaustivity 3+1 scale is not needed in most scenarios to compare XML retrieval approaches

– “too small” may be simulated by some threshold length

• INEX 2006 used only the specificity dimension to “measure” relevance– The same highlighting approach is used

• Use of a highlighting procedure simplifies everything and is enough to “properly” compare the effectiveness of XML retrieval systems

Page 28: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Metrics

Page 29: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Measuring effectiveness: Metrics

• A research problem in itself!• Quantizations reflecting preference scenarios

• Metrics inex_eval - official INEX metric through 2004 inex_eval_ng (consider overlap & size) ERR (expected ratio of relevant units) XCG (XML cumulative gain) - official INEX metric 2005 t2i (tolerance to irrelevance) PRUM (Precision Recall with User Modelling) HiXEval …..

Page 30: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

XML retrieval allows users to retrieve document components that are more focussed, e.g. a section of a book instead of an entire book

BUT: what about if the chapter or one the subsections is returned?

Near-misses

(3,3)

(3,2)

(3,1)

(1,3)

(exhaustivity, specificity) as defined in 2004

Page 31: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Retrieve the bestbest XML elements according to content and structure criteria (2004 scale):

• Most exhaustive and the most specific = (3,3)

• Near misses = (3,3) + (2,3) (1,3) specific• Near misses = (3, 3) + (3,2) (3,1) exhaustive• Near misses = (3, 3) + (2,3) (1,3) (3,2) (3,1) (1,2)

near-misses

Page 32: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Quantization functions - reward near misses (2004 scale)

Strict - no reward

General - some rewards

quantstrict e,s 1 if e,s (3,3)

0 otherwise

quantgen e,s

1.00 if e,s (3,3)

0.75 if e,s 2,3 , 3,2 , 3,1 0.50 if e,s 1,3 , 2,2 , 2,1 0.25 if e,s 1,1 , 1,2 0.00 if e,s 0,0

Page 33: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Other INEX tracks• Interactive (2004 - 2006)• Relevance feedback (2004 - 2006)• Natural language query processing (2004 - 2006)• Heterogeneous collection (2004 - 2006)

• Multimedia track (2005 - )• Document mining (2005 - ) together with PASCAL network -

http://xmlmining.lip6.fr/

• User - case studies (2006)• XML entity ranking (2006 - )

• Other tracks under discussion for 2007, including a book search track

Page 34: INEX 2002 - 2006:  Understanding XML Retrieval Evaluation

Looking Forward

• Much recent work on evaluation• Larger more realistic collection - Wikipedia

– More assessed topics!– Better suite for analysis and reusability

• Better understanding of– information needs and retrieval scenarios– measuring effectiveness

• Introduction of a passage retrieval task in INEX 2007

Questions?