JULIO GONZALO, VÍCTOR PEINADO, PAUL CLOUGH & JUSSI KARLGREN CLEF 2009, CORFU iCLEF 2009 overview tags : image_search, multilinguality, interactivity, log_analysis,

Slides:



Advertisements
Similar presentations
Multilinguality & Semantic Search Eelco Mossel (University of Hamburg) Review Meeting, January 2008, Zürich.
Advertisements

Agent-Based Architecture for Intelligence and Collaboration in Virtual Learning Environments Punyanuch Borwarnginn 5 August 2013.
LogCLEF 2009 Log Analysis for Digital Societies (LADS) Thomas Mandl, Maristella Agosti, Giorgio Maria Di Nunzio, Alexander Yeh, Inderjeet Mani, Christine.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Statistical Machine Translation Part II – Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Developing and Evaluating a Query Recommendation Feature to Assist Users with Online Information Seeking & Retrieval With graduate students: Karl Gyllstrom,
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
The Challenges of Multilingual Search Paul Clough The Information School University of Sheffield ISKO UK conference 8-9 July 2013.
Presentation Title Presentation Subtitle and/or Conference Name Place Day Month Year First Name Last Name Job Title.
IR Challenges and Language Modeling. IR Achievements Search engines  Meta-search  Cross-lingual search  Factoid question answering  Filtering Statistical.
Semantic Web and Web Mining: Networking with Industry and Academia İsmail Hakkı Toroslu IST EVENT 2006.
Rutgers Components Phase 2 Principal investigators –Paul Kantor, PI; Design, modelling and analysis –Kwong Bor Ng, Co-PI - Fusion; Experimental design.
Intelligent Information Retrieval CS 336 Lisa Ballesteros Spring 2006.
Cross Language IR Philip Resnik Salim Roukos Workshop on Challenges in Information Retrieval and Language Modeling Amherst, Massachusetts, September 11-12,
INFO 624 Week 3 Retrieval System Evaluation
© Tefko Saracevic, Rutgers University1 digital libraries and human information behavior Tefko Saracevic, Ph.D. School of Communication, Information and.
Collaborative Cross-Language Search Douglas W. Oard University of Maryland, College Park May 14, 2015SICS Workshop.
 Ad-hoc - This track tests mono- and cross- language text retrieval. Tasks in 2009 will test both CL and IR aspects.
 Official Site: facility.org/research/evaluation/clef-ip-10http:// facility.org/research/evaluation/clef-ip-10.
Usability Methods: Cognitive Walkthrough & Heuristic Evaluation Dr. Dania Bilal IS 588 Spring 2008 Dr. D. Bilal.
User-Centred Interface Design for Cross-Language Information Retrieval Presented at SIGIR2002 Tampere - Finland Clarity is funded by EU - IST Program Contract.
Advisor: Hsin-Hsi Chen Reporter: Chi-Hsin Yu Date:
Library User Studies Hsin-liang Chen SLIS, Indiana University.
CS598CXZ Course Summary ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
Search Engines and Information Retrieval Chapter 1.
CLEF – Cross Language Evaluation Forum Question Answering at CLEF 2003 ( Bridging Languages for Question Answering: DIOGENE at CLEF-2003.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
Impressions of 10 years of CLEF Donna Harman Scientist Emeritus National Institute of Standards and Technology.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
CLEF 2004 – Interactive Xling Bookmarking, thesaurus, and cooperation in bilingual Q & A Jussi Karlgren – Preben Hansen –
Usability testing. Goals & questions focus on how well users perform tasks with the product. – typical users – doing typical tasks. Comparison of products.
The PATENTSCOPE search system: CLIR February 2013 Sandrine Ammann Marketing & Communications Officer.
The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield
SUMMON ® 2.0 DISCOVERY REINVENTED. What is Summon 2.0? A new, streamlined, modern interface New and enhanced features providing layers of contextual guidance.
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
MIRACLE Multilingual Information RetrievAl for the CLEF campaign DAEDALUS – Data, Decisions and Language, S.A. Universidad Carlos III de.
Comparing syntactic semantic patterns and passages in Interactive Cross Language Information Access (iCLEF at the University of Alicante) Borja Navarro,
UA in ImageCLEF 2005 Maximiliano Saiz Noeda. Index System  Indexing  Retrieval Image category classification  Building  Use Experiments and results.
Elaine Ménard & Margaret Smithglass School of Information Studies McGill University [Canada] July 5 th, 2011 Babel revisited: A taxonomy for ordinary images.
UNED at iCLEF 2008: Analysis of a large log of multilingual image searches in Flickr Victor Peinado, Javier Artiles, Julio Gonzalo and Fernando López-Ostenero.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
CLEF 2007 Workshop Budapest, Hungary, 19–21 September 2007 Nicola Ferro Information Management Systems (IMS) Research Group Department of Information Engineering.
Clarity Cross-Lingual Document Retrieval, Categorisation and Navigation Based on Distributed Services
Cross-Language Evaluation Forum CLEF 2003 Carol Peters ISTI-CNR, Pisa, Italy Martin Braschler Eurospider Information Technology AG.
Measuring How Good Your Search Engine Is. *. Information System Evaluation l Before 1993 evaluations were done using a few small, well-known corpora of.
QA Pilot Task at CLEF 2004 Jesús Herrera Anselmo Peñas Felisa Verdejo UNED NLP Group Cross-Language Evaluation Forum Bath, UK - September 2004.
What’s happening in iCLEF? (the iCLEF Flickr Challenge) Julio Gonzalo (UNED), Paul Clough (U. Sheffield), Jussi Karlgren (SICS), Javier Artiles (UNED),
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
Acceso a la información mediante exploración de sintagmas Anselmo Peñas, Julio Gonzalo y Felisa Verdejo Dpto. Lenguajes y Sistemas Informáticos UNED III.
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
Combining Text and Image Queries at ImageCLEF2005: A Corpus-Based Relevance-Feedback Approach Yih-Cheng Chang Department of Computer Science and Information.
Statistical Machine Translation Part II: Word Alignments and EM Alex Fraser Institute for Natural Language Processing University of Stuttgart
Multilingual Information Retrieval using GHSOM Hsin-Chang Yang Associate Professor Department of Information Management National University of Kaohsiung.
The CLEF 2005 interactive track (iCLEF) Julio Gonzalo 1, Paul Clough 2 and Alessandro Vallin Departamento de Lenguajes y Sistemas Informáticos, Universidad.
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
ASSOCIATIVE BROWSING Evaluating 1 Jin Y. Kim / W. Bruce Croft / David Smith by Simulation.
Usability Engineering Dr. Dania Bilal IS 587 Fall 2007.
User Errors in Formulating Queries and IR Techniques to Overcome Them Birger Larsen Information Interaction and Information Architecture Royal School of.
What is Research Design? RD is the general plan of how you will answer your research question(s) The plan should state clearly the following issues: The.
Summon® 2.0 Discovery Reinvented
Evaluation of IR Systems
College of Information
User Requirements in the Cultural Heritage Domain
CSE 635 Multimedia Information Retrieval
HCI Evaluation Techniques
Lab 2: Information Retrieval
Turn Page to View Presentation of Package
Presentation transcript:

JULIO GONZALO, VÍCTOR PEINADO, PAUL CLOUGH & JUSSI KARLGREN CLEF 2009, CORFU iCLEF 2009 overview tags : image_search, multilinguality, interactivity, log_analysis, web2.0

What CLIR researchers assume User needs information. Machine searches. User is happy (or not).

But finding is a matter of two smart slow Fast stupid Room for collaboration!

“Users screw things up” Can’t be resetDifferences between systems dissappearDifferences between interactive systems too! Who needs QA systems having a search engine and a user?

But CLIR is different

Help!

iCLEF methodology: hypothesis-driven hypothesis Reference & contrastive systems, topics, users latin-square pairing between system/topic/user Features:  Hypothesis-based (vs. operational)  Controlled (vs. ecological)  Deductive (vs. inductive)  Sound

iCLEF : tasks On newswire Cross-Language Document Selection Cross-Language query formulation and refinement Cross-Language Question Answering On image archives Cross-Language Image search.

Practical outcome!

iCLEF : problems Unrealistic search scenario, user sample opportunistic Experimental design not cost-effective  Only one aspect of CLIR at a time  High cost of recruiting, training, observing users.

Pick a document for “saffron”

Pick an illustration for “saffron”

Flickr

Topics Methodology  Ad hoc: find as many photographs of (different) european parliaments as possible.  Creative: find five illustrations for this article about saffron in Italy.  Visual: What is the name of the beach where this crab is lying on?  Participants must propose their own methodology and experiment design iCLEF 2006

Explored issues How users deal with native/passive/unknown languages? Do they actually use CLIR facilities when available? user’s behaviour Satisfaction (all tasks) Completeness (creative,ad-hoc) Quality (creative) user’s perceptions How many facets were retrieved (creative, ad- hoc) Was the image found? (visual) search effectiveness

iCLEF 2008/2009 Produce reusable dataset search log analysis task. Much larger set of users online game

iCLEF 2008/2009: Log Analysis Online game: see this image? Find it! (in any of six languages)Game interface features ML search assistanceUsers register with a language profileDataset: rich search log All search interactions Explicit success/failure Post-search questionnaires Queries Easy to find with the appropriate tags (  typically 3 tags) Hint mechanism (first target language, then tags)

Simultaneous search in six languages

Boolean search with translations

Relevance feedback

Assisted query translation

User profiles

User rank (Hall of Fame)

Group rank

Hint mechanism

Language skills bias in 2008

Selection of topics (images) No English annotations (new for 2009) Not buried in search results Visual cues No named entities

users / 41 teams 5101 complete search sessions Linguistics students, photography fans, IR researchers from industry and academia, monitored groups, other 130 users / 18 teams 2410 complete search sessions CS & linguistics students, photography fans, IR researchers from industry and academia, monitored groups, other. Harvested logs

Language skills bias in 2009

Log statistics

Distribution of users

Interface Native languages Language skills

English Spanish Language skills (II)

German Dutch Language skills (III)

French Italian Language skills (and IV)

Participants (I): log analysis Goal: correlation between lexical ambiguity in queries and search success Methodology: analysis of full search log University of Alicante Goal: correlations between several search parameters and search success Methodology: own set of users, search log analysis UAIC Goal: correlation between search strategies and search success Methodology: analysis of full search log UNED Goal: study confidence and satisfaction from search logs Methodology: analysis of full search log SICS

Participants (II): other strategies Goal: focus on users’ trust and confidence to reveal their perceptions of the task. Methodology: Own set of users, own set of queries, training, observational study, retrospective thinking aloud, questionnaires. Manchester Metropolitan University Goal: understanding challenges when searching images that have multilingual annotations. Methodology: Own set of users, training, questionnaires, interviews, observational analysis. University of North Texas

Discussion logs = “iCLEF legacy”  442 users w. heterogeneous language skills  7511 search sessions w. questionnaires iCLEF has been a success in terms of providing insights into interactive CLIR … and a failure in terms of gaining adepts?

So long!

And now… the iCLEF Bender Awards

VÍCTOR PEINADO, FERNANDO LÓPEZ-OSTENERO, JULIO GONZALO ICLEF2009 Analysis of Multilingual Image Search Sessions

Search Log Analysis 98 users with ≥ 15 sessions ≥ 1M log lines have been processed ( ) 5,243 search sessions w. questionnaires analysis of sessions comparing: Active / passive / unknown language profiles Successful / unsuccessful

Success vs. Language Skills

Cognitive Effort vs. Language Skills

Use of CL Facilities vs. Language Skills

Success / Failure

Cognitive Effort vs. Success

Use of CL facilities vs. success

Questionnaires Selecting/finding appropriate translations for the query terms is the most challenging aspect of the task In iCLEF09 80% of the users agree with the usefulness of the personal dictionary vs 59% who preferred the additional query terms suggested by the system 78% of iCLEF08 users missed results organized in tabs, 80% of iCLEF09 users complained about bilingual dictionaries iCLEF08 users claimed to have used their knowledge of target languages (90%), while iCLEF09 users opted for using additional dictionaries and other on-line sources (82%)

Discussion Easy task? (≥ 80% success for all profiles) Direct relation between use of CL facilities and lack of target language skills Positive correlation between use of relevance feedback and success CL facilities highly appreciated in a multilingual search scenario

Success rate and language profiles

Success rate vs. number of hints

Cognitive cost and language profiles

Search modes and language profiles Una vez que se conoce el idioma destino: Modo multilingüe: pasivos  23% + que activos Desconocidos  61% + que activos Modo monolingüe: pasivos  4% - que activos (!!!) Desconocidos  23% - que activos (!!!)

Learning effects: success and #hints

Learning effects: cognitive cost

Questionnaires after success Highest difficulty is cross-linguality for the “unknown” group.

Aprendizaje: exploración ranking

Assisted query translation vs. Relevance feedback

Perceived utility(I)

Perceived utility(II)