The CLEF 2005 Cross-Language Image Retrieval Track Organised by Paul Clough, Henning Müller, Thomas Deselaers, Michael Grubinger, Thomas Lehmann, Jeffery.

Slides:



Advertisements
Similar presentations
Using Large-Scale Web Data to Facilitate Textual Query Based Retrieval of Consumer Photos.
Advertisements

Pseudo-Relevance Feedback For Multimedia Retrieval By Rong Yan, Alexander G. and Rong Jin Mwangi S. Kariuki
Sheffield at ImageCLEF 2003 Paul Clough and Mark Sanderson.
Search Results Need to be Diverse Mark Sanderson University of Sheffield.
ImageCLEF breakout session Please help us to prepare ImageCLEF2010.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
Morris LeBlanc.  Why Image Retrieval is Hard?  Problems with Image Retrieval  Support Vector Machines  Active Learning  Image Processing ◦ Texture.
1 Multimedia I: Image Retrieval in Biomedicine William Hersh, MD Professor and Chair Department of Medical Informatics & Clinical Epidemiology Oregon Health.
Information Retrieval Concerned with the: Representation of Storage of Organization of, and Access to Information items.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
MANISHA VERMA, VASUDEVA VARMA PATENT SEARCH USING IPC CLASSIFICATION VECTORS.
Advance Information Retrieval Topics Hassan Bashiri.
 Ad-hoc - This track tests mono- and cross- language text retrieval. Tasks in 2009 will test both CL and IR aspects.
 Official Site: facility.org/research/evaluation/clef-ip-10http:// facility.org/research/evaluation/clef-ip-10.
Search is not only about the Web An Overview on Printed Documents Search and Patent Search Walid Magdy Centre for Next Generation Localisation School of.
LREC Combining Multiple Models for Speech Information Retrieval Muath Alzghool and Diana Inkpen University of Ottawa Canada.
Evaluation David Kauchak cs458 Fall 2012 adapted from:
Evaluation David Kauchak cs160 Fall 2009 adapted from:
Evaluating the Contribution of EuroWordNet and Word Sense Disambiguation to Cross-Language Information Retrieval Paul Clough 1 and Mark Stevenson 2 Department.
MediaEval Workshop 2011 Pisa, Italy 1-2 September 2011.
August 21, 2002Szechenyi National Library Support for Multilingual Information Access Douglas W. Oard College of Information Studies and Institute for.
CLEF Ǻrhus Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Oier Lopez de Lacalle, Arantxa Otegi, German Rigau UVA & Irion: Piek Vossen.
Text- and Content-based Approaches to Image Retrieval for the ImageCLEF 2009 Medical Retrieval Track Matthew Simpson, Md Mahmudur Rahman, Dina Demner-Fushman,
Impressions of 10 years of CLEF Donna Harman Scientist Emeritus National Institute of Standards and Technology.
1 Intra- and interdisciplinary cross- concordances for information retrieval Philipp Mayr GESIS – Leibniz Institute for the Social Sciences, Bonn, Germany.
1 The Domain-Specific Track at CLEF 2008 Vivien Petras & Stefan Baerisch GESIS Social Science Information Centre, Bonn, Germany Aarhus, Denmark, September.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
JULIO GONZALO, VÍCTOR PEINADO, PAUL CLOUGH & JUSSI KARLGREN CLEF 2009, CORFU iCLEF 2009 overview tags : image_search, multilinguality, interactivity, log_analysis,
Cross-Language Evaluation Forum CLEF Workshop 2004 Carol Peters ISTI-CNR, Pisa, Italy.
Concept-based Image Retrieval: The ARRS GoldMiner ® Image Search Engine Charles E. Kahn Jr., MD, MS Medical College of Wisconsin Milwaukee, Wisconsin,
Applying the KISS Principle with Prior-Art Patent Search Walid Magdy Gareth Jones Dublin City University CLEF-IP, 22 Sep 2010.
The CLEF 2003 cross language image retrieval task Paul Clough and Mark Sanderson University of Sheffield
Image Retrieval in Radiology: The ImageCLEF Challenge Charles E. Kahn, Jr. Medical College of Wisconsin Jayashree Kalpathy-Cramer Oregon Health & Science.
MIRACLE Multilingual Information RetrievAl for the CLEF campaign DAEDALUS – Data, Decisions and Language, S.A. Universidad Carlos III de.
Math Information Retrieval Zhao Jin. Zhao Jin. Math Information Retrieval Examples: –Looking for formulas –Collect teaching resources –Keeping updated.
UA in ImageCLEF 2005 Maximiliano Saiz Noeda. Index System  Indexing  Retrieval Image category classification  Building  Use Experiments and results.
UNED at iCLEF 2008: Analysis of a large log of multilingual image searches in Flickr Victor Peinado, Javier Artiles, Julio Gonzalo and Fernando López-Ostenero.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
How robust is CLIR? Proposal for a new robust task at CLEF Thomas Mandl Information Science Universität Hildesheim 6 th Workshop.
Database & Information Systems Group University of Basel October 10th, 2007 Similarity Search Michael Springmann PhD Seminar October 11 th, 2007.
Information Retrieval at NLC Jianfeng Gao NLC Group, Microsoft Research China.
TREC-CHEM The TREC Chemical IR Track Mihai Lupu 1, John Tait 1, Jimmy Huang 2, Jianhan Zhu 3 1 Information Retrieval Facility 2 York University 3 University.
Using Domain Ontologies to Improve Information Retrieval in Scientific Publications Engineering Informatics Lab at Stanford.
CLEF Kerkyra Robust – Word Sense Disambiguation exercise UBC: Eneko Agirre, Arantxa Otegi UNIPD: Giorgio Di Nunzio UH: Thomas Mandl.
From Text to Image: Generating Visual Query for Image Retrieval Wen-Cheng Lin, Yih-Chen Chang and Hsin-Hsi Chen Department of Computer Science and Information.
Thomas Mandl: GeoCLEF Track Overview Cross-Language Evaluation Forum (CLEF) Thomas Mandl, (U. Hildesheim) 8 th Workshop.
Medical retrieval 2008 New data set with almost 66,000 images Thirty topics were made available, ten in each of three categories: visual, mixed, and semantic.
Information Retrieval
What’s happening in iCLEF? (the iCLEF Flickr Challenge) Julio Gonzalo (UNED), Paul Clough (U. Sheffield), Jussi Karlgren (SICS), Javier Artiles (UNED),
Topic creation for medical image retrieval benchmarks ImageCLEF/MUSCLE workshop, Alicante, Henning Müller, Bill Hersh.
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
1 13/05/07 1/20 LIST – DTSI – Interfaces, Cognitics and Virtual Reality Unit The INFILE project: a crosslingual filtering systems evaluation campaign Romaric.
Cs Future Direction : Collaborative Filtering Motivating Observations:  Relevance Feedback is useful, but expensive a)Humans don’t often have time.
Combining Text and Image Queries at ImageCLEF2005: A Corpus-Based Relevance-Feedback Approach Yih-Cheng Chang Department of Computer Science and Information.
Thomas Mandl: Robust CLEF Overview 1 Cross-Language Evaluation Forum (CLEF) Thomas Mandl Information Science Universität Hildesheim
The CLEF 2005 interactive track (iCLEF) Julio Gonzalo 1, Paul Clough 2 and Alessandro Vallin Departamento de Lenguajes y Sistemas Informáticos, Universidad.
Analysis of Experiments on Hybridization of different approaches in mono and cross-language information retrieval DAEDALUS – Data, Decisions and Language,
Manual Query Modification and Automated Translation to Improve Cross-language Medical Image Retrieval Jeffery R. Jensen William R. Hersh Department of.
1 Knowledge-Based Medical Image Indexing and Retrieval Caroline LACOSTE Joo Hwee LIM Jean-Pierre CHEVALLET Daniel RACOCEANU Nicolas Maillot Image Perception,
Semantic search-based image annotation Petra Budíková, FI MU CEMI meeting, Plzeň,
1 DAFFODIL Effective Support for Using Digital Libraries Norbert Fuhr University of Duisburg-Essen, Germany.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
CLEF Workshop ECDL 2003 Trondheim Michael Kluck slide 1 Introduction to the Monolingual and Domain-Specific Tasks of the Cross-language.
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
CLEF Budapest1 Measuring the contribution of Word Sense Disambiguation for QA Proposers: UBC: Agirre, Lopez de Lacalle, Otegi, Rigau, FBK: Magnini.
Experiments for the CL-SR task at CLEF 2006
Introduction to Information Retrieval
Introduction to Search Engines
Future Direction : Collaborative Filtering
Active AI Projects at WIPO
Presentation transcript:

The CLEF 2005 Cross-Language Image Retrieval Track Organised by Paul Clough, Henning Müller, Thomas Deselaers, Michael Grubinger, Thomas Lehmann, Jeffery Jensen and William Hersh

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Overview Image Retrieval and CLEF Motivations Tasks in 2005 Ad-hoc retrieval of historic photographs and medical images Automatic annotation of medical images Interactive task Summary and future work

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Image Retrieval and CLEF Cross-language image retrieval Images often accompanied by text (used for retrieval) Began in 2003 as pilot experiment Aims of ImageCLEF Investigate retrieval combining visual features and associated text Promote the exchange of ideas Provide resources for IR evaluation

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Motivations Image retrieval a good application for CLIR Assume images are language-independent Many images have associated text (e.g. captions, metadata, Web page links) CLIR has potential benefits for image vendors and users Image retrieval can be performed using Low-level visual features (e.g. texture, colour and shape) Abstracted features expressed using text Combining both visual and textual approaches

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 ImageCLEF participants from 11 countries Specific domains and tasks Retrieval of historic photographs (St Andrews) Retrieval and annotation of medical images (medImageCLEF and IRMA) Additional co-ordinators William Hersh and Jeffrey Jensen (OHSU) Thomas Lehmann and Thomas Deselaers (Aachen) Michael Grubinger (Melbourne) Links with MUSCLE NoE including pre-CLEF workshop

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Ad-hoc retrieval from historic photographs Paul Clough (University of Sheffield) Michael Grubinger (Victoria University)

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 From: St Andrews Library historic photographic collection イングランドにある灯台の写真 Изображения английских маяков Fotos de faros ingleses Pictures of English lighthouses Kuvia englantilaisista majakoista Bilder von englischen Leuchttürmen صور لمنارات انجليزيه St Andrews image collection

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Topics 28 search tasks (topics) Consist of title, narrative and example images Topics more general than 2004 and more “visual” e.g. waves breaking on beach, dog in sitting position Topics translated by native speakers 8 languages for title & narrative (e.g. German, Spanish, Chinese, Japanese) 25 languages for title (e.g. Russian, Bulgarian, Norwegian, Hebrew, Croatian) 2004 topics and qrels used as training data

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Relevance judgements Staff from Sheffield University were assessors Assessors judged topic pools Top 50 images from all 349 runs Average of 1,376 images per pool 3 assessments per image (inc. topic creator) Ternary relevance judgements Qrels: images judged as relevant/partially relevant by topic creator and at least one other assessor

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Submissions & Results (1) 11 groups (5 new*) CEA* NII* Alicante CUHK* DCU Geneva Indonesia* Miracle NTU Jaen* UNED DimensionType#Runs (%)Avg. MAP LanguageEnglish119 (34%) Non-English230 (66%) Run typeAutomatic349 (100%) Feedback (QE)Yes142 (41%) No207 (59%) ModalityImage4 (1%) Text318 (91%) Text + Image27 (8%) Initial QueryImage4 (1%) Title274 (79%) Narrative6 (2%) Title + Narr57 (16%) Title + Image4 (1%) All4 (1%)0.3953

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Submissions & Results (2) Language#RunsMax. MAPGroupInitial QueryFeedbackModality English CUHKTitle+imgYesText+img Chinese (trad.) NTUTitle+narrYesText+img Spanish (Lat. Am.) Alicante/JaenTitleYesText Dutch Alicante/JaenTitleYesText Visual NTUVisualYesImage German Alicante/JaenTitleYesText Spanish (Euro.) UNEDTitleYesText Portuguese MiracleTitleNoText Greek DCUTitleYesText French JaenTitle+narrYesText Japanese AlicanteTitleYesText Russian DCUTitleYesText Italian MiracleTitleYesText

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Submissions & Results (3) Language#RunsMax. MAPGroupInitial QueryFeedbackModality Chinese (simpl.) AlicanteTitleYesText Indonesian IndonesiaTitleNoText+img Turkish MiracleTitleNoText Swedish JaenTitleNoText Norwegian MiracleTitleNoText Filipino MiracleTitleNoText Polish MiracleTitleNoText Romanian MiracleTitleNoText Bulgarian MiracleTitleNoText Czech MiracleTitleNoText Croatian MiracleTitleNoText Finnish MiracleTitleNoText Hungarian MiracleTitlenoText

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Summary Most groups focused on text retrieval Fewer combined runs than 2004 But still gives highest average MAP Translation main focus for many groups 13 languages have at least 2 groups More use of title & narrative than 2004 As Relevance feedback (QE) improves results Topics still dominated by semantics But typical of searches in this domain

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Ad-hoc medical retrieval task Henning Müller (University Hospitals Geneva) William Hersh, Jeffrey Jensen (OHSU)

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Collection 50,000 medical images 4 sub-collections with heterogeneous annotation Radiographs, photographs, Powerpoint slides and illustrations Mixed languages for annotations (French, German and English) In 2004 only 9,000 images available

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Search topics Topics based on 4 axes Modality (e.g. x-ray, CT, MRI) Anatomic region shown in image (e.g. head, arm) Pathology (disease) shown in image Abnormal visual observation (e.g. enlarged heart) Different types of topic identified from survey Visual (11) – visual approaches only expected to perform well Mixed (11) – text and visual approaches expected to perform well Semantic (3) – visual approaches not expected to perform well Topics consist of annotation in 3 languages and 1-3 query images

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 An example (topic # 20 - mixed) Show me microscopic pathologies of cases with chronic myelogenous leukemia. Zeige mir mikroskopische Pathologiebilder von chronischer Leukämie. Montre-moi des images de la leucémie chronique myélogène.

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Relevance assessments Medical doctors made relevance judgements Only one per topic for money and time constraints Some additional to verify consistency Relevant/partially relevant/non relevant For ranking only relevant vs. non-relevant Image pools created from submissions Top 40 images from 134 runs Average of 892 images per topic to assess

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Submissions 13 groups submitted runs (24 registered) Resources very interesting but lack of manpower 134 runs submitted Several categories for submissions Manual vs. Automatic Data source used Visual/textual/mixed All languages could be used or a single one

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Results (1) Mainly automatic and mixed submissions some further to be classified as manual Large variety of text/visual retrieval approaches Ontology-based Simple tf/idf weighting Manual classification before visual retrieval Query typesAutomaticManual Visual283 Textual141 Mixed862

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Results (2) – highest MAP Query typesAutomaticManual VisualI2Rfus.txt i2r-vk-avg.txt TextualIPALI2R_Tn OHSUmanual.txt MixedIPALI2R_TIan OHSUmanvis.txt 0.160

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Average results per topic type

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Summary Text-only approaches perform better than image-only But some visual systems have high early precision Depends on the topics formulated Visual systems very bad on semantic queries Best overall systems use combined approaches GIFT as a baseline system used by many participants and still best visual completely automatic Few manual runs

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Automatic Annotation Task Thomas Deselaers, Thomas Lehmann (RWTH Aachen University)

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Automatic annotation Goal Compare state-of-the-art classifiers for medical image annotation task Purely visual task Task 9,000 training & 1,000 test medical images from Aachen University Hospital 57 classes identifying modality, body orientation, body region and biological system (IRMA code) e.g. 01: plain radiography, coronal, cranuim, musculosceletal system Classes in English and German and unevenly distributed

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Example of IRMA code Example: Example: radiography, plain, analog, overview radiography, plain, analog, overview coronal, AP, supine coronal, AP, supine abdomen, middle abdomen, middle uropoetic system uropoetic system

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Example Images

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Participants Groups 26 registered 12 submitted runs Runs In total 41 submitted CEA (France) CINDI (Montreal,CA) medGift (Geneva, CH) Infocomm (Singapore, SG) Miracle (Madrid, ES) Umontreal (Montreal, CA) Mt. Holyoke College (Mt. Hol., US) NCTU-DBLAB (TW) NTU (TW) RWTH Aachen CS (Aachen, DE) IRMA Group (Aachen, DE) U Liège (Liège, BE)

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Results... Baseline error rate: 36.8%

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 Conclusions Continued global participation from variety of research communities Improvements in ad-hoc medical task Realistic topics Larger medical image collection Introduction of medical annotation task Overall combining text and visual approaches works well for ad-hoc task

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 ImageCLEF2006 and beyond …

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 ImageCLEF 2006 … New ad-hoc IAPR collection of 25,000 personal photographs Annotations in English, German and Spanish Medical ad-hoc Same data; new topics Medical annotation Larger collection; more fine-grained classification New interactive task Using Flickr.com … more in iCLEF talk

22/09/05 ImageCLEF: cross-language image retrieval at CLEF2005 … and beyond Image annotation task Annotate general images with simple concepts Using the LTU 80,000 Web images (~350 categories) MUSCLE collaboration Create visual queries for ad-hoc task (IAPR) Funding workshop in 2006 All tasks involve cross-language in some way