Evaluating Systems for Multilingual and Multimodal Information Access

Evaluating Systems for Multilingual and Multimodal Information Access

  • Thomas Deselaers
  • Nicola Ferro
  • Julio Gonzalo
  • Mikko Kurimo
  • Thomas Mandl
  • Vivien Petras
Publisher:SpringerISBN 13: 9783642044472ISBN 10: 3642044476

Paperback & Hardcover deals ―

Amazon IndiaGOFlipkart GOSnapdealGOSapnaOnlineGOJain Book AgencyGOBooks Wagon₹10,536Book ChorGOCrosswordGODC BooksGO

e-book & Audiobook deals ―

Amazon India GOGoogle Play Books ₹127.2Audible GO

* Price may vary from time to time.

* GO = We're not able to fetch the price (please check manually visiting the website).

Know about the book -

Evaluating Systems for Multilingual and Multimodal Information Access is written by Thomas Deselaers and published by Springer. It's available with International Standard Book Number or ISBN identification 3642044476 (ISBN 10) and 9783642044472 (ISBN 13).

The ninth campaign of the Cross-Language Evaluation Forum (CLEF) for European languages was held from January to September 2008. There were seven main eval- tion tracks in CLEF 2008 plus two pilot tasks. The aim, as usual, was to test the p- formance of a wide range of multilingual information access (MLIA) systems or s- tem components. This year, 100 groups, mainly but not only from academia, parti- pated in the campaign. Most of the groups were from Europe but there was also a good contingent from North America and Asia plus a few participants from South America and Africa. Full details regarding the design of the tracks, the methodologies used for evaluation, and the results obtained by the participants can be found in the different sections of these proceedings. The results of the CLEF 2008 campaign were presented at a two-and-a-half day workshop held in Aarhus, Denmark, September 17–19, and attended by 150 resear- ers and system developers. The annual workshop, held in conjunction with the European Conference on Digital Libraries, plays an important role by providing the opportunity for all the groups that have participated in the evaluation campaign to get together comparing approaches and exchanging ideas. The schedule of the workshop was divided between plenary track overviews, and parallel, poster and breakout sessions presenting this year’s experiments and discu- ing ideas for the future. There were several invited talks.