Professional Documents
Culture Documents
TREC Revised
TREC Revised
PROJECT REPORT
ON
TREC
SUBMITTED BY
SUBHAYAN CHATTERJEE
SESSION: 2019-2021
JADAVPUR UNIVERSITY
GUIDED BY
1
Acknowledgement
Date: 10.04.2020
2
CONTENT
Introduction 4
Objectives 4
TREC Facts 4
Origin of TREC 5
Goals for the TREC experiments 5
Information activities of TREC 6
TREC Tracks 6
TREC Collections 8
TREC Topics 9
Relevance Judgements in TREC 9
TREC 1 10
TREC 2 10
TREC 3-5, 6 11
TREC 7 11
TREC 8-12 12
TREC Yearly Conference Cycle 12
TREC Publications 13
Celebrating 25 years of TREC 13
Evaluation of TREC 14
Benefits of TREC experiments 15
Conclusion 16
References 17
3
INTRODUCTION
OBJECTIVES
TREC FACTS
Non-English retrieval
4
ORIGIN OF TREC
5
Information activities of TREC
TREC TRACKS
Live QA Track
OpenSearch Track
Tasks Track
6
Contextual Suggestion Track
Live QA Track
OpenSearch Track
Tasks Track
PAST TRACKS
Chemical Track
Crowdsourcing Track
Genomics Track
Enterprise Track
Cross-LanguageTrack
FedWeb Track
Filtering Track
HARD Track
7
Interactive Track
Legal Track
Novelty Track
Session Track
Spam Track
Terabyte Track
Video Track
Web Track
TREC COLLECTIONS
8
TREC TOPICS
It is impossible to calculate the absolute recall for each query. TREC uses a
Specific method called pooling for calculating relative recall as opposed to
absolute recall. In this method of estimating recall, all the relevant
documents that occurred in the top l00 documents for each system and for
each query are combined together to produce a ‘pool’ of relevant. By
pooling all the results from all the participating teams, one can expect that
most of the relevant documents in the collection have been found. The ad
hoc tasks in TREC are evaluated using a package called trec_eval, which
reports about 85 different numbers for a run, including recall and precision
measures at various cut-off points and a single value summary measure
from recall and precision.
9
TREC 1
In November I992, TREC-l (the first Text Retrieval Conference) was held at
NIST The conference, co-sponsored by DARPA and NIST, brought together
information retrieval researchers to compare the results of their different
systems when used on a large new test collection (called the TIPSTER
collection). The first conference attracted 28 groups from academia and
industry, and generated widespread interest from the information retrieval
community.
Harman reports that the draft results of the TREC-1 experiments revealed the
following facts:
TREC 2
Took place in August I993. In addition to 22 of the TREC-l groups, nine new
groups took part, bringing the total number of participating groups to 31.
The participants were able to choose from three levels of participation;
category A, full participation; category B, full participation using one-
quarter of the full document set; and category C, for evaluation only. Two
types of retrieval were examined: retrieval using an ‘ad hoc’ query, such as
a researcher might use in a library environment, and retrieval using a
‘routing’ query, such as a profile to filter some incoming document stream.
The number of documents to be returned was increased from 200 per topic
to 1000; and the total database size was increased from roughly l gigabyte
to 3 gigabytes.
10
TREC 3 TO TREC 5
TREC-3 introduced new topics with shorter descriptions, allowing for more
innovative topic expansion ideas. First two TRECs used very long topics
(averaging about 130 terms), in TREC-3 they were made shorter by
excluding some keywords, and in TREC-4 they were made even shorter to
investigate the problems with very short user statements (containing
around ten terms). TREC-5 included both short and long versions of the
topics with the goal of carrying out deeper investigations into which types
of techniques work well on various lengths of topics.
TREC 6
TREC 7
In addition to the main ad hoc task, TREC-7 contained seven tracks out of
which two tracks - query track and very large corpus track - were new. The
goal of the query track was to create a large query collection. The query
track was designed as a means of creating a large set of different queries
for an existing TREC topic set, topics 1 to 50.
11
TREC-8 to TREC-12
12
TEXT RETRIEVAL CONFERENCE PUBLICATIONS
Presentations
Proceedings
Some important TREC publications include books like ‘TREC Experiment and
Evaluation in Information Retrieval’, edited by Ellen M. Voorhees and Donna K.
Harman.
Webcast of Celebration
Celebration Agenda
13
OTHER EVALUATIONS OF TREC
The 10th edition of the annual meeting of Forum for Information Retrieval
Evaluation (fire.irsi.res.in). Since its inception in 2008, FIRE has had a strong
focus on shared tasks, similar to those offered at Evaluation forums like
TREC, CLEF and NTCIR. The shared tasks focus on solving specific problems
in the area information access and more importantly help in generating
evaluation datasets for the research community.
14
2) To increase communication among industry, academia, and government
by creating an open forum for the exchange of research ideas
1. Boolean retrieval
9. Dictionary-based stemming
10. Question-answering
15
CONCLUSION
One can conclude that TREC has been a vehicle not only for improving retrieval
technology, but also for providing a better understanding of removal
evaluation. TREC series of experiments have brought together researchers
from across the world to work on common. Goal is build up large text
collection.
16
REFERENCES
https://en.wikipedia.org/wiki/Text_Retrieval_Conference
https://trec.nist.gov/evals.html
17