Professional Documents
Culture Documents
Retrieval Evaluation
Introduction to Retrieval evaluation
retrieval system that are used to assess how well the search/ browse
results satisfied the users query intent.
The evaluation is a systematic determination of subjects merit, worth
objectives and results of any search action that has been completed.
Retrieval evaluation …
The evaluation of information retrieval system measure which of the two
existing system perform better and try to assess how the level of
performance of a given can be improved.
Thus, the first type of evaluation which should be considered is a functional
analysis in which the specified system functionalities are tested one by one.
The evaluation metrics may be online metric focused at users interactions
with the search system (search engine models) while offline metrics
measure relevance from text/document collection, likely each result or
search engine results page (SERP) .
Retrieval evaluation …
It is a simple procedure which can be quite useful for catching
programming errors. Given that the system has passed the functional
analysis phase, one should proceed to evaluate the performance of the
system.
Lancaster state that the evaluation of information retrieval system can
it’s impact.
The evaluation helps to investigate the degree to which the state goals have
- A document collection
different viewpoints.
evaluation.
Retrieval performance Evaluation
batch mode (i.e., the user submits a query and receives an answer back) or of
a whole interactive session (i.e., the user specifies his information need
through a series of interactive steps with the system).
Further, the retrieval task could also comprise a combination of these two
strategies.
Batch and interactive query tasks are quite distinct processes and thus their
design, guidance provided by the system, and duration of the session are
critical aspects which should be observed and measured.
In a batch session, none of these aspects is nearly as important as the quality
experiments.
Retrieval performance Evaluation (Online Metrics)
Online metrics are generally created from search logs. The metrics are
where the judges score the quality of the search results. Mainly the very
common retrieval performance evaluation measures / criteria are: -
- Recall
- Precision
- Fallout
- Generality
query and 60 of these items are retrieved in a given search, then the recall is
stated to be 60% in other words the system has been able to retrieve 60% of
the relevant items. That is Recall = 60/100 *100 = 60
Therefore,
and 60 of them are relevant and retrieved for the user according his/her
request.
Retrieval performance Evaluation …
Precision: it is defined as the proportion of documents retrieved that is relevant.
Example:- in a given search the system retrieves 90 items, out of these 45 are
Therefore,
Fall-out:- Fall-out ration is a proportion of non-relevant items that have been
retrieved in a given transaction/ search.
Generality ratio is the relevant items that have been retrieved in a given search.
o F-score/ F-measure : it is a weighted harmonic mean of precision and recall,
Query:-
1. Write and discuss the relationship F-measure and E-measure of recall
and precision?
2. Discuss about the confusion matrices of recall and precision?
Limitation of Recall and Precision
A user is going to prepare a state-of-the-art report on a topic would like to have all the items
- Effectiveness : It is the level up to which the given system attain it’s objective. It
measures how far it can retrieve relevant information while with holding non-relevant information.
- Usability
- Satisfaction and
- Cost
Other Performance Evaluation criteria's
user interacts with the system. It takes into account the user and their
expectations, skills and experiences.
Satisfaction:
- Searching tasks
- Searching settings
Cost: users may experience costs in terms of any payment that they