Beruflich Dokumente
Kultur Dokumente
Term paper on
Evaluation of Information Retrieval
Systems
Submitted to
Prof. Mandar. R. Mutailkdesai
By
M Ramprabakaran
M.Tech.., ISiM
On the rise of new trends in the World Wide Web, Information retrieval as experiences a
tremendous changes. Evaluation is more important factor in information retrieval. This paper
summarizes evaluation of information retrieval and various approaches to Information retrieval
systems. Limitation is based on the change in the data being searched.
Introduction
Evaluati Modifying
ng the the query
user data No
Yes
Fig 1
Stop
(i) Coverage: To provide the relevant information, the coverage will include all the
content that is required to provide.
(ii) Time lag: The average time taken by the system to process by providing the seeking
information.
(iii) Form: form represent the way of producing the output.
(iv) Effort: The user requests the information and retrieves his requested answer.
(v) Recall: The recall denotes the fraction of relevant document, which has been
retrieved.
(vi) Precision: The precision denotes the fraction of retrieved, which is relevant.[1]
Tab1
So the binary relevance is better than the random samples. It maximizes the performances. For
effective information retrieval three things are necessary they are:
1. Collection of document.
2. Consists of queries.
3. Relevant judgment i.e. Relevant and non-Relevant.[5]
Relevant Document
Number of Documents
Fig 2
Effectiveness can be achieved by recall and precision. In the above diagram, among the number
of documents, on the basis of recall, the fraction of relevant document has been retrieved.
Among the retrieved document, precision can be achieved. The performances can be measured in
four dimensional spaces, i.e.by specifying relevant Vs retrieved.
Retrieved
[1] Fig 3
It can be illustrated by following example: Consider the set of documents F0, F1, F2, F3, F4, F5,
F6, F7, F8, F9 In which query retrieves, the following,
The Relevant document that is obtained is A+B, The not relevant document obtained is C+D,
The Retrieved document that is obtained is A+C, and the unretrieved document obtained is B+D.
A+B = 12,
C+D = 13
A+C = 15,
B+D = 10.
If the recall is increases, i.e. non-relevant document also increases, the precision will be
decreases. [1],[2]
If you’re searching the information on Google search certain information, among them
almost first 11 results are would be relevant and retrieved i.e. true positive results will be
obtained. So precision increases, recall also increases. The ranked retrieval document is done by
interpolated precision, focus on the highest precision i.e. maximum value of relevant document
that is retrieved. So it is rated on the scale, consider the following the table:
Tab 2
In the retrieved document, first document is rated as 1 as per the 11 document are rated. So in
this case at the first precision will be increases at the particular point the precision will be
1.2
0.8
0.6
0.4
0.2
0
1 2 3 4 5 6 7 8 9 10 11
Fig 4
At the certain circumstances, precision decreases, recall will be remains the same. After
11 documents precision will be start falling on. Most of the documents will not be relevant,
hence it is considered as false positive. To determine the percentage of false positive it can be
calculated using the formula
2 PR
P+R
Consider an example, if there are 8 relevant and 10 not relevant document, the precision and
recall can be calculated using the formula precision = true positive / (true positive + false
positive), Recall = true positive / (true positive + false negative).
Cranfield experiment
The cranfield comprises of two steps, first is cranfield 1 and second is cranfield 2. The cranfield
2 is the subset of cranfield 1.
In single term, if the user seeks only one word say desk, so the cranfield 2 will retrieve only the
document regarding to desk.( if the document contains desktop, desktop monitor, and other
information regarding to desktop will be deleted) Cranfield 2 will send it to cranfield 1, so
cranfield 1 gives a better performances and effectiveness.
In Multi term, if the user seeks Multi word say Sachin and Dravid, so the cranfield 2 will retrieve
only the document regarding to Sachin and Dravid.(In these case, if the document contains
Sachin, Shewag and Dravid, cranfield 2 will be rejected) cranfield 2 will send it to cranfield 1, so
cranfield 1 gives a better performances and effectiveness.
In thesaurus based controlling language, if the user seeks for “shopal filters are used for edge
detection”, (say) so the cranfield 2 will split the words as shopal and filters and are and used and
for and edge and detection, the cranfield 2 will support up to 32 splitting words. It will seek the
Cranfield mainly focuses on true positive words, so the recall and precision be
increasing.[1],[2],[8]
STAIRS
SMART
The expansion of SMART is System for Mechanical analysis and retrieval of text. SMART
uses page ranking algorithms. The Google and yahoo uses SMART algorithm. Each and every
metadata, is ranked, the Google and Yahoo uses hits for every metadata the user is viewing upon.
The maximum number of times the user uses is maximum hits; the lowest time is minimum hits,
if the user is not viewing the hits, the document will be deleted.[2],[3]
The expansion of INEX is imitative for the evaluation of XML retrieval, It is mainly used
in cite seer, IEEE Explore, ACM portal. The article is stored by using XML, It has a front matter
and body, the body contains section, the section contains title and paragraph, the abstract and
introduction will be added in the title part, the content will be added in paragraph. The front
matter contains journal title and article title. The INEX mainly focus on relevance and coverage.
The formula for INEX is
The ( Q rel(c ) cov(c )) doenotes how much document had covered and how much document is
relevant. The coverage is achived by 4 ways
TREC
The expanision of TREC Text Retrieval Conference. It is used for searching the journal
articles in google or yahoo. It will link into the web site like citeseer, IEEE exploxer. Ones it
opens it allow the user to provide the login form, after he can view his article.
The TREC uses Routing and Adhoc. Routing is used to search for the newer documents, it will
send the query i.e the seeking article to ad hoc, ad hoc will found out the releated article and
gives it back to the Routing. The Routing will send the related articles to the search engine i.e
Google, Yahoo.[1],[2]
Cognitive paradigm
The cognative paradigm mainly focus on the user evaluation criteria for information, the
identifying factor includes coverage, requirements/need, ease of access, social presure, time
consideration. [1]
Evaluation is very important factor for information retrieval. Evaluation is process not
only focusing on precision and recall but also the set of relevant document and retrieved
document. Evaluating the relevant document and retrieved document is very difficult process.
Research works are going upon for evaluating the retrieved document and relevant document.
While retrieving the document it follows queries, the queering should be done twice, it is based
on average of recall and precision, it give a efficient and flexible results, it also avoids
redundancy. [1]
5. http://nlp.stanford.edu/IR-book/html/htmledition/information-retrieval-system-
evaluation-1.html
6. http://nlp.stanford.edu/IR-book/html/htmledition/evaluation-of-unranked-retrieval-sets-
1.html
7. http://nlp.stanford.edu/IR-book/html/htmledition/evaluation-of-ranked-retrieval-results-
1.html
8. http://blog.codalism.com/?p=845
9. http://www.infoplease.com/ce6/sci/A0825197.html
10. http://nlp.stanford.edu/IR-book/html/htmledition/evaluation-of-xml-retrieval-1.html