Retrieval evaluation measures, Management Information Sys

Retrieval Evaluation Measures 

Objective retrieval evaluation measures have been used from the beginning to assess the retrieval system performance. Many different parameters can in principle be used to measure retrieval performance. Of these measures, the best known, and the most widely used, are: recall (R) and precision (P), reflecting the proportion of relevant items retrieved in answer to a search request, and the proportion of retrieved items that are relevant, respectively. The main assumption behind the use of measures such as recall and precision is that the average user is interested in retrieving large amounts of relevant materials (producing a high recall performance) while at the same time rejecting a large proportion of the extraneous items (producing high precision). These assumptions may not always be satisfied. Nevertheless, recall-precision measurements have formed the basis for the evaluation of the better known studies of both operational and laboratory-type retrieval systems. 

It may be mentioned that recall-precision measurements have not proved universally acceptable. Objections of a theoretical and practical nature have been raised. The most serious questions relate to the fact that recall, in particular, is apparently incompatible .with the utility theoretic approach to retrieval, which forms the basis of a good deal of existing information retrieval theory. Under the utility theoretical perspective, retrieval effectiveness is measured by determining the utility to the users of the documents retrieved in answer to a user query. The problem is that a document may be relevant to a query while nevertheless proving useless for a variety of reasons. The system utility might be optimised in some cases by bringing a single relevant document to the user's attention, at which point the recall might be very low. Hence, recall and relevance are quite different notions. 

The most important factor in determining how well a system is working is the relevance of the items retrieved from the system in response to a query from the user. Relevancy of an item, however, is not a binary evaluation, but a continuous function between the items being exactly what is being looked for and is being totally unrelated. To discuss relevance, it is necessary to define the context under which the concept is used. From a human judgement stand point, relevancy can be considered: 

  1. Subjective, i.e., depends upon a specific user's judgement. 
  2. Situational, i.e., relates to a user's requirements. 
  3. Cognitive, i,e., depends on human perception and behaviour. 
  4. Temporal, i.e., changes over time. 
  5. Measurable, i.e., observable at points in time, 

The subjective nature of relevance judgements has been documented by Saracevic and was shown in TREC experiments. (Saracevic, 1995). In a dynamic environment each user has his own understanding.of the requirement and the threshold on what is acceptable. Based upon his cognitive model of the information and the problem, the user judges a particular item to be relevant or not. Some users consider the information they already know to he non-relevant to their information need. Also, judgement of relevance can vary over time. Thus, relevance judgement is measurable at a point in time constrained by the particular users and their thresholds on acceptability of information. 

Another method of specifying relevance is from information, system and situational views. Here again, the information view is subjective in nature and pertains to human judgement of the conceptual relatedness between an item and the search. It involves the user's personal 'judgement of the relevancy of the item to the user's information need. When information professionals assist the user, it is assumed that they can reasonably predict whether certain information will satisfy the user's needs. Ingwersen (1992) categorises the informations view into four categories of 'aboutness': 

  1. 'Author Aboutness'- determined by the author's language as matched by the system in natural language retrieval. 
  2. 'Index Aboutness'- determined by the indexer's transformation of the author's natural language into a controlled vocabulary. 'Request Aboutness'- determined by the user's or intermediary's processing of a search statement into a query. 
  3. 'User Aboutness'- determined by the indexer's attempt to represent the document according to presupposition about what the user will want to know. 

In this context, the system view relates to a match between query terms and terms within an item. It can be objectively observed, tested without relying on human judgement. On the other hand, the situation view pertains to relationship between information and the user's information problem situation. It assumes that only users can make valid judgements regarding the suitability of information to solve their information need. Lancaster and Warner refer to information and situation views as relevance and pertinence respectively. Pertinence can be defined as those items that satisfy the user's information need at the time of retrieval. Pertinence depends on each situation. 

It may be mentioned here that evaluation of IR Systems is, essential to understand the source of weakness in existing systems and to improve their effectiveness. The standard measures of Precision, Recall and Relevance have been used for the last 25 years as the major measures of algorithmic effectiveness. 

The major problem associated with evaluation of IR Systems is the subjective nature of the information. There is no deterministic methodology for understanding what is relevant to a user's search. Users have trouble in translating their mental perception of information being sought into the written language of a search statement. When fe. is are needed users are able to provide a specific relevance judgement on an item. But when general information is needed relevancy goes from a classification process to a continuous function. They are not able to easily distinguish relevant items from non-relevant ones. 

The Text Retrieval Evaluation Conferences (TRCSs) provide yearly forums where developers of algorithms can share their techniques with their peers and contribute theories of evaluation, which may lead to the design and development of effective and efficient ER systems in future.  

Posted Date: 10/24/2012 3:28:44 AM | Location : United States

Related Discussions:- Retrieval evaluation measures, Assignment Help, Ask Question on Retrieval evaluation measures, Get Answer, Expert's Help, Retrieval evaluation measures Discussions

Write discussion on Retrieval evaluation measures
Your posts are moderated
Related Questions
Question : a) What generic strategy or business model should Super Foods Company Ltd. adopt in order to move to an on online grocery? Justify your answer. b) How does the b

Question : a) What is the relationship between management information systems and transaction processing systems ? b) What are the advantages of web-based interactive mar

Problem: (i) ‘Many Information Systems projects in the public sector have failed due to poor change management procedures'. Discuss this statement with examples. (ii) An I

can someone contact me at

Question: a) "Many organisations have been using the Internet, Intranet and Extranet to realise strategic gains in their competitive position. To achieve and maintain such a

INTRANETS   In this section, we shall try to know what Intranets are and how they are beneficial. In simple terms, an Intranet is an internal communication mechanism of an orga

KWIC (Key Word in Context) Index   A KWIC index is the most common type of natural language indexes. It has become popular because it is straight-forward and relatively easier

Consider a traditional manufacturing company that wanted to build a social business strategy. What might be a reasonable business strategy, and how would organizational and IS stra

Case Study: Viruses Robert Morris is the son of a well respected computer expert. Robert inserted a virus in a computer network that impacted negatively on over 6000 users. He

Question 1: A bank is in the process of reviewing its IT Systems. It currently hosts a number of legacy systems which supports different aspects of its business functions namel