Results 1 
8 of
8
Some inconsistencies and misidentified modelling assumptions in probabilistic information retrieval
 A CM Transactions on Information Systems
, 1995
"... Research in the probabilistic theory of information retrieval involves the construction of mathematical models based on statistical assumptions. One of the hazards inherent in this kind of theory construction is that the assumptions laid down maybe inconsmtent in unanticipated ways with the data to ..."
Abstract

Cited by 27 (0 self)
 Add to MetaCart
Research in the probabilistic theory of information retrieval involves the construction of mathematical models based on statistical assumptions. One of the hazards inherent in this kind of theory construction is that the assumptions laid down maybe inconsmtent in unanticipated ways with the data to which they are applied. Another hazard is that the stated assumptions may not be those on which the derived modeling equations or resulting experiments are actually based. Both kinds of mistakes have been made m past research on probabihstic reformation retrieval. One consequence of these errors is that the statistical character of certain probabilistic IR models, including the socalled Binary Independence model, has been seriously misapprehended Categories and Subject Descriptors: H. 1.2 [Models and Principles]: User/Machine Systems;
Term Dependence: Truncating the Bahadur Lazarsfeld Expansion
 Information Processing and Management
, 1994
"... The performance of probabilistic information retrieval systems is studied where differing statistical dependence assumptions are used when estimating the probabilities inherent in the retrieval model. Experimental results using the Bahadur Lazarsfeld expansion suggest that the greatest degree of ..."
Abstract

Cited by 16 (7 self)
 Add to MetaCart
The performance of probabilistic information retrieval systems is studied where differing statistical dependence assumptions are used when estimating the probabilities inherent in the retrieval model. Experimental results using the Bahadur Lazarsfeld expansion suggest that the greatest degree of performance increase is achieved by incorporating term dependence information in estimating . It is suggested that incorporating dependence in to degree 3 be used; incorporating more dependence information results in relatively little increase in performance. Experiments examine the span of dependence in natural language text, the window of terms in which dependencies are computed and their effect on information retrieval performance. Results provide additional support for the notion of a window of to terms in width; terms in this window may be most useful when computing dependence. 2 1 Introduction Those who study information retrieval often assume that the features or terms use...
The Maximum Entropy Approach and Probabilistic IR Models
 ACM TRANSACTIONS ON INFORMATION SYSTEMS
, 1998
"... The Principle of Maximum Entropy is discussed and two classic probabilistic models of information retrieval, the Binary Independence Model of Robertson and Sparck Jones and the Combination Match Model of Croft and Harper are derived using the maximum entropy approach. The assumptions on which the cl ..."
Abstract

Cited by 12 (0 self)
 Add to MetaCart
The Principle of Maximum Entropy is discussed and two classic probabilistic models of information retrieval, the Binary Independence Model of Robertson and Sparck Jones and the Combination Match Model of Croft and Harper are derived using the maximum entropy approach. The assumptions on which the classical models are based are not made. In their place, the probability distribution of maximum entropy consistent with a set of constraints is determined. It is argued that this subjectivist approach is more philosophically coherent than the frequentist conceptualization of probability that is often assumed as the basis of probabilistic modeling and that this philosophical stance has important practical consequences with respect to the realization of information retrieval research.
Monitoring UserSystem Performance in Interactive Retrieval Tasks
 PROC. RIAO 2004
, 2004
"... Monitoring usersystem performance in interactive search is a challenging task. Traditional measures of retrieval evaluation, based on recall and precision, are not of any use in real time, for they require a priori knowledge of relevant documents. This paper shows how a Shannon entropybased measur ..."
Abstract

Cited by 4 (1 self)
 Add to MetaCart
Monitoring usersystem performance in interactive search is a challenging task. Traditional measures of retrieval evaluation, based on recall and precision, are not of any use in real time, for they require a priori knowledge of relevant documents. This paper shows how a Shannon entropybased measure of usersystem performance naturally falls in the framework of (interactive) probabilistic information retrieval. The value of entropy of the distribution of probability of relevance associated with the documents in the collection can be used to monitor search progress in live testing, to allow for example the system to select an optimal combination of search strategies. User profiling and tuning parameters of retrieval systems are other important applications.
Probabilistic Information Retrieval Model for Dependency Structured Indexing System
 In Proceedings of the ACM SIGIRâ€™02 Workshop on Mathematical/Formal Methods in Information Retrieval, 2002. Proceedings of the Third NTCIR Workshop
, 2002
"... statistically independent from each another. However, independence assumption is obviously and openly understood to be wrong, so we present a new method of incorporating term dependence in probabilistic retrieval model by adapting a structural index system using dependency parse tree and the Chow Ex ..."
Abstract

Cited by 2 (1 self)
 Add to MetaCart
statistically independent from each another. However, independence assumption is obviously and openly understood to be wrong, so we present a new method of incorporating term dependence in probabilistic retrieval model by adapting a structural index system using dependency parse tree and the Chow Expansion to compensate the weakness of the assumption. In this paper, we describe a theoretic process to apply the Chow Expansion to the general probabilistic models and the stateoftheart 2Poisson model, and we reexamine the weight of phrase terms. Through the experiments on document collections, ETRIKEMONG in Korean, we demonstrate that the incorporation of term dependences using the Chow Expansion contribute to the improvement of performance in Probabilistic IR systems. Keywords term dependence, phrasal indexing, Chow Expansion, probabilistic model, 2Poisson model 1.
A Model for the Stopping Behavior of Users of Online Systems
, 1987
"... We examine a model in which the user of an online system continually updates his/her estimated probability of success, and quits or continues according to the expected utility of each action. The prior distribution of the unknown probability is a beta distribution, with mean determined by the a prio ..."
Abstract

Cited by 2 (1 self)
 Add to MetaCart
We examine a model in which the user of an online system continually updates his/her estimated probability of success, and quits or continues according to the expected utility of each action. The prior distribution of the unknown probability is a beta distribution, with mean determined by the a priori expectation of success, and variance determined by the confidence with which the user has that prior expectation. The stopping criterion depends upon the accumulated number of positive and negative reinforcements, and is a straight line in a suitable coordinate system.
Distance, Minimum CrossEntropy, and Path methods. Background and Purpose of the Study
, 1988
"... The maximum entropy principle may be applied to the design of probabilistic retrieval systems. When there are inconsistent expert judgments, the resulting optimization problem cannot be solved. The inconsistency of the expert judgments can be revealed by solving a linear programming formulation. In ..."
Abstract
 Add to MetaCart
The maximum entropy principle may be applied to the design of probabilistic retrieval systems. When there are inconsistent expert judgments, the resulting optimization problem cannot be solved. The inconsistency of the expert judgments can be revealed by solving a linear programming formulation. In the case of inconsistent judgment, four plausible schemes are proposed in order to find revised judgments which are consistent with the true data structure but still reflect the original expert judgment. These schemes are the Interactive, Minimum