Probability Kinematics in Information. Retrieval: a case study. F. Crestani. Universita' di Padova, Italy. C.J. van Rijsbergen.

Size: px
Start display at page:

Download "Probability Kinematics in Information. Retrieval: a case study. F. Crestani. Universita' di Padova, Italy. C.J. van Rijsbergen."

Transcription

1 Probability Kinematics in Information Retrieval: a case study F. Crestani Dipartimento di Elettronica e Informatica Universita' di Padova, Italy C.J. van Rijsbergen Department of Computing Science University of Glasgow, Scotland Abstract In this paper we discuss the dynamics of probabilistic term weights in dierent IR retrieval models. We present four dierent models based on dierent notions of retrieval. Two of these models are classical probabilistic models long in use in IR, the two others are based on a logical technique of evaluating the probability of a conditional called Imaging, one is a generalisation of the other. We analyse the transfer of probabilities occuring in the representation space at retrieval time for these four models, compare their retrieval performance using classical test collections, and discuss the results. 1

2 Contents 1 Introduction 3 2 The representation space Possible World Semantics : : : : : : : : : : : : : : : : : : : : : : The term space : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6 3 Probability kinematics for retrieval models Retrieval by Joint Probability : : : : : : : : : : : : : : : : : : : : Retrieval by Conditional Probability : : : : : : : : : : : : : : : : Retrieval by Logical Imaging : : : : : : : : : : : : : : : : : : : : : Retrieval by General Logical Imaging : : : : : : : : : : : : : : : : 13 4 Probability, similarity and opinionated functions 17 5 A case study 19 6 Experiments with test collections 22 7 Comparative evaluation 24 8 Conclusions 26 A A case study of probability dynamics 29 A.1 Retrieval by Joint Probability : : : : : : : : : : : : : : : : : : : : 30 A.2 Retrieval by Conditional Probability : : : : : : : : : : : : : : : : 31 A.3 Retrieval by Logical Imaging : : : : : : : : : : : : : : : : : : : : : 32 A.4 Retrieval by General Logical Imaging : : : : : : : : : : : : : : : : 33 2

3 1 Introduction In Information Retrieval (IR), probabilistic modelling means the use of a model to rank documents in decreasing order of their evaluated or, more often, estimated probability of relevance to a user's information need that is expressed by means of a query. In an IR system based on a probabilistic model, the user is always guided to examine rst the documents which are the most likely to be relevant to their need, that is those at the top of the ranked list. The use of a probabilistic model in IR assures that we can obtain \optimal retrieval performance" if we rank documents according to their probability to be judged relevant to a query [15]. However, this rule, called The Probability Ranking Principle, refers only to \optimal retrieval", which is dierent from \perfect retrieval". Optimal retrieval can be dened precisely for probabilistic IR because optimality can be proved theoretically, owing to a provable relationship between ranking and the probabilistic interpretation of precision and recall. Perfect retrieval relates to the objects of the IR systems themselves, i.e. documents and information needs, but as IR systems use representations of these objects, perfect retrieval is not an appropriate goal for computer-based systems. Despite that, probabilistic models based on the Probability Ranking Principle have been shown to give the highest levels of retrieval eectiveness currently available. Although there are a few operative IR systems based on probabilistic or semiprobabilistic models, there are still obstacles to getting many probabilistic models accepted in the commercial IR world. One major obstacle is that of nding methods for estimating the probabilities of relevance that are both eective and computationally ecient. Past and present research has made much use of formal probability theory and statistics in order to solve the problems of estimation. In mathematical terms the problem consists in estimating the probability P (R j q; d), i.e. the probability of relevance given a query q and a document d for every document in the collection, and ranking the documents according to this measure. This is very dicult because of the large number of variables involved in the representation of documents in comparison with the small amount of feedback data available about the relevance of documents (sometimes refered to as the \curse of dimensionality"). In 1986 Van Rijsbergen [22] proposed for use in IR techniques based on nonclassical conditional logic. This would enable the estimation of P (R j q; d) by the evaluation of P (d! q) instead, so proposing to use the probability of a conditional to estimate the conditional probability. The evaluation of the probability P (d! q) should follow the following logical uncertainty principle: \Given any two sentences x and y; a measure of the uncertainty of 3

4 y! x related to a given data set is determined by the minimal extent to which wehave to add information to the data set, to establish the truth of y! x." That proposal initiated a new line of research (see for example [12, 13, 3, 2]) but in that paper nothing was said about how \uncertainty" and \minimal" might be quantied. A few years later, moving into Modal Logic, Van Rijsbergen proposed to estimate the probability of the conditional by a process called Logical Imaging [23]. This techniques has been explored in more details later in [5]. In this paper we explore further the use of the probability of a conditional, namely P (d! q), to estimate the conditional probability P (R j q; d). We propose the use in IR of a technique called General Logical Imaging (or simply \General Imaging"), proposed by Gardenfors [7] in the context of the Belief Revision theory. This techniques is a generalisation of Imaging [10] that enables a more general redistribution of probabilities than Imaging. We analyse and compare the probability kinematics of Imaging and General Imaging with more classical probabilistic models that have been the basis of many IR probabilistic models: the Joint Probability model and the Conditional Probability model. The paper is structured as follows. Section 2 describes the model that will be used to represent documents and queries in the rest of the paper. Section 3 describes the probability kinematics of four dierent retrieval models. An example of the dierent results that can be achieved using these four models is given in Section 5. The retrieval performance of these models are evaluated using the probability and similarity functions described in Section 4 on some standard test collections described in Section 6. The results are presented, compared, and discussed in Section 7. Section 8 ends the paper giving the conclusions of the experimental investigation. 2 The representation space In probabilistic IR the task of the system can be formalised as follows. If we assume binary relevance judgements, i.e. R the set of possible relevance judgements contains only the two possible judgements: relevant (R) and not-relevant (R), then according to The Probability Ranking Principle, the task of the system is to rank the documents according to their probability of being relevant P (R j q;d), where q and d are the real query and the real document. Of course we can only estimate this probability by using the available query and documents representations, q and d. The probability P (R j q; d), is the Retrieval Status Value (RSV) that will be used to rank documents. 4

5 The diculty of applying probabilistic IR lies therefore in two dierent problems: estimation. representation; The problem of estimating P (R j q; d) is tackled in this paper from a theoretical point of view. In the past many researchers have tried to estimate P (R j q; d) in many ways. It is in fact from these attempts that many probabilistic models for IR have sprung. In Section 3 we will report on four models, we will analyse their dierences, and we will draw some interesting conclusions. Representing documents and queries is very dicult to tackle. We will ignore the problems related to representation in this paper. We will use a representation model that helps us in the analysis of the probabilistic models presented in Section 3. This representation model is based on a particular semantics: the \Possible World Semantics" [11]. 2.1 Possible World Semantics Possible World Semantics (PWS) was introduced by Kripke [9] in the context of Modal Logic. In this semantics the truth value of a logical sentence is evaluated in the context of a world. The word \world" has been used by a number of logicians in this connection, and seems to be the most convenient one, but perhaps some such phrase as \conceivable or envisageable state of aairs" ([8], p. 75) would convey the idea more clearly. PWS has been used in Modal systems to give a semantics for Necessity (a sentence is true in every possible world) and Possibility (a sentence is true in at least one possible world) 1. Without entering into the details of this semantics, we would like to point out the main reasons why we use it. We use PWS because it enables the evaluation of a conditional sentence without explicitly dening the operator \!". What it requires is a clustering on the space of events (worlds) by means of a primitive relation of neighbourhood. This clustering enables us to dene an accessibility relation that is necessary for the evaluation of the conditional. According to the PWS the truth value of the conditional y! x inaworld w is equivalent to the truth value of the consequent x in the closest world w y to w where the antecedent y is true [19]. In Section 3.3 we will explain how we can use this result in the context of IR, but rst let us examine how we can use the PWS to model our probabilistic retrieval space. 1 Here we simply refer to the Modal System S5 and not to more complex models. 5

6 2.2 The term space One of the most often used IR models is the Vector Space Model (VSM) [17]. In this model a document is represented by means of a vector whose elements are numbers representing the presence/absence of certain features in that document like for example, the presence or absence of some index terms 2. The document representation space is therefore multidimensional, with as many dimensions as the number of features used to represent documents. A document is represented in this space as a single point. A query is also a point, and document retrieval is performed as a function of the distance in the representation space between query and documents. The semantics of the VSM is therefore that of a multidimensional geometrical space. Many IR models use this representation space as the underlying space. In this paper we use a dierent approach. The semantics of our representation space is based on the PWS. We can use the PWS in the context of IR by considering a term as a possible world. This view was proposed in [5]. According to this view, a term is represented as a \vector of documents". This is the inverse of the representation model used in the VSM. Intuitively this can be understood as \if you want toknow the meaning of a term then look at all the documents in which that term occurs". This idea is not new in IR (see for example [1, 14]) and it has been widely used for the evaluation of term{term similarity (see Section 4). More formally,we assume a set of terms T, the set of our possible worlds. We also assume a \prior" probability distribution P assigning to each term (world) t 2 T a probability P (t) so that P P (t) =1. We then assume we have a document collection D. The document we use are represented using terms in T,as it is common in IR, but in our representation space we use documents to represent terms, in such a way that a term t is represented by a binary vector whose elements denote the presence or absence of documents in the term representation. Finally, we assume we have a query q also represented using terms in T. 3 Probability kinematics for retrieval models In the following sections we examine the dierentmovements of probability transfers that takes place in four dierent retrieval models. Our purpose is to show how the probability associated with terms changes and shifts in dierent ways in dierent models. We do not intend to associate directly any of the models here with existing IR models. However, these four models can be considered to 2 For simplicity of exposition we only consider the binary case. 6

7 be the archetypes of the most common IR models. The probability kinematics of the rst two models shows how VSM and probabilistic retrieval model may be explained in terms of the framework adopted for probability transfer. The last two models are new and are based on a completely dierent approach for transfer of probability. Their origin lies in the eld of non-classical logics. We show that in principle, i.e. without entering into complex \ad hoc" weighting and retrieval schemas, these last two models perform better than the rst two. This result suggests that an improvements in retrieval eectiveness can be obtained by designing IR systems based on probabilistic models that use a completely new kind of probability kinematics. We explain the changes in the probability space by using the representation model described in Section 2, and taking as an example a particular document d and a query q. We suppose we have a document d represented by terms t 1, t 5, and t 6 and a query q represented by the t 1, t 4, and t 6. Each of these terms has a \prior" probability associated with it, this is indicated by P (t). In the following we show how the RSV of document d is evaluated by dierent retrieval models and we concentrate our attention on how the probabilities associated with terms change and move from term to term during the evaluation of the RSV. We will indicate the new \posterior" probability associated with terms by P d (t) to highlight the fact that it is obtained by taking into consideration a particular document d. In the following sections we make extensive use of the following function: I(w; y) =( 1 if w occurs in y 0 otherwise We will use this function to evaluate to presence or absence of a representation element w in the representation of y. In the context of the PWS, assuming that y is a logical sentence and w is a possible world, we can also interpret I(w; y) as follows: I(w; y) =( 1 if y is true at w 0 otherwise 3.1 Retrieval by Joint Probability We call Retrieval by Joint Probability (RbJP) the ranking and retrieval of documents obtained estimating the probability of relevance with the probability of the joint event consisting of having both the query and the document true for a term. P (R j q; d) P(q; d) 7

8 t P (t) I(t; d) I(t; q) P (t) I(t; d) I(t; q) Table 1: Evaluation of P (q; d) RbJP can be obtained by evaluating the RSV of a document in accordance with the following formula: P (q; d) = X t P(t)I(t; d) I(t; q) where, for a document d, we compute the sum of the probabilities of all terms that are both present in that document and in the query. In the PWS this is equivalent to the sum of the probabilities of the worlds in which both the document and the query are true. In RbJP model there is no transfer of probabilities. The \prior" probability P (t) associated with term t does not change, it remains the same. That is why we do not need to evaluate the \posterior" probability P d (t). It is important to note that the RbJP model is the archetype of many IR models currently in use. Most IR models that are based on the evaluation of a similarity between documents and query are based upon the idea of a joint probability measure. Both Dice's and Jaccard's coecients, for example, are based on this idea, as it can be easily seen once we remove the normalisation factors (see [21], p. 39). The Cosine Correlation used by the the Vector Space Model is again a normalised version RbJP. Of course, we do not intend to undervalue the importance of normalisation factors, but we just want to point out that the probability kinematics of all these IR models does not change once a normalisation factor is introduced, it substantially remains the same as that for RbJP. To show howweevaluate the RSV in the case of RbJP, we show an example in Table 1, where we report the evaluation of P (q; d). The evaluation process is the following: 1. Identify the terms occurring in the document d (third column of the table). 2. Identify the terms occurring in the query q (fourth column). 8

9 p1 t1 t2 p2 p1 t1 t2 p2 p3 t3 t4 p4 p3 t3 q p4 t4 p5 t5 d t6 p6 p5 t5 t6 p6 a b Figure 1: Graphical interpretation of the evaluation of P (q; d). 3. Evaluate the P (d; q) by summing the probability of all terms present in both document and query (fth column). A graphical interpretation of RbJP using the PWS is given in Figure 1, where each term is represented by a world with its \prior" probability measure expressing the importance of the term in the term space T. The shadowed terms are those occurring in document d (Figure 1(a)), P (q; d) is obtained by summing the probability of all terms occurring in the document and in the query representations, that is summing the probabilities of the shadowed terms also occurring in q (Figure 1(b)). 3.2 Retrieval by Conditional Probability In the case of Retrieval by Conditional Probability (RbCP) we estimate the probability of relevance of a document by evaluating the conditional probability of the query given that document. P (R j q; d) P(q j d) P(q j d) can be evaluated as follows: P (q j d) = P d (q) = P t P d (t) I(t; q) = P t P (t)(1 + d ) I(t; q) where P d (t) is the \posterior" probability distribution over the set of terms occurring in d obtained by conditioning on the document d itself, and d is the ratio 9

10 t P (t) I(t; d) P d (t) I(t; q) P d (t) I(t; q) Table 2: Evaluation of P (q j d) by which the \prior" probability is modied. The value d is the ration between the sum of the probabilities of the terms not occurring in d and the sum of the probabilities of those occurring in d: d = 62d P (t) 2d P (t) It should be noticed that the transfer of probabilities that takes place in this model provides the minimal revision of the \prior" probability necessary to make d certain that does not distort the prole of probability ratios. In fact the \posterior" probability is directly proportional to the \prior" probability, so leaving constant the ratios among probabilities associated with the terms after the contraction of the representation space due to the certainty of the conditional event d. Let us now see an example of this evaluation using query q and document d. Table 2 reports the evaluation of P (q j d). The evaluation process is the following: 1. Identify the terms occurring in the document d (third column of the table). 2. Evaluate the \posterior" probability P d (t) by transferring the probabilities from terms not occurring in the document to terms occurring in it. The probabilities are transfered in a proportional way, so that each term occurring in the document d receives a portion of the sum of the probability of the terms not occuring in the document proportional to its \prior"probability (fourth column). 3. Evaluate I(t; q) for each term, i.e. determine the terms occurring in the query (fth column). 4. Evaluate P d (t) I(t; q) for all terms (sixth column) and evaluate P d (q) by summation (bottom of sixth column). 10

11 p1 t1 t2 p2 p1 t1 t2 p2 p 1 t1 p3 t3 t4 p4 p3 t3 p4 t4 q p5 t5 d t6 p6 p5 p6 p 5 p 6 t5 t6 t5 t6 a b c Figure 2: Graphical interpretation of the evaluation of P (q j d). It is interesting to see a graphical interpretation of this process. In Figure 2(a) each term is represented by a world with its \prior" probability expressing the importance of the term in the term space. The shadowed terms occur in document d. The conditioning process transfers the probability from terms not occurring in the document d to those occurring in it as depicted in Figure 2(b). In Figure 2(c) the terms with null probability disappear and those occurring in the query q are taken into consideration. The \posterior" probabilities P d (t) of terms occurring in the query are summed to evaluate P (q j d). 3.3 Retrieval by Logical Imaging Imaging is a process developed in the framework of Modal Logic that enables the evaluation of a conditional sentence without explicitly dening the operator \!" [19]. Imaging has been extended to the case where there is a probability distribution on the worlds by Lewis [10]. In this case the evaluation of P (y! x) causes a shift of the original probability P from a world w to the closest world w y where y is true. Probability is neither created nor destroyed, it is simply moved from a \not-y-world" to a \y-world" to derive a new probability distribution P y. This process is called \deriving P y from P by imaging on y". We will not go into a detailed explanation of the Imaging process. The interested reader can look at papers by Stalnaker [19], Lewis [10], and Gardenfors [7] for more details. We use Imaging in IR with the purpose of estimating the probability of relevance of a document by means of the probability of the conditional d! q: P (R j q; d) P(d! q) We call Retrieval by Logical Imaging (RbLI) the model based on such RSV. A detailed explanation of this model can be found in [5]. Briey, in RbLI we consider 11

12 t P (t) I(t; d) t d P d (t) I(t; q) P d (t) I(t; q) Table 3: Evaluation of P (d! q) by imaging on d the process of Imaging on d over all the possible terms t in T. More formally: P (d! q) = P d (q) = P t P d (t) I(t; q) = P t P (t) I(t d ;q) where t d is the closest term to t for which d is true, or in other words, the most similar term to t that occurs in the document d. The application of the above technique to IR requires an appropriate measure of similarityover the term space T to enable the identication of t d.we will tackle this problem in Section 4. It should be noticed that Imaging provides the minimal revision of the \prior" probability in the sense that it involves no gratuitous movement of probability from world to dissimilar worlds. In fact, the revision of the \prior" probability necessary to make d certain is obtained by adopting the least drastic change in the probability space. This is achieved by transfering probabilities from each term not occuring in the document d to its closest (most similar) term occurring in it, so that the total amount of the distance covered in the transfer is minimal. A detailed comparison between Conditionalisation and Imaging can be found in [6]. For a practical example of the evaluation of RbLI let us suppose wehave the same query q and document d of the previous sections. Table 3 reports the evaluation of P (d! q) by imaging on d. The evaluation process is the following: 1. Identify the terms occurring in the document d (third column of the table). 2. Determine for each term in T the t d, i.e. the most similar term to t for which I(t; d) =1. This is done using a similarity measure on the term space (fourth column). 3. Evaluate P d (t) by transferring the probabilities from terms not occurring in the document to terms occurring in it (fth column). 12

13 p1 t1 t2 p2 p1 t1 p2 t2 p 1 t1 p3 t3 p5 t5 d t4 t6 p6 p4 t3 p3 p5 t5 p4 t4 t6 p6 q p 5 p 6 t5 t6 a b c Figure 3: Graphical interpretation of the evaluation of P (d! q) by imaging on d. 4. Evaluate I(t; q) for each term, i.e. identify the terms occurring in the query (sixth column). 5. Evaluate P d I(t; q) for all terms (seventh column) and evaluate P d (q) by summation (bottom of seventh column). A graphical interpretation of this process is depicted in Figure 3. As we said, we assume wehave a measure of similarity on the term space, using it we can transfer probability from each term not occurring in the document d to its most similar one occurring in d. After the transfer of probability terms with null probability disappear and those occurring in the query q are taken into consideration, so that their \posterior" probabilities P d (t) can be summed to evaluate P d (q). 3.4 Retrieval by General Logical Imaging The idea of General Imaging originated from an attempt to overcome one of the restrictive assumptions Lewis made for Stalnaker's semantics of conditionals [10]. The assumption is related to the \uniqueness" of the world w y, that is the uniqueness of the world most similar to w where y is true. In [7] p. 110, Gardenfors propose a generalisation of the Imaging process that does not rely on this assumption 3. The starting point of the generalisation is the use of a (degenerate) probability function to represent the fact that in any possible world w a proposition y is either true or false. Using our formalism I(w; y) =P w (y). In this case P w (y) =1if yis true in w, and P w (y) =0if yis false in w. Lewis called such probability function opinionated because \it would represent 3 In that book he also characterised this generalisation of Imaging in terms of a homomorphic condition that does not presuppose any kind of possible world semantics, but we will remain faithful to our semantics in the rest of this paper. 13

14 the beliefs of someone who was absolutely certain that the world w was actual and who therefore held a rm opinion about every question" (see [10], p. 145). Suppose we have a set of possible worlds W. We can generalise Imaging by considering the fact that, instead of having P w (y) = 1 only for a single w y,we can have P w (y) > 0 for a few worlds. More formally: P w (y) =( 0 if y is false at w > 0 otherwise with the requirement that P w P w (y) = 1. We then have: P w (y! x) = X w P w y(x) Now we assume a probability distribution over the set of possible worlds W (the \prior" probability) so that, according to the classical rules of probability, we have P w P (w) = 1. Hence we dene P (y) as follows: P (y) = X w P(w)P w (y) From this probability distribution we can derive a new probability distribution P 0 so that: P 0 (w 0 )= X w P(w)P w (w 0 )I(w 0 ;w) where: I(w 0 ;w)=( 1 if w 0 belongs to the set W y of w 0 otherwise with W y as the set of the closest worlds to w where y is true. It could be proved, with a demonstration similar to the one reported in [10], p. 142, that P (y! x) =P 0 (x) or, using a terminology more appropriate to highlight the Imaging process on y, that: P (y! x) =P y (x) 14

15 t P (t) I(t; d) t d P d (t) I(t; q) P d (t) I(t; q) ; ; ; Table 4: Evaluation of P (d! q) by general imaging on d where P y is the \posterior" probability distribution derived from the \prior" probability P by General Imaging on y. In other words, P y can be obtained by transfering the probability from every world w to the W y, which in this case is the set of most similar (closest) worlds to w where y is true. The transfer of probability is performed according to the opinionated probability function P w. Lewis Imaging, as reported in Section 3.3, is a special case of General Imaging when P w (y) = 1 for some w. Retrieval by General Logical Imaging (RbGLI) can be regarded a the process of applying General Imaging on d in order to evaluate the RSV of the document d. More formally: P (d! q) = P d (q) = P t P d (t) I(t; q) = P t P (t) ( P t 0P t0 (d)) I(t d ;q) = P t ( P t 00 (t) P(t 0 )) I(t d ;q) = P t;t 0 P t0 (t) P (t 0 ) I(t d ;q) where P t0 (d) is the opinionated probabilityofdin t 0, and P t0 (t) is the opinionated probability of term t in t 0 which is necessary to evaluate the former, and I(t d ;q) is as dened in Section 3.3. The application of the above technique to IR requires again an appropriate measure of similarity over the term space T to enable the identication of the set of closest terms to t where d is true. We will tackle this problem in Section 4. Table 4 reports an example of the evaluation of P (d! q) by general imaging on d. The evaluation process is the following: 1. Identify the terms occurring in the document d (third column of the table). 15

16 p1 t1 t2 p2 p1 t1 t2 p2 p 1 t1 p3 t3 t4 p4 p3 t3 p4 t4 q p5 t5 d t6 p6 p5 p6 p 5 p 6 t5 t6 t5 t6 a b c Figure 4: Graphical interpretation of the evaluation of P (d! q) by general imaging on d. 2. Determine for each term not occuring in the document (with I(t; d) =0) the most similar terms (in this example only two terms) occuring in the document (those with I(t; d) = 1). This is done using a similarity measure on the term space (fourth column). 3. Evaluate P d (t) by transferring the probabilities from terms not occurring in the document to terms occurring in it (fth column) using an opinionated probability function, also called transfer function. In this example the transfer function prescribes that the most similar term to the one under consideration gets 2=3 of its probability, while the second most similar gets the remaining 1=3. 4. Evaluate I(t; q) for each term, i.e. determine the terms occurring in the query (sixth column). 5. Evaluate the probabilities P d I(t; q) for all the terms (seventh column) and evaluate P d (q) by summation (bottom of seventh column). A graphical interpretation of this process is depicted in Figure 4. As it can be seen in the picture, in the case of RbGLI the transfer of probability is performed from each term not occurring in the document d to the k t most similar terms occurring in d. In the above example k t = 2 for every term, but k t can be set to any other integer number as long as 1 k t l t, where l t is the number of documents in which the term t occurs. The value of k t can theoretically be dierent for every term. For k t = 1 for every term RbGLI defaults to RbLI. For k t = l t for every term the transfer may seem similar to RbCP, but it must be noted that the probability transfer in RbGLI is performed by taking into account the similaritybetween terms and not the ratio \prior" probabilities as in the case of RbCP. 16

17 4 Probability, similarity and opinionated functions In order to perform an experimental investigation into the previous four retrieval model we have three requirements: 1. a \prior" probability distribution over the set of worlds which should reect the importance of each world (the mass if we take an analogy with planets and stars) in the universe; 2. a measure of similarity (which is related to distance) between worlds; 3. an opinionated probability function over the set of possible worlds for every document in the collection. Notice that the rst requirement is present in everyone of the four models, the second one for RbLI and RbGLI, and the third only for RbGLI. According to our view in which a term is a world, these two requirements become: a probability distribution, a measure of similarity, and a set of opinionated functions on the term space T. The problem of determining an appropriate \prior" probability distribution over the set of terms used to index a document collection is one of the oldest in IR and many models have been proposed for this purpose. The problem could be translated into nding a measure of the importance of the term in the term space, where this importance is related to the ability of the term to discriminate between relevant and not relevant documents. In IR several discrimination measures have been proposed (see for example [21, 16]). For the experiments reported in this paper we used the Inverse Document Frequency (IDF), a measure which assigns high discrimination power to terms with low and medium collection frequency. IDF is dened as: IDF(t)= log n N where n is the number of documents in which t occurs, and N is the number of documents in the collection. Strictly speaking, this is not a probability measure since P t IDF(t) 6=1,however we can assume it to be monotone to P (t). We can use this estimate because we require only a ranking of the documents, the exact probabilityvalues are not used. We also chose this measure because it does not require relevance information. At 17

18 this stage of our work we prefer not to require relevance information which should come from Relevance Feedback with the user. The problem of measuring the similarity between terms and its use to dene the accessibility among worlds is more dicult. It is very important to chose the appropriate measure since much of RbLI and RbGLI depends on it. In this paper we decided to use the Expected Mutual Information Measure (EMIM) between terms as the measure of accessibility between worlds. The EMIM between two terms is often interpreted as a measure of the statistical information contained in the rst term about the other one (or vice versa, it being a symmetric measure). EMIM is dened as follows: I(i; j) = X t i ;t j P (t i ;t j ) log P(t i;t j ) P(t i )P(t j ) where i and j are binary variables representing terms. When we apply this measure to binary variables we can estimate EMIM between two terms using the technique proposed in [20]. This technique makes use of co-occurrence data which can be simply derived by a statistical analysis of the term occurrences in the collection. Using this measure we can then evaluate for every term a ranking of all the other terms according to their decreasing level of similarity with it. We store this information in a le which is used at run-time to determine for a term its nearest neighbour occurring in the document under consideration. The denition of a possibly dierent opinionated function on the term space for every document in the collection seems to be a very heavy requirements. However, it is quite common in IR to use indexing techniques that assign to every term a dierent weight for each document. The problems here is to identify the most appropriate measure given the semantics of the opinionated function. We will not tackle this problem in this paper. In the evaluation reported in Section 7 we use a monotonically decreasing transfer function that transfers from a term t a decreasing fraction of its probability to all the other terms in the term space once they are ordered in decreasing order of similarity. In particular, to simplify computations, in the evaluation of P (d! q) by general imaging on d, from each term not occuring in d we transfer its probability only to the rst 10 most similar terms occuring in d. The transfer function we use works in such away that the ith of these 10 terms ordered in decreasing order of similarity always gets a fraction of P (t) that is the double of the one that the i + 1th gets. A graphical model of this function is depicted in Figure 5. 18

19 Perc. of probability received Terms in order of similarity Figure 5: A graphical model of the term probability redistribution on its 10 closest terms. 5 A case study In order to understand better the dierences in the retrieval results that can be achieved by using dierent probabilistic models to evaluate the RSV, we will now give a simple example using a very small collection made on only 5 documents, and a term space made up of only 6 terms. We will show that we can obtain dierent rankings and dierent retrieval of these documents according to the retrieval model used. Let us suppose we have the following term space, T = ft 1 ;t 2 ;t 3 ;t 4 ;t 5 ;t 6 g and the following document collection. D = fd 1 ;d 2 ;d 3 ;d 4 ;d 5 g Let us suppose that both documents and query are represented by means of terms in T as explained in Section 2.2 as follows: d 1 = ft 1 ;t 5 ;t 6 g d 2 =ft 1 ;t 6 g d 3 =ft 1 ;t 2 ;t 3 ;t 5 g d 4 =ft 1 ;t 2 ;g d 5 =ft 1 ;t 3 ;t 6 g 19

20 and q = ft 1 ;t 4 ;t 6 g What our IR system is supposed to do is provide the user who submitted the query q with a ranking of the documents in D so that the ones most relevant to the query will be at the top of the ranking. The ranking is produced according to the RSV that is evaluated according to a particular retrieval model. In order to use the RbLI and RbGLI models we also need a measure of similarity between terms. This have been evaluated using a similarity function S(t i ;t j ). The similarity between terms is used to rank for every terms all other terms in T so that the closest one will be at the top of the ranking. For this example the ranking produced after evaluating the term{term similarity is the following: t 1 t 2 t 5 t 6 t 4 t 3 t 2 t 4 t 1 t 6 t 5 t 3 t 3 t 4 t 5 t 6 t 2 t 1 t 4 t 3 t 2 t 5 t 1 t 6 t 5 t 6 t 3 t 2 t 1 t 4 t 6 t 5 t 3 t 1 t 4 t 2 In order to use RbGLI we also need an opinionated function. We used this simple one: the probability ismoved from one term only to its two closest ones so that the closest gets 2=3 of the term probability while the second closest gets 1=3 of it. We do not report here the details of all the evaluations to determine the RSV of each document for each model. We suggest the interested reader to look at the Appendix of this paper. We will only discuss the results. The rankings produced by the retrieval models are the following: RbJP d 1 = d 2 = d 5 d 3 = d 4 RbCP d 2 d 5 d 4 d 1 d 3 RbLI d 2 d 5 d 1 d 4 d 3 RbGLI d 2 d 5 d 1 d 4 d 3 We can draw some interesting observations from these results: 1. RbJP produces some misleading results. Documents d 1, d 2, and d 5 are all ranked at the top and at the same level of estimated relevance. This is because the all share the same 2 terms in common with the query. The fact 20

21 that document d 2 have all of its terms in common with the query, while d 1 and d 5 have only 2 out of 3 terms in common with the query is not taken into the due account. Also documents d 3, and d 4 are ranked together since they have only one and the same term in common with the query. In this case, however, this is done quite correctly. 2. RbCP, RbLI, and RbGLI put at the top of the ranking the document d 2. This is correct since the document representation, i.e. terms t 1 and t 6,is completely included in the query representation. However the fact that the query also mention some other index term is not taken into consideration. In [5] we pointed out, in complete agreement with [13], how this fact could be taken into account byevaluating P (q! d). This could be done either using Imaging or General Imaging. A model that combines both P (d! q) and P (q! d) seems to be a very powerful tool for IR. 3. Both RbLI and RbGLI rank document d 1 before document d 4, as opposed to RbCP that ranks d 1 after d 4. The behaviour of RbCP is not correct since document d 1 shares 2 terms with the query while d 4 shares only 1. The reason why RbCP does that is due to the fact that d 4 is represented by only 2 terms, while d 1 is represented by 3 terms. This fact causes a movement of probability insuchaway that term t 1 (the only term d 4 has in common with the query) receives quite a large share of the probability moved from all other terms in the terms space that are not present ind 4. This is due to the fact that the \prior" probability oft 1 is double that of t 2 (the other term occuring in d 4 ), so the probability moved over to t 1 is double of that moved over to t 2. In RbLI and RbGLI the movement of probability is more balanced and it is also heavily inuenced by the fact that most terms in T are more similar to t 2 than to t 1. The term t 2 therefore gets a larger share of the probability moved from terms not occuring in d 4 to terms occuring in d 4 (see Appendix). 4. Both documents d 3 and d 4 have only one term in common with the query. They have been ranked at the same level by RbJP, while they have been ranked quite dierently by RbCP, RbLI and RbGLI. RbCP ranks d 4 quite high up in the ranking for reasons explained above, while d 3 is ranked last since the probability form terms not occuring in d 3 is moved over to a larger number of terms than d 4 and it is therefore more widespread. In particular, quite a lot of the probably moved around ends up on term t 5, which has a large prior probability. In RbLI and RbGLI the movement of probability is governed by the similarity between terms instead that by the \prior" probability. The fact that document d 4 is ranked before document d 3 is completely due to the similarities among terms in the term space. Dierent values of similarity among terms could have caused a dierent ranking. In the case of document d 3 none of the terms not occuring in it is very closed 21

22 to term t 1 (the term d 3 has in common with the query) and therefore no probability ismoved over to it. The same can be said of document d 4 where most of the probability moved around ends up on term t 2 that is not occuring in the query. 5. There is quite an interesting interpretation of the behaviour of RbLI and RbGLI in the context of word sense disambiguation. Let us imagine wehave a term t x that can be interpreted in two dierent senses: S 1 and S 2. Let suppose sense S 1 is more common than S 2 into the collection of documents D. In this case a similarity measure based on co-occurrence will pick up mainly the sense S 1 for t x, so giving high levels of similaritybetween t x and other terms used in the same context and with the same sense. Let now suppose term t x is used in a query; let us see how two documents using t x in the two dierent sense will be retrieved. Imagine, for simplicity, that the two documents d s1 and d s2 have the same number of terms related two their respective senses. According to RbLI and RbGLI term t x will receivea certain amount of probability coming from terms that are close to it but not occurring in the document. In document d s1 there will certainly be some terms closely related to the S 1 sense of t x, their probability will therefore be retained and not transfered over to t x. On the other hand, document d s2 will have fewer terms close to t x and there will be a large amount of probability shifted from terms close to t x but not occuring in d s2. This movement of probability will cause term t x to have a larger probability in d s2 than in d s1, and so document d s2 will ranked higher than d s1 in the list of (estimated) relevant documents. RbLI and RbGLI have therefore the eect of putting at the top of the ranking documents with uncommon senses of terms used in the query. This eect is however leveled down in queries and documents with a large number of terms. From this case study we believe we can conclude that RbLI and RbGLI perform more correctly than RbJP and RbCP. However, before giving such a strong statement we should test these four retrieval models using some standard testing technique, that is using some document test collection and produce comparative precision and recall tables. We will do this in the next following sections. 6 Experiments with test collections In order to study the retrieval eectiveness of the four models under considerations a series of tests were performed using some document test collections. Given the heavy computations necessary to evaluate the EMIM between every two terms, and give the theoretical nature of our goals, we decided to use some 22

23 Data Craneld CACM NPL documents queries terms in doc terms in query avg. doc. length avg. query length avg. rel. doc Table 5: Test collections data small but well studied test collection. The choice falls on the following three test collections that have been extensively used in the eld of IR: the Craneld 1400, the CACM, and the NPL test collections. The main characteristics of these three test collections are summarised in Table 5. Among the large number of references to these test collections we will only mention the followings: [4, 18]. Precision and recall 4 tables have been evaluated using their standard denition. The method of linear interpolation has been used to determine the standard values corresponding to intervals of 10% in recall gures. We are aware of the fact that linear interpolation estimates the best possible performance between two adjacent observed points (see [21], p. 152). However, since our experiments have the purpose to compare the retrieval eectiveness of the four proposed retrieval models only with each other, we think that this does not provide any misleading result. On the contrary, it enables us to compare the models taking into account their best possible performance. At the bottom of the precision and recall tables we also report two synthetic measures of eectiveness. They are the average precision, which is the simple average of all the precision values, and the average F measure (see [21], pp. 174), which is a combined eectiveness measure that takes into consideration both precision and recall as follows: F = 1 ( 1 P )+(1 ) 1 R where P and R are respectively the precision and recall values, and is a parameter that enables us to give more importance either to precision (when! 0) or to recall (when! 1). We use the value =0:5 so that equal importance to 4 Precision is the proportion of the retrieved set of documents that is relevant to the query. Recall is the proportion of all documents in the collections that are relevant to a query and that are actually retrieved. 23

24 Craneld 1400 Recall % Precision % RbJP RbLI RbCP RbGLI Avg %chg Max F %chg Table 6: Comparison of the performance of RbJP, RbLI, RbCP, and RbGLI using the Craneld 1400 test document collection precision and recall is given. We evaluated the F value for each pair of precision and recall values, and what is reported in the tables is the maximum values for each model. The percentage increase from one model to another is also reported. 7 Comparative evaluation We performed a comparative evaluation of the retrieval eectiveness of the four models presented above using the document collections and the experimental settings reported in Section 4 and 6. The results are reported in Tables 6, 7, and 8. The results are also reported in the form of Recall/Precision graphs in Figures 6, 7,and 8. We can observe that: any model in which there is a probability transfer (like RbLI, RbCP, and RbGLI) performs better than a model in which there is no such transfer (RbJP); any model that requires the probability transfer to be from one term to a set of terms (called \one-to-many" transfer, like in RbCP and RbGLI) performs better than models in which either there is no tranfer (RbJP) or 24

25 CACM Recall % Precision % RbJP RbLI RbCP RbGLI Avg %chg Max F %chg Table 7: Comparison of the performance of RbJP, RbLI, RbCP, and RbGLI using the CACM test document collection NPL Recall % Precision % RbJP RbLI RbCP RbGLI Avg %chg Max F %chg Table 8: Comparison of the performance of RbJP, RbLI, RbCP, and RbGLI using the NPL test document collection 25

26 Cranfield P/R graph RbJP RbLI RbCP RbGLI 70 Precision % Recall % Figure 6: Precision and recall graphs for the Craneld test collection. the transfer is from one term to a single other term (called \one-to-one" transfer, like in RbLI); a one-to-many transfer that takes into account the similarity between the donor and the receivers (RbGLI) performs better than a one-to-many tranfer that takes into account the probability ratio between the receivers (RbCP). These ndings are consistent over the three document collections, as the percentange changes in the average precision show. Despite the fact that we are using a simple term weighting schema and that we are experimenting of small test collections, we think we can nonetheless conclude that it is important to study further the probability kinematics of IR probabilistic retrieval models in order to be able to improve their performance. An interesting result of our study on the kinematics of the four models presented is that we can obtain higher levels of retrieval eectiveness by taking into consideration the similarity between the objects involved in the transfer of probability. We intend to pursue this result to design a new probabilistic model for IR. 8 Conclusions In this present study of the probability kinematics in IR, we believe we have shown that in principle a probability transfer that takes into account a measure of similaritybetween the donor and the recipient is more eective in the context of IR than a probability transfer that does not take that into account. Most current probabilistic retrieval models are based on a probability kinematics that does 26

27 CACM P/R graph RbJP RbLI RbCP RbGLI 70 Precision % Recall % Figure 7: Precision and recall graphs for the CACM test collection NPL P/R graph RbJP RbLI RbCP RbGLI 70 Precision % Recall % Figure 8: Precision and recall graph for the NPL test collection 27

28 not take into account similarity between terms or between documents, unless \ad hoc" weighting schemas, mostly based on clustering, are used. We would therefore like to suggest a further investigation into more complex and optimised models for probabilistic retrieval, where probability kinematics follows a nonclassical approach. The use of General Imaging is one of such approaches, but other ones can be developed using results achieved in other elds, such as Logics and Belief Revision theory. Our current results, summarised in this paper, seems to suggest that an improvements in retrieval eectiveness can be obtained by designing IR systems using probabilistic models that are based upon a dierent kind of probability kinematics. In our future work we intend to purse further this research direction by implementing and testing a new probabilistic IR system based upon General Imaging. Acknowledgements The authors would like to thank Iain Campbell and Mark Sanderson for the interesting and lively discussions on the topics of this paper. We will thank them even if they do not seem to believe that in IR as in most other elds theory ought to come before experimentation. 28

The troubles with using a logical model of IR on a large collection of documents

The troubles with using a logical model of IR on a large collection of documents The troubles with using a logical model of IR on a large collection of documents Fabio Crestani Dipartimento di Elettronica ed Informatica Universitá a degli Studi di Padova Padova - Italy Ian Ruthven

More information

A Formalization of Logical Imaging for IR using QT

A Formalization of Logical Imaging for IR using QT A Formalization of Logical Imaging for IR using QT September 1st, 2008 guido@dcs.gla.ac.uk http://www.dcs.gla.ac.uk/ guido Information Retrieval Group University of Glasgow QT in IR Why QT? General guidelines

More information

Information Retrieval Basic IR models. Luca Bondi

Information Retrieval Basic IR models. Luca Bondi Basic IR models Luca Bondi Previously on IR 2 d j q i IRM SC q i, d j IRM D, Q, R q i, d j d j = w 1,j, w 2,j,, w M,j T w i,j = 0 if term t i does not appear in document d j w i,j and w i:1,j assumed to

More information

Chap 2: Classical models for information retrieval

Chap 2: Classical models for information retrieval Chap 2: Classical models for information retrieval Jean-Pierre Chevallet & Philippe Mulhem LIG-MRIM Sept 2016 Jean-Pierre Chevallet & Philippe Mulhem Models of IR 1 / 81 Outline Basic IR Models 1 Basic

More information

Ranked Retrieval (2)

Ranked Retrieval (2) Text Technologies for Data Science INFR11145 Ranked Retrieval (2) Instructor: Walid Magdy 31-Oct-2017 Lecture Objectives Learn about Probabilistic models BM25 Learn about LM for IR 2 1 Recall: VSM & TFIDF

More information

A Preference Semantics. for Ground Nonmonotonic Modal Logics. logics, a family of nonmonotonic modal logics obtained by means of a

A Preference Semantics. for Ground Nonmonotonic Modal Logics. logics, a family of nonmonotonic modal logics obtained by means of a A Preference Semantics for Ground Nonmonotonic Modal Logics Daniele Nardi and Riccardo Rosati Dipartimento di Informatica e Sistemistica, Universita di Roma \La Sapienza", Via Salaria 113, I-00198 Roma,

More information

Stochastic dominance with imprecise information

Stochastic dominance with imprecise information Stochastic dominance with imprecise information Ignacio Montes, Enrique Miranda, Susana Montes University of Oviedo, Dep. of Statistics and Operations Research. Abstract Stochastic dominance, which is

More information

Can Vector Space Bases Model Context?

Can Vector Space Bases Model Context? Can Vector Space Bases Model Context? Massimo Melucci University of Padua Department of Information Engineering Via Gradenigo, 6/a 35031 Padova Italy melo@dei.unipd.it Abstract Current Information Retrieval

More information

Information Retrieval and Web Search Engines

Information Retrieval and Web Search Engines Information Retrieval and Web Search Engines Lecture 4: Probabilistic Retrieval Models April 29, 2010 Wolf-Tilo Balke and Joachim Selke Institut für Informationssysteme Technische Universität Braunschweig

More information

HELP US REMOVE OUTDATED BELIEFS?

HELP US REMOVE OUTDATED BELIEFS? CAN INFORMATION THEORY HELP US REMOVE OUTDATED BELIEFS? Erik J. Olsson Lund University Cognitive Science Kungshuset, Lundagård S-222 22 Lund, Sweden E-mail Erik.Olsson@filosofi.uu.se Abstract: It is argued

More information

IR Models: The Probabilistic Model. Lecture 8

IR Models: The Probabilistic Model. Lecture 8 IR Models: The Probabilistic Model Lecture 8 ' * ) ( % $ $ +#! "#! '& & Probability of Relevance? ' ', IR is an uncertain process Information need to query Documents to index terms Query terms and index

More information

Minimum and maximum values *

Minimum and maximum values * OpenStax-CNX module: m17417 1 Minimum and maximum values * Sunil Kumar Singh This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 2.0 In general context, a

More information

Lecture 3: Probabilistic Retrieval Models

Lecture 3: Probabilistic Retrieval Models Probabilistic Retrieval Models Information Retrieval and Web Search Engines Lecture 3: Probabilistic Retrieval Models November 5 th, 2013 Wolf-Tilo Balke and Kinda El Maarry Institut für Informationssysteme

More information

Significance Testing with Incompletely Randomised Cases Cannot Possibly Work

Significance Testing with Incompletely Randomised Cases Cannot Possibly Work Human Journals Short Communication December 2018 Vol.:11, Issue:2 All rights are reserved by Stephen Gorard FRSA FAcSS Significance Testing with Incompletely Randomised Cases Cannot Possibly Work Keywords:

More information

A mathematical model of social conformity as a function of group size

A mathematical model of social conformity as a function of group size A mathematical model of social conformity as a function of group size To Chin, Yu April 20, 2013 This paper aims to model mathematically the dependence of conformity on majority size with the intention

More information

Modal Logic. UIT2206: The Importance of Being Formal. Martin Henz. March 19, 2014

Modal Logic. UIT2206: The Importance of Being Formal. Martin Henz. March 19, 2014 Modal Logic UIT2206: The Importance of Being Formal Martin Henz March 19, 2014 1 Motivation The source of meaning of formulas in the previous chapters were models. Once a particular model is chosen, say

More information

In this initial chapter, you will be introduced to, or more than likely be reminded of, a

In this initial chapter, you will be introduced to, or more than likely be reminded of, a 1 Sets In this initial chapter, you will be introduced to, or more than likely be reminded of, a fundamental idea that occurs throughout mathematics: sets. Indeed, a set is an object from which every mathematical

More information

Information Retrieval

Information Retrieval Introduction to Information Retrieval Lecture 11: Probabilistic Information Retrieval 1 Outline Basic Probability Theory Probability Ranking Principle Extensions 2 Basic Probability Theory For events A

More information

Russell s logicism. Jeff Speaks. September 26, 2007

Russell s logicism. Jeff Speaks. September 26, 2007 Russell s logicism Jeff Speaks September 26, 2007 1 Russell s definition of number............................ 2 2 The idea of reducing one theory to another.................... 4 2.1 Axioms and theories.............................

More information

Logic and Philosophical Logic. 1 Inferentialism. Inferentialism and Meaning Underdetermination

Logic and Philosophical Logic. 1 Inferentialism. Inferentialism and Meaning Underdetermination Logic and Philosophical Logic Inferentialism and Meaning Underdetermination AC Paseau alexanderpaseau@philosophyoxacuk 28 January 2019 In the rst half of today's class, we looked at Tarski's account of

More information

Non-independence in Statistical Tests for Discrete Cross-species Data

Non-independence in Statistical Tests for Discrete Cross-species Data J. theor. Biol. (1997) 188, 507514 Non-independence in Statistical Tests for Discrete Cross-species Data ALAN GRAFEN* AND MARK RIDLEY * St. John s College, Oxford OX1 3JP, and the Department of Zoology,

More information

Some Formal Analysis of Rocchio s Similarity-Based Relevance Feedback Algorithm

Some Formal Analysis of Rocchio s Similarity-Based Relevance Feedback Algorithm Some Formal Analysis of Rocchio s Similarity-Based Relevance Feedback Algorithm Zhixiang Chen (chen@cs.panam.edu) Department of Computer Science, University of Texas-Pan American, 1201 West University

More information

DECISIONS UNDER UNCERTAINTY

DECISIONS UNDER UNCERTAINTY August 18, 2003 Aanund Hylland: # DECISIONS UNDER UNCERTAINTY Standard theory and alternatives 1. Introduction Individual decision making under uncertainty can be characterized as follows: The decision

More information

Roots of Unity, Cyclotomic Polynomials and Applications

Roots of Unity, Cyclotomic Polynomials and Applications Swiss Mathematical Olympiad smo osm Roots of Unity, Cyclotomic Polynomials and Applications The task to be done here is to give an introduction to the topics in the title. This paper is neither complete

More information

CHAPTER 2 INTRODUCTION TO CLASSICAL PROPOSITIONAL LOGIC

CHAPTER 2 INTRODUCTION TO CLASSICAL PROPOSITIONAL LOGIC CHAPTER 2 INTRODUCTION TO CLASSICAL PROPOSITIONAL LOGIC 1 Motivation and History The origins of the classical propositional logic, classical propositional calculus, as it was, and still often is called,

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Ontology on Shaky Grounds

Ontology on Shaky Grounds 1 Ontology on Shaky Grounds Jean-Pierre Marquis Département de Philosophie Université de Montréal C.P. 6128, succ. Centre-ville Montréal, P.Q. Canada H3C 3J7 Jean-Pierre.Marquis@umontreal.ca In Realistic

More information

Lecture 9: Probabilistic IR The Binary Independence Model and Okapi BM25

Lecture 9: Probabilistic IR The Binary Independence Model and Okapi BM25 Lecture 9: Probabilistic IR The Binary Independence Model and Okapi BM25 Trevor Cohn (Slide credits: William Webber) COMP90042, 2015, Semester 1 What we ll learn in this lecture Probabilistic models for

More information

BETWEEN THE LOGIC OF PARMENIDES AND THE LOGIC OF LIAR

BETWEEN THE LOGIC OF PARMENIDES AND THE LOGIC OF LIAR Bulletin of the Section of Logic Volume 38:3/4 (2009), pp. 123 133 Kordula Świȩtorzecka BETWEEN THE LOGIC OF PARMENIDES AND THE LOGIC OF LIAR Abstract In the presented text we shall focus on some specific

More information

0 o 1 i B C D 0/1 0/ /1

0 o 1 i B C D 0/1 0/ /1 A Comparison of Dominance Mechanisms and Simple Mutation on Non-Stationary Problems Jonathan Lewis,? Emma Hart, Graeme Ritchie Department of Articial Intelligence, University of Edinburgh, Edinburgh EH

More information

Part I Qualitative Probabilistic Networks

Part I Qualitative Probabilistic Networks Part I Qualitative Probabilistic Networks In which we study enhancements of the framework of qualitative probabilistic networks. Qualitative probabilistic networks allow for studying the reasoning behaviour

More information

Knowledge Discovery in Data: Overview. Naïve Bayesian Classification. .. Spring 2009 CSC 466: Knowledge Discovery from Data Alexander Dekhtyar..

Knowledge Discovery in Data: Overview. Naïve Bayesian Classification. .. Spring 2009 CSC 466: Knowledge Discovery from Data Alexander Dekhtyar.. Spring 2009 CSC 466: Knowledge Discovery from Data Alexander Dekhtyar Knowledge Discovery in Data: Naïve Bayes Overview Naïve Bayes methodology refers to a probabilistic approach to information discovery

More information

Information Retrieval and Topic Models. Mausam (Based on slides of W. Arms, Dan Jurafsky, Thomas Hofmann, Ata Kaban, Chris Manning, Melanie Martin)

Information Retrieval and Topic Models. Mausam (Based on slides of W. Arms, Dan Jurafsky, Thomas Hofmann, Ata Kaban, Chris Manning, Melanie Martin) Information Retrieval and Topic Models Mausam (Based on slides of W. Arms, Dan Jurafsky, Thomas Hofmann, Ata Kaban, Chris Manning, Melanie Martin) Sec. 1.1 Unstructured data in 1620 Which plays of Shakespeare

More information

Rough Sets for Uncertainty Reasoning

Rough Sets for Uncertainty Reasoning Rough Sets for Uncertainty Reasoning S.K.M. Wong 1 and C.J. Butz 2 1 Department of Computer Science, University of Regina, Regina, Canada, S4S 0A2, wong@cs.uregina.ca 2 School of Information Technology

More information

TEACHER NOTES FOR ADVANCED MATHEMATICS 1 FOR AS AND A LEVEL

TEACHER NOTES FOR ADVANCED MATHEMATICS 1 FOR AS AND A LEVEL April 2017 TEACHER NOTES FOR ADVANCED MATHEMATICS 1 FOR AS AND A LEVEL This book is designed both as a complete AS Mathematics course, and as the first year of the full A level Mathematics. The content

More information

Computational Tasks and Models

Computational Tasks and Models 1 Computational Tasks and Models Overview: We assume that the reader is familiar with computing devices but may associate the notion of computation with specific incarnations of it. Our first goal is to

More information

An Attempt To Understand Tilling Approach In proving The Littlewood Conjecture

An Attempt To Understand Tilling Approach In proving The Littlewood Conjecture An Attempt To Understand Tilling Approach In proving The Littlewood Conjecture 1 Final Report For Math 899- Dr. Cheung Amira Alkeswani Dr. Cheung I really appreciate your acceptance in adding me to your

More information

Characterization of Semantics for Argument Systems

Characterization of Semantics for Argument Systems Characterization of Semantics for Argument Systems Philippe Besnard and Sylvie Doutre IRIT Université Paul Sabatier 118, route de Narbonne 31062 Toulouse Cedex 4 France besnard, doutre}@irit.fr Abstract

More information

Designing and Evaluating Generic Ontologies

Designing and Evaluating Generic Ontologies Designing and Evaluating Generic Ontologies Michael Grüninger Department of Industrial Engineering University of Toronto gruninger@ie.utoronto.ca August 28, 2007 1 Introduction One of the many uses of

More information

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Probabilistic Reasoning. (Mostly using Bayesian Networks) Probabilistic Reasoning (Mostly using Bayesian Networks) Introduction: Why probabilistic reasoning? The world is not deterministic. (Usually because information is limited.) Ways of coping with uncertainty

More information

A note on fuzzy predicate logic. Petr H jek 1. Academy of Sciences of the Czech Republic

A note on fuzzy predicate logic. Petr H jek 1. Academy of Sciences of the Czech Republic A note on fuzzy predicate logic Petr H jek 1 Institute of Computer Science, Academy of Sciences of the Czech Republic Pod vod renskou v 2, 182 07 Prague. Abstract. Recent development of mathematical fuzzy

More information

Proof techniques (section 2.1)

Proof techniques (section 2.1) CHAPTER 1 Proof techniques (section 2.1) What we have seen so far: 1.1. Theorems and Informal proofs Argument: P 1 P n Q Syntax: how it's written Semantic: meaning in a given interpretation Valid argument:

More information

Information Retrieval

Information Retrieval Introduction to Information Retrieval Lecture 12: Language Models for IR Outline Language models Language Models for IR Discussion What is a language model? We can view a finite state automaton as a deterministic

More information

Introduction. Chapter 1

Introduction. Chapter 1 Chapter 1 Introduction In this book we will be concerned with supervised learning, which is the problem of learning input-output mappings from empirical data (the training dataset). Depending on the characteristics

More information

Relevant Logic. Daniel Bonevac. March 20, 2013

Relevant Logic. Daniel Bonevac. March 20, 2013 March 20, 2013 The earliest attempts to devise a relevance logic that avoided the problem of explosion centered on the conditional. FDE, however, has no conditional operator, or a very weak one. If we

More information

Exhaustive Classication of Finite Classical Probability Spaces with Regard to the Notion of Causal Up-to-n-closedness

Exhaustive Classication of Finite Classical Probability Spaces with Regard to the Notion of Causal Up-to-n-closedness Exhaustive Classication of Finite Classical Probability Spaces with Regard to the Notion of Causal Up-to-n-closedness Michaª Marczyk, Leszek Wro«ski Jagiellonian University, Kraków 16 June 2009 Abstract

More information

Retrieval by Content. Part 2: Text Retrieval Term Frequency and Inverse Document Frequency. Srihari: CSE 626 1

Retrieval by Content. Part 2: Text Retrieval Term Frequency and Inverse Document Frequency. Srihari: CSE 626 1 Retrieval by Content Part 2: Text Retrieval Term Frequency and Inverse Document Frequency Srihari: CSE 626 1 Text Retrieval Retrieval of text-based information is referred to as Information Retrieval (IR)

More information

Conditional probabilistic reasoning without conditional logic

Conditional probabilistic reasoning without conditional logic Conditional probabilistic reasoning without conditional logic Fabrizio Sebastiani Istituto di Elaborazione dell Informazione Consiglio Nazionale delle Ricerche Via S. Maria, 46-56126 Pisa (Italy) E-mail:

More information

Latent Semantic Analysis. Hongning Wang

Latent Semantic Analysis. Hongning Wang Latent Semantic Analysis Hongning Wang CS@UVa Recap: vector space model Represent both doc and query by concept vectors Each concept defines one dimension K concepts define a high-dimensional space Element

More information

A FRAGMENT OF BOOLE S ALGEBRAIC LOGIC SUITABLE FOR TRADITIONAL SYLLOGISTIC LOGIC

A FRAGMENT OF BOOLE S ALGEBRAIC LOGIC SUITABLE FOR TRADITIONAL SYLLOGISTIC LOGIC A FRAGMENT OF BOOLE S ALGEBRAIC LOGIC SUITABLE FOR TRADITIONAL SYLLOGISTIC LOGIC STANLEY BURRIS 1. Introduction Boole introduced his agebraic approach to logic in 1847 in an 82 page monograph, The Mathematical

More information

A Little Deductive Logic

A Little Deductive Logic A Little Deductive Logic In propositional or sentential deductive logic, we begin by specifying that we will use capital letters (like A, B, C, D, and so on) to stand in for sentences, and we assume that

More information

Introduction: What Is Modal Logic?

Introduction: What Is Modal Logic? Introduction: What Is Modal Logic? Strictly speaking, modal logic studies reasoning that involves the use of the expressions necessarily and possibly. The main idea is to introduce the symbols (necessarily)

More information

Logicality of Operators

Logicality of Operators Logicality of Operators Tomoya Sato Abstract Characterizing logical operators has been crucially important in the philosophy of logic. One reason for this importance is that the boundary between logically

More information

Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events

Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events Massimo Franceschet Angelo Montanari Dipartimento di Matematica e Informatica, Università di Udine Via delle

More information

1 The Well Ordering Principle, Induction, and Equivalence Relations

1 The Well Ordering Principle, Induction, and Equivalence Relations 1 The Well Ordering Principle, Induction, and Equivalence Relations The set of natural numbers is the set N = f1; 2; 3; : : :g. (Some authors also include the number 0 in the natural numbers, but number

More information

Modern Information Retrieval

Modern Information Retrieval Modern Information Retrieval Chapter 3 Modeling Introduction to IR Models Basic Concepts The Boolean Model Term Weighting The Vector Model Probabilistic Model Retrieval Evaluation, Modern Information Retrieval,

More information

On the Revision of Probabilistic Belief States. Craig Boutilier. Department of Computer Science. University of British Columbia

On the Revision of Probabilistic Belief States. Craig Boutilier. Department of Computer Science. University of British Columbia On the Revision of Probabilistic Belief States Craig Boutilier Department of Computer Science University of British Columbia Vancouver, British Columbia CANADA, V6T 1Z4 Abstract In this paper we describe

More information

National Accelerator Laboratory

National Accelerator Laboratory Fermi National Accelerator Laboratory FERMILAB-Conf-97/149-E DØ QCD Results Using the? Jet-Finding Algorithm in p p Collisions at p s = 1800 GeV D. Lincoln For the D Collaboration University of Michigan

More information

A version of for which ZFC can not predict a single bit Robert M. Solovay May 16, Introduction In [2], Chaitin introd

A version of for which ZFC can not predict a single bit Robert M. Solovay May 16, Introduction In [2], Chaitin introd CDMTCS Research Report Series A Version of for which ZFC can not Predict a Single Bit Robert M. Solovay University of California at Berkeley CDMTCS-104 May 1999 Centre for Discrete Mathematics and Theoretical

More information

Perfect matchings in highly cyclically connected regular graphs

Perfect matchings in highly cyclically connected regular graphs Perfect matchings in highly cyclically connected regular graphs arxiv:1709.08891v1 [math.co] 6 Sep 017 Robert Lukot ka Comenius University, Bratislava lukotka@dcs.fmph.uniba.sk Edita Rollová University

More information

Uncertainty and Rules

Uncertainty and Rules Uncertainty and Rules We have already seen that expert systems can operate within the realm of uncertainty. There are several sources of uncertainty in rules: Uncertainty related to individual rules Uncertainty

More information

INTENSIONS MARCUS KRACHT

INTENSIONS MARCUS KRACHT INTENSIONS MARCUS KRACHT 1. The Way Things Are This note accompanies the introduction of Chapter 4 of the lecture notes. I shall provide some formal background and technology. Let a language L be given

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

SEMANTICAL CONSIDERATIONS ON NONMONOTONIC LOGIC. Robert C. Moore Artificial Intelligence Center SRI International, Menlo Park, CA 94025

SEMANTICAL CONSIDERATIONS ON NONMONOTONIC LOGIC. Robert C. Moore Artificial Intelligence Center SRI International, Menlo Park, CA 94025 SEMANTICAL CONSIDERATIONS ON NONMONOTONIC LOGIC Robert C. Moore Artificial Intelligence Center SRI International, Menlo Park, CA 94025 ABSTRACT Commonsense reasoning is "nonmonotonic" in the sense that

More information

Mathematics: applications and interpretation SL

Mathematics: applications and interpretation SL Mathematics: applications and interpretation SL Chapter 1: Approximations and error A Rounding numbers B Approximations C Errors in measurement D Absolute and percentage error The first two sections of

More information

Richard DiSalvo. Dr. Elmer. Mathematical Foundations of Economics. Fall/Spring,

Richard DiSalvo. Dr. Elmer. Mathematical Foundations of Economics. Fall/Spring, The Finite Dimensional Normed Linear Space Theorem Richard DiSalvo Dr. Elmer Mathematical Foundations of Economics Fall/Spring, 20-202 The claim that follows, which I have called the nite-dimensional normed

More information

Graph Theory. Thomas Bloom. February 6, 2015

Graph Theory. Thomas Bloom. February 6, 2015 Graph Theory Thomas Bloom February 6, 2015 1 Lecture 1 Introduction A graph (for the purposes of these lectures) is a finite set of vertices, some of which are connected by a single edge. Most importantly,

More information

PV211: Introduction to Information Retrieval

PV211: Introduction to Information Retrieval PV211: Introduction to Information Retrieval http://www.fi.muni.cz/~sojka/pv211 IIR 11: Probabilistic Information Retrieval Handout version Petr Sojka, Hinrich Schütze et al. Faculty of Informatics, Masaryk

More information

6.867 Machine learning, lecture 23 (Jaakkola)

6.867 Machine learning, lecture 23 (Jaakkola) Lecture topics: Markov Random Fields Probabilistic inference Markov Random Fields We will briefly go over undirected graphical models or Markov Random Fields (MRFs) as they will be needed in the context

More information

Boolean and Vector Space Retrieval Models

Boolean and Vector Space Retrieval Models Boolean and Vector Space Retrieval Models Many slides in this section are adapted from Prof. Joydeep Ghosh (UT ECE) who in turn adapted them from Prof. Dik Lee (Univ. of Science and Tech, Hong Kong) 1

More information

Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events

Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events Massimo Franceschet Angelo Montanari Dipartimento di Matematica e Informatica, Università di Udine Via delle

More information

PRIME GENERATING LUCAS SEQUENCES

PRIME GENERATING LUCAS SEQUENCES PRIME GENERATING LUCAS SEQUENCES PAUL LIU & RON ESTRIN Science One Program The University of British Columbia Vancouver, Canada April 011 1 PRIME GENERATING LUCAS SEQUENCES Abstract. The distribution of

More information

CHAPTER 1 INTRODUCTION TO BRT

CHAPTER 1 INTRODUCTION TO BRT CHAPTER 1 INTRODUCTION TO BRT 1.1. General Formulation. 1.2. Some BRT Settings. 1.3. Complementation Theorems. 1.4. Thin Set Theorems. 1.1. General Formulation. Before presenting the precise formulation

More information

An Alternative To The Iteration Operator Of. Propositional Dynamic Logic. Marcos Alexandre Castilho 1. IRIT - Universite Paul Sabatier and

An Alternative To The Iteration Operator Of. Propositional Dynamic Logic. Marcos Alexandre Castilho 1. IRIT - Universite Paul Sabatier and An Alternative To The Iteration Operator Of Propositional Dynamic Logic Marcos Alexandre Castilho 1 IRIT - Universite Paul abatier and UFPR - Universidade Federal do Parana (Brazil) Andreas Herzig IRIT

More information

cse541 LOGIC FOR COMPUTER SCIENCE

cse541 LOGIC FOR COMPUTER SCIENCE cse541 LOGIC FOR COMPUTER SCIENCE Professor Anita Wasilewska Spring 2015 LECTURE 2 Chapter 2 Introduction to Classical Propositional Logic PART 1: Classical Propositional Model Assumptions PART 2: Syntax

More information

Towards Collaborative Information Retrieval

Towards Collaborative Information Retrieval Towards Collaborative Information Retrieval Markus Junker, Armin Hust, and Stefan Klink German Research Center for Artificial Intelligence (DFKI GmbH), P.O. Box 28, 6768 Kaiserslautern, Germany {markus.junker,

More information

Genuine atomic multicast in asynchronous distributed systems

Genuine atomic multicast in asynchronous distributed systems Theoretical Computer Science 254 (2001) 297 316 www.elsevier.com/locate/tcs Genuine atomic multicast in asynchronous distributed systems Rachid Guerraoui, Andre Schiper Departement d Informatique, Ecole

More information

ONLINE APPENDICES FOR INCENTIVES IN EXPERIMENTS: A THEORETICAL INVESTIGATION BY AZRIELI, CHAMBERS & HEALY

ONLINE APPENDICES FOR INCENTIVES IN EXPERIMENTS: A THEORETICAL INVESTIGATION BY AZRIELI, CHAMBERS & HEALY ONLINE APPENDICES FOR INCENTIVES IN EXPERIMENTS: A THEORETICAL INVESTIGATION BY AZRIELI, CHAMBERS & HEALY Appendix B. Modeling Games as Decisions In this appendix we describe how one can move seamlessly

More information

Lecture 14 - P v.s. NP 1

Lecture 14 - P v.s. NP 1 CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanford.edu) February 27, 2018 Lecture 14 - P v.s. NP 1 In this lecture we start Unit 3 on NP-hardness and approximation

More information

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE

DEPARTMENT OF COMPUTER SCIENCE Autumn Semester MACHINE LEARNING AND ADAPTIVE INTELLIGENCE Data Provided: None DEPARTMENT OF COMPUTER SCIENCE Autumn Semester 203 204 MACHINE LEARNING AND ADAPTIVE INTELLIGENCE 2 hours Answer THREE of the four questions. All questions carry equal weight. Figures

More information

Boxlets: a Fast Convolution Algorithm for. Signal Processing and Neural Networks. Patrice Y. Simard, Leon Bottou, Patrick Haner and Yann LeCun

Boxlets: a Fast Convolution Algorithm for. Signal Processing and Neural Networks. Patrice Y. Simard, Leon Bottou, Patrick Haner and Yann LeCun Boxlets: a Fast Convolution Algorithm for Signal Processing and Neural Networks Patrice Y. Simard, Leon Bottou, Patrick Haner and Yann LeCun AT&T Labs-Research 100 Schultz Drive, Red Bank, NJ 07701-7033

More information

Expressive Power, Mood, and Actuality

Expressive Power, Mood, and Actuality Expressive Power, Mood, and Actuality Rohan French Abstract In Wehmeier (2004) we are presented with the subjunctive modal language, a way of dealing with the expressive inadequacy of modal logic by marking

More information

THE ROLE OF COMPUTER BASED TECHNOLOGY IN DEVELOPING UNDERSTANDING OF THE CONCEPT OF SAMPLING DISTRIBUTION

THE ROLE OF COMPUTER BASED TECHNOLOGY IN DEVELOPING UNDERSTANDING OF THE CONCEPT OF SAMPLING DISTRIBUTION THE ROLE OF COMPUTER BASED TECHNOLOGY IN DEVELOPING UNDERSTANDING OF THE CONCEPT OF SAMPLING DISTRIBUTION Kay Lipson Swinburne University of Technology Australia Traditionally, the concept of sampling

More information

DR.RUPNATHJI( DR.RUPAK NATH )

DR.RUPNATHJI( DR.RUPAK NATH ) Contents 1 Sets 1 2 The Real Numbers 9 3 Sequences 29 4 Series 59 5 Functions 81 6 Power Series 105 7 The elementary functions 111 Chapter 1 Sets It is very convenient to introduce some notation and terminology

More information

Introduction to Metalogic

Introduction to Metalogic Introduction to Metalogic Hans Halvorson September 21, 2016 Logical grammar Definition. A propositional signature Σ is a collection of items, which we call propositional constants. Sometimes these propositional

More information

FACTORIZATION AND THE PRIMES

FACTORIZATION AND THE PRIMES I FACTORIZATION AND THE PRIMES 1. The laws of arithmetic The object of the higher arithmetic is to discover and to establish general propositions concerning the natural numbers 1, 2, 3,... of ordinary

More information

rules strictly includes the class of extended disjunctive programs. Interestingly, once not appears positively as above, the principle of minimality d

rules strictly includes the class of extended disjunctive programs. Interestingly, once not appears positively as above, the principle of minimality d On Positive Occurrences of Negation as Failure 3 Katsumi Inoue Department of Information and Computer Sciences Toyohashi University of Technology Tempaku-cho, Toyohashi 441, Japan inoue@tutics.tut.ac.jp

More information

Precis of Aristotle s Modal Syllogistic

Precis of Aristotle s Modal Syllogistic Philosophy and Phenomenological Research Philosophy and Phenomenological Research Vol. XC No. 3, May 2015 doi: 10.1111/phpr.12185 2015 Philosophy and Phenomenological Research, LLC Precis of Aristotle

More information

Module 7 D and Equivalent Systems

Module 7 D and Equivalent Systems Module 7 D and Equivalent Systems G J Mattey May 2, 2007 Contents 1 he Semantical System DI 1 2 he Derivational System DD 4 3 he Axiom System D 6 4 Applications of the D-Systems 6 In this and subsequent

More information

2 Exercises 1. The following represent graphs of functions from the real numbers R to R. Decide which are one-to-one, which are onto, which are neithe

2 Exercises 1. The following represent graphs of functions from the real numbers R to R. Decide which are one-to-one, which are onto, which are neithe Infinity and Counting 1 Peter Trapa September 28, 2005 There are 10 kinds of people in the world: those who understand binary, and those who don't. Welcome to the rst installment of the 2005 Utah Math

More information

A Dynamic Logic of Iterated Belief Change Bernd van Linder Utrecht University Department of Computer Science P.O. Box TB Utrecht The Nethe

A Dynamic Logic of Iterated Belief Change Bernd van Linder Utrecht University Department of Computer Science P.O. Box TB Utrecht The Nethe A Dynamic Logic of Iterated Belief Change Bernd van Linder Utrecht University Department of Computer Science P.O. Box 80.089 3508 TB Utrecht The Netherlands Email: bernd@cs.ruu.nl Abstract In this paper

More information

A Structuralist Account of Logic

A Structuralist Account of Logic Croatian Journal of Philosophy Vol. VIII, No. 23, 2008 Majda Trobok, Department of Philosophy University of Rijeka A Structuralist Account of Logic The lynch-pin of the structuralist account of logic endorsed

More information

Neale and the slingshot Fabrice Correia

Neale and the slingshot Fabrice Correia a Neale and the slingshot Fabrice Correia 'Slingshot arguments' is a label for a class of arguments which includes Church's argument to the effect that if sentences designate propositions, then there are

More information

Predicates, Quantifiers and Nested Quantifiers

Predicates, Quantifiers and Nested Quantifiers Predicates, Quantifiers and Nested Quantifiers Predicates Recall the example of a non-proposition in our first presentation: 2x=1. Let us call this expression P(x). P(x) is not a proposition because x

More information

Bayesian vs frequentist techniques for the analysis of binary outcome data

Bayesian vs frequentist techniques for the analysis of binary outcome data 1 Bayesian vs frequentist techniques for the analysis of binary outcome data By M. Stapleton Abstract We compare Bayesian and frequentist techniques for analysing binary outcome data. Such data are commonly

More information

Price Competition and Endogenous Valuation in Search Advertising

Price Competition and Endogenous Valuation in Search Advertising Price Competition and Endogenous Valuation in Search Advertising Lizhen Xu Jianqing Chen Andrew Whinston Web Appendix A Heterogeneous Consumer Valuation In the baseline model, we assumed that consumers

More information

Comparing Measures of Central Tendency *

Comparing Measures of Central Tendency * OpenStax-CNX module: m11011 1 Comparing Measures of Central Tendency * David Lane This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 1.0 1 Comparing Measures

More information

Propositional Logic Truth-functionality Definitions Soundness Completeness Inferences. Modal Logic. Daniel Bonevac.

Propositional Logic Truth-functionality Definitions Soundness Completeness Inferences. Modal Logic. Daniel Bonevac. January 22, 2013 Modal logic is, among other things, the logic of possibility and necessity. Its history goes back at least to Aristotle s discussion of modal syllogisms in the Prior Analytics. But modern

More information

The WENO Method for Non-Equidistant Meshes

The WENO Method for Non-Equidistant Meshes The WENO Method for Non-Equidistant Meshes Philip Rupp September 11, 01, Berlin Contents 1 Introduction 1.1 Settings and Conditions...................... The WENO Schemes 4.1 The Interpolation Problem.....................

More information

Pell's Equation. Luke Kanczes

Pell's Equation. Luke Kanczes Pell's Equation Luke Kanczes Introduction. Pell's Equation Pells equation is any Diophantine equation which takes the form [] x Dy = () for positive integers x and y, where D is a xed positive integer

More information