An Argumentation-Based Interpreter for Golog Programs

Size: px
Start display at page:

Download "An Argumentation-Based Interpreter for Golog Programs"

Transcription

1 An Argumentation-Based Interpreter for Golog Programs Michelle L. Blom and Adrian R. Pearce NICTA Victoria Research Laboratory Department of Computer Science and Software Engineering The University of Melbourne {mlblom, Abstract This paper presents an argumentation-based interpreter for Golog programs. Traditional Golog interpreters are not designed to find the most preferred executions of a program from the perspective of an agent. Existing techniques developed to discover these executions are limited in terms of how the preferences of an agent can be expressed, and the variety of preference types that can be used to guide search for a solution. The presented work combines the use of argumentation to compare executions relative to a set of general comparison principles, and the theory behind best first search to reduce the cost of the search process. To the best of our knowledge this is the first work to integrate argumentation and the interpretation of Golog programs, and to use argumentation as a tool for best first search. 1 Introduction Golog [Levesque et al., 1997] is a logic programming language for agents designed to represent complex actions and procedures in the situation calculus. Each execution of a Golog program represents a way of achieving a particular goal. A traditional Golog interpreter is designed to find the possible executions of a program, not to identify those that are the most preferred. Several approaches have been developed to achieve this task, and extend Golog with preferences. DTGolog a decision theoretic extension of Golog maximises utility in the interpretation of Golog programs [Boutilier et al., 2000]. In [Fritz and McIlraith, 2006], DTGolog is extended with the incorporation of qualitative preferences in the form of temporal logic constraints of varying priority. In [Sohrabi et al., 2006], best first search is applied in the interpretation of Golog programs to discover a web service composition that is the most preferred given temporal logic preference formulae. The variety of preference types that can be expressed in these existing approaches is limited. The utilities of [Boutilier et al., 2000] and the qualitative preferences of [Fritz and McIlraith, 2006; Sohrabi et al., 2006] do not express all the ways in which executions of a program can be compared. These approaches, for example, cannot capture comparative preference derived from the relationships that exist between one choice and another. Examples of such preferences include those based on: intangible, immeasurable factors such as taste and perception; decision heuristics encapsulating past decision-making experience; and lexicographic or semi-lexicographic [Tversky, 1969] comparison rules. In this paper, we present an argumentation-based Golog interpreter that aims to address this limitation, while maintaining the advantage of reduced search cost through best first search. Argumentation techniques have gained widespread use in both agent and multi-agent systems as tools for the support of more human-like decision-making processes [Ouerdane et al., 2007]. In this paper, we first develop an argumentation framework for decision-making (ADF ) that constructs a strict preference profile (a dominance graph) over a set of executions A of a Golog program. This framework allows an agent to compare executions according to whichever principles it desires (denoted comparison principles), including those describing comparative preference. We do not assume that the resulting profile is transitive or acyclic, and as such we apply choice theory techniques to define the most preferred subset of A (MP S(A)). We apply a choice rule, such as the Schwartz [Schwartz, 1972] criterion, to a dominance graph over A to select these most preferred alternatives. We then characterise the interpretation of a Golog program as best first search (using the transition semantics of [de Giacomo et al., 2000]). At each stage in the search for a most preferred execution of a program δ, an instantiation of our ADF provides an estimate of which available transitions will lead to an execution in MP S(A), where A denotes the executions of δ. The concept of speculation (a tool for predicting future opportunities in search) and a prioritised selection algorithm are used to select a transition to perform at each stage of the search process. To select a transition, an agent speculates on (estimates) the nature of the preferences that exist between complete executions of a program, based on the preferences that exist between partial executions at each stage in search. Our argumentation-based best first search interpreter is shown to always find an execution in the most preferred subset of a program s execution set A (provided A = ). The remainder of this paper is structured as follows. In Section 2, we provide an introduction to the Golog programming language. In Section 3, we describe in more detail the existing preference-based Golog interpreters, their limitations, and our motivation for using argumentation-based best

2 first search to address them. Our ADF, defining MP S(A) for a set of program executions A, is outlined in Section 4. In Section 5, we describe argumentation-based best first search. 2 Golog, and The Situation Calculus Golog programs are described using axioms of the situation calculus, together with constructs to encode procedures, actions, iteration, and non-deterministic choice [Levesque et al., 1997]. The situation calculus enables reasoning about actions, which change the state of the world, and situations, which describe a sequence of actions applied to an initial state. In the situation calculus, a collection of action precondition, successor state, foundational domain independent, and initial state axioms characterise the domain of an application. The constructs present in the Golog programming language are described in [de Giacomo et al., 2000; Levesque et al., 1997] and are: a (primitive action); φ? (test that φ holds in the current situation); δ 1 ; δ 2 (execute program δ 1 then δ 2 ); δ 1 δ 2 (choose to execute program δ 1 or δ 2 ); π v.δ (choice of arguments v in δ); δ (execute δ zero or more times); if φ then δ 1 else δ 2 (synchronised conditional); while φ do δ (synchronised loop); and proc P ( v)δ end (procedure). Given a domain theory, D, expressed in the situation calculus, a legal execution of a Golog program δ is a sequence of actions a such that: D = Do(δ, s o, do( a, s o)), where do( a, s o ) is a legal terminating situation (execution) of δ when executed in the situation s 0 1. Each pair (δ, s), where δ is a program to be executed in situation s, is called a configuration. The transition semantics for the interpretation of Golog programs [de Giacomo et al., 2000] (providing an implementation of Do) defines when an agent can transition from one configuration (δ, s) to another (δ, s ) by performing one step in the program δ, resulting in the situation s and the program δ left to perform. For each form a Golog program may assume, δ f, a Trans axiom is defined to determine which configurations can be transitioned to from (δ f, s) by performing a single step of δ f in situation s. A configuration (δ f, s) is final (Final(δ f, s)) if no steps remain to be performed in δ f. A selection of Trans and Final axioms are shown below 2. Given a program of the form δ f = a (where a is a primitive action), the configuration (δ f, s) can transition if it is possible to perform action a in situation s (P oss(a[s], s)), T rans(a, s, δ, s ) P oss(a[s], s) δ = nil s = do(a[s], s) and is not final as the action a remains to be performed. For the sequence δ f = δ 1 ; δ 2 (where δ 1 is performed first and then δ 2 ), the configuration (δ f, s) transitions as follows, T rans(δ 1 ; δ 2, s, δ, s ) γ.δ = (γ; δ 2 ) T rans(δ 1, s, γ, s ) F inal(δ 1, s) T rans(δ 2, s, δ, s ) and is final if both δ 1 and δ 2 are final in s: F inal(δ 1 ; δ 2, s) F inal(δ 1, s) F inal(δ 2, s) 1 do({a 1, a 2,..., a n 1, a n}, s) is shorthand for do(a n, do(a n 1,... do(a 2, do(a 1, s)))). 2 A full account of the transition semantics can be found in [de Giacomo et al., 2000]. 3 Related Work In rational choice theory, it is assumed that the preferences of a decision maker can be expressed as a utility function that assigns a value to each decision alternative. Utility maximisation produces a complete and transitive ranking from which a most preferred alternative can be selected. Golog and Preferences Developed in this vein is DTGolog a decision theoretic extension of Golog [Boutilier et al., 2000]. DTGolog a prominent approach for finding the most preferred executions of a Golog program combines the theory of Markov decision processes (MDPs) with the Golog programming language. Solving an MDP involves finding an optimal policy an optimal mapping between the states an agent might find itself in and the actions that should be performed in those states. DTGolog specifies an MDP with an action theory (expressed in the situation calculus) augmented with stochastic actions (actions with uncertain outcomes) and a reward function mapping numerical rewards to situations. A policy for a program δ is optimal if it maximises the expected reward or utility achieved over a given horizon (maximum number of actions performable), and the probability with which δ will be successfully executed. The incorporation of qualitative preferences in the interpretation of Golog programs [Fritz and McIlraith, 2006; Sohrabi et al., 2006] and planning in general [Son and Pontelli, 2004; Bienvenu et al., 2006] is an advantage when quantitative preferences (such as utilities or rewards) are not available or do not capture all of the preferences of an agent. Fritz and McIlraith extend DTGolog with qualitative preferences in [Fritz and McIlraith, 2006]. These preferences are expressed in terms of basic desire formulae (temporal logic constraints on situations) of varying priority. These formulae are compiled into a DTGolog program to be executed synchronously with the program under interpretation δ. This synchronous execution aims to find the most quantitatively preferred policies of δ amongst those that satisfy (if not all, at least some of) the qualitative constraints. Sohrabi et al develop GologPref [Sohrabi et al., 2006], a Golog interpreter designed to discover a most preferred web service composition using the best first search strategy of PPLAN [Bienvenu et al., 2006]. In [Bienvenu et al., 2006], user preferences are expressed in a language that extends PP [Son and Pontelli, 2004]. PP describes a hierarchy of preference constructs: basic desire formulae outlining temporal logic constraints on plan trajectories; atomic preferences specifying an ordering on basic desires according to their importance; and general preference formulae allowing the combination of atomic preferences with operators analogous to,, and. PPLAN introduces additional mechanisms of preference combination and assigns weights (denoting degrees of importance) to basic desires in atomic preference formulae. These weights are used to compute a value for each partial plan on a search frontier, where preferences not already violated are assumed to be satisfied at a future point on the trajectory of a partial plan. PPLAN selects the most promising partial plan for exploration at each stage in search. In contrast with the work of [Sohrabi et al., 2006], DT- Golog based approaches [Boutilier et al., 2000; Fritz and

3 McIlraith, 2006] consider the entire execution space in the search for one that is most preferred. This need arises, however, from their incorporation of stochastic actions. In this paper, we do not consider the use of stochastic actions. Limitations of Existing Work In [Delgrande et al., 2007], a distinction between absolute preference (describing the desirability of a particular state of affairs) and comparative preference (describing the desirability of one state of affairs relative to others) is made. The approaches described above support only absolute preference, and disallow those that imply the merit of one partial plan or execution is dependent on the alternatives it is being compared with. In addition to the varieties of preference that can often only be described comparatively such as preference over tastes and perception preference over alternatives involving multiple attributes is often expressed in terms of comparative rules to reduce the cognitive load on a decision maker [Bar-Hillel and Margalit, 1988]. An example is the application of a majority rule over multiple attributes. In this scenario, one alternative is preferred to another if it is better on a majority of attributes. Approximation methods used to ascertain a preference for one alternative over another such as the majority rule described above, and the cancelling out of dimensions with similar differences [Tversky, 1969] are common in human decision-making [Tversky, 1969]. Argumentation and Best First Search Argumentation techniques provide a general means by which preference between alternatives can be specified or ascertained. In argumentation-based decision-making, preference amongst alternatives is constructed as a part of the decision-making process, and is based on the competing interests, motivations, values, or goals, of the decision maker. Given a set of program executions, reasons may be formed in support of preferences between its members based on the variety of principles an agent may use to compare them. In this paper, we combine the advantages of argumentation as a tool for expressing complex preferences, and best first search a technique applied in existing work to reduce the cost of search for a most preferred execution. In place of independently evaluating partial executions with respect to a set of absolute preferences, arguments supporting preferences between partial executions are used to guide the search. 4 Argumentation over Executions In this section, we define an argumentation framework for decision-making (ADF) that constructs preference relationships between decision alternatives given a general set of principles by which they are compared. We apply this framework, in conjunction with choice theory, to find the most preferred subset of an execution set A for a Golog program δ. We first provide the following definition of a comparison principle, and assume that an agent a = C, I is equipped with a finite set of such principles C, and a (transitive, irreflexive, and asymmetric) partial ordering I over the principles in C. Given c 1, c 2 C, I(c 1, c 2 ) denotes that principle c 1 is more important to a than principle c 2. Definition 1 Given decision alternatives d 1, d 2 D, a comparison principle is a relation C(d 1, d 2, φ) where φ is a conditional expression involving d 1 and d 2. If φ(d 1, d 2 ) holds, we say that d 1 is preferred to d 2 (P(d 1, d 2 )). We now introduce a running example that will used throughout the remainder of this paper. Example 1 (The Travelling Agent) An agent a has the following program (denoted δ) outlining what to do to travel from point A to B to C. Agent a can drive or walk from A to B, and then take the train (preceded by eating cake) or skip (followed by eating fruit) from B to C. proc travel(a, B, C) [drive(a, B) walk(a, B)] ; [[eat(cake); train(b, C)] [skip(b, C); eat(f ruit)]] end We make the assumption that all actions are possible in any situation, with the exception of taking the train which is not possible in a situation that has involved driving. Given an initial situation s 0, A = home (h), B = university (u), C = cinema (c), the program δ has three executions: s 1 = do({drive(h, u), skip(u, c), eat(fruit)}, s 0 ), s 2 = do({walk(h, u), eat(cake), train(u, c)}, s 0 ), s 3 = do({walk(h, u), skip(u, c), eat(fruit)}, s 0 ). Agent a has three principles C i (s j, s k, φ i ) by which it compares these executions. By the first comparison principle, agent a prefers one execution over another if it involves eating tastier food. φ 1 (do( a 1, s 0 ), do( a 2, s 0 )) eat(x) a 1 eat(y) a 2 x > taste y where x > taste y denotes that x tastes better than y. The second principle expresses that walking and skipping together is preferred to driving. φ 2 (do( a 1, s 0 ), do( a 2, s 0 )) walk(x i, y i ) a 1 skip(x j, y j ) a 1 drive(x k, y k ) a 2 By the third principle, agent a prefers executions involving less tiring activities. φ 3 (do( a 1, s 0 ), do( a 2, s 0 )) x a 1 y a 2 x < tiring y where x < tiring y denotes that x is less tiring than y. There are several existing frameworks that have been developed to argumentatively determine the best decisions in an alternative set based on the preferences of an agent (see [Ouerdane et al., 2007] for a review). Each of these approaches is designed to induce a preference ordering over a set of available decisions, from which the best (most preferred) decisions can be inferred. This preference ordering is formed by analysing and comparing a set of arguments (pro and con each decision, for example) constructed in light of the positive and negative aspects of each decision. Our aim is to provide an agent with the flexibility to compare executions with whichever principles it desires. Rather than construct arguments in support of and against individual

4 decisions (pros and cons), comparison principles are instantiated in our ADF to form arguments in support of preference relationships between decisions (executions). Each argument in our ADF represents a reason why one execution s i A is preferred to another s j A (P(s i, s j )). Each argument takes the form J, P(s i, s j ) where s i, s j A and J is a justification for the truth of the conclusion P(s i, s j ). Each justification J represents the instantiation of a comparison principle c C used by an agent to determine if one execution is preferred to another. An argument a 1 = J 1, P(s i, s j ) AR is attacked (rebutted) by all arguments with an opposing conclusion a 2 = J 2, P(s j, s i ) AR (denoted attacks(a 2, a 1 )). An argument a 1 defeats an argument a 2 iff attacks(a 1, a 2 ) holds and the principle that derives a 2 is not more important to the decision maker (agent) than the principle that derives a 1. The structure of our ADF is based on the structure of an audience-specific value-based argumentation framework [Bench-Capon, 2003]. The conclusions of undefeated arguments (arguments not defeated by another argument) in an ADF represent a strict preference profile (a dominance graph) P over the available decisions (executions). Definition 2 An ADF is a 5-tuple: ADF = AR, attacks, C, pri, I where AR is a finite set of arguments, attacks is an irreflexive binary attack relation on AR, C is a finite set of comparison principles, pri maps each argument in AR to the principle in C it is derived from, and I C C is a (transitive, irreflexive, and asymmetric) partial ordering of the principles in C. An argument a AR defeats an argument b AR iff attacks(a, b) and I(pri(b), pri(a)) holds. We assume that an agent has its own choice rule (Ch) for selecting a most preferred subset of A (MP S(A)) based on the preferences that have been found to exist between its elements (Ch(A, P) = MP S(A)). As preference is based on a set of general comparison principles, we do not assume that P is transitive or acyclic. An agent may, for example, represent Ch(A, P) with the Schwartz set [Schwartz, 1972]. Definition 3 Given a set of alternatives D, and a dominance graph over D, P: (i) an alternative γ D dominates an alternative β D iff (γ, β) P; (ii) a Schwartz set component of D is a subset S D such that: (a) no alternative γ D where γ / S dominates an alternative β S; and (b) no non-empty subset S S satisfies property (a); and (iii) the Schwartz set Sc of D is the union of all Schwartz set components of D. If a preference cycle exists between decision alternatives, the Schwartz set infers that they are each at least as good as each other [Schwartz, 1972]. Example 2 (The Travelling Agent cont.) For agent a: fruit > taste cake, drive < tiring walk, and I = {(C 1, C 2 ), (C 1, C 3 ), (C 2, C 3 )}. In an ADF defined over A = {s 1, s 2, s 3 }, a is able to instantiate its comparison principles to produce five arguments in support of preference relationships between elements of A (AR = {A i } i=1,...,5 ). Argument 1, A 1 = C 1, P(s 1, s 2 ), denotes that s 1 is preferred to s 2 as it involves eating tastier food. Similarly, A 2 = C 1, P(s 3, s 2 ). Argument 3, A 3 = C 2, P(s 3, s 1 ), denotes that s 3 is preferred to s 1 as s 3 involves walking and skipping while s 1 involves driving. Argument 4, A 4 = C 3, P(s 1, s 2 ), denotes that s 1 is preferred to s 2 as it is less tiring. Similarly, A 5 = C 3, P(s 1, s 3 ). By Definition 2, A 3 and A 5 attack each other with A 3 defeating A 5 (as I(C 3, C 2 )), and P = {(s 1, s 2 ), (s 3, s 2 ), (s 3, s 1 )}. Given the Schwartz set (Definition 3) as Ch, the most preferred subset of A, MP S(A), is Ch(A, P) = {s 3 }. 5 Argumentation-Based Best First Search In this section, we use the transition semantics defined in [de Giacomo et al., 2000] to characterise the interpretation of a Golog program as best first search. Given a program δ to interpret (and an initial situation s 0 ), the state-space of the problem is the set of configurations (δ i, s i ) that can be transitioned to from (δ, s 0 ) by performing a sequence of actions in δ. Given a configuration (δ i, s i ), a legal moves function MOVES(δ i, s i ) returns the set of configurations that may be transitioned to from (δ i, s i ) by performing the first step of δ i in the situation s i. Search proceeds by building a tree (with root node (δ, s 0 )), whose paths represent partial or complete executions of δ. For each fringe (set of leaf configurations) of this tree, Fringe, an evaluation function EVAL(Fringe, P) selects which configuration in Fringe to transition to next given a dominance graph P over Fringe. If a final configuration (δ r, s r ) is selected by EVAL, the execution represented by this configuration s r forms the result of search. Our aim is to allow an agent a = C, I to discover a final configuration representing an execution in M P S(A) where A are the executions of δ given s 0. Algorithm 1 outlines the program interpretation process, where: EXPAND(Fringe, γ) removes the configuration γ = (δ i, s i ) from the search fringe and replaces it with MOVES(δ i, s i ); ARGUE-ADF(Fringe, C, I) determines the undefeated conclusions P of an ADF defined over the configurations in Fringe; and EVAL(Fringe, P) selects a configuration in Fringe to transition to next based on the dominance graph P. Algorithm 1 PROGRAM INTERPRETATION PROCESS Input: (δ, s 0 ), C, I, Output: Best Best (δ, s 0 ), Fringe {(δ, s 0 )} while Final(Best) do Fringe EXPAND(Fringe, Best) P ARGUE-ADF(Fringe, C, I) Best EVAL(Fringe, P) end while ARGUE-ADF(Fringe, C, I) Each fringe in the search tree constructed by EXPAND consists of a set of configurations (δ i, s i ) where each (δ i, s i ) is reachable from (δ, s 0 ) by performing a sequence of actions in δ. ARGUE-ADF constructs a dominance graph over the configurations in a fringe, with the objective of providing advice to EVAL in its selection of the next configuration to transition to (expand). An agent a = C, I determines P by forming an ADF (Definition 2) over the configurations in Fringe. Each argument in the ADF expresses a preference for one configuration γ 1 Fringe over another γ 2 Fringe (based on the ex-

5 ecutions or partial executions they represent), and describes a reason why γ 1 is more likely to lie on a path to a most preferred final configuration (execution) than γ 2. If arguments in the ADF over Fringe are based solely on actions that have been performed on a path from (δ, s 0 ) to a configuration in Fringe, the search for a most preferred execution of δ becomes greedy. To achieve admissibility, we must ensure that an agent can construct an argument preferring one configuration γ 1 Fringe to another γ 2 Fringe if, according to C, γ 1 leads to a configuration that is preferred to γ 2. We do not want an agent to search ahead in a program to determine which actions are actually possible, as this would defeat the purpose of best first search. Instead, we allow an agent to speculate on (estimate) which actions could be performed after making a transition, and construct arguments accordingly. To formally define a speculation, we introduce a relaxation of the transition semantics of [de Giacomo et al., 2000]. In this relaxed semantics, a TransR axiom is defined for each of the constructs in Section 2. This semantics is defined such that: if a configuration γ can transition to a configuration γ by Trans, γ can transition to γ by TransR; and if a configuration γ cannot transition to a configuration γ by Trans, it may transition to γ by TransR. The relaxed semantics overestimates what is possible in a program. For each construct defined in Section 2, with the exception of primitive actions, TransR and Trans are equivalent. Given a primitive action δ = a, (δ, s) transitions as follows, T ransr(a, s, δ, s ) δ = nil s = do(a[s], s) and differs from its Trans counterpart in that it does not check if an action is possible. Definition 4 (Speculation) For a configuration γ = (δ, s ), γ s = (δ s, s s ) is a speculation of γ if it is final (F inal(δ s, s s )) and a member of the reflexive, transitive closure of TransR (TransR ) where TransR (δ, s, δ s, s s ) holds if (δ s, s s ) is reachable from (δ, s ) by repeated application of TransR 3. With the introduction of speculations the set of arguments AR in the ADF defined over a fringe is formed as follows: for each pair of configurations on a fringe (γ 1, γ 2 ) (where both γ 1 and γ 2 are final or non-final, or γ 1 is non-final and γ 2 final), an argument c, P(γ 1, γ 2 ) (as described in Section 4) is constructed in support of P(γ 1, γ 2 ) if a speculation of γ 1 is preferred to γ 2 according to a comparison principle c C. EVAL(Fringe, P) EVAL computes a set E MP S representing the configurations in Fringe an agent estimates to lie on a path from (δ, s 0 ) to a final configuration (δ, s ) where s MP S(A) and A are the executions of δ. E MP S is the result of applying a choice rule Ch to the dominance graph P induced by an ADF over Fringe in ARGUE-ADF: E MP S def = Ch(Fringe, P) Fringe EVAL selects a configuration in Fringe to transition to next, using the advice in E MP S. As P is based on speculations that may not represent legally reachable configurations, some 3 We consider only Golog programs that admit a finite number of executions, and for which there is no situation in which iteration and loop constructs admit an infinite number of iterations. of the preferences in P are transient. A transient preference between configurations γ 1 and γ 2 does not exist between all configurations reachable from γ 1 and all those reachable from γ 2, limiting an agent s ability to predict the nature of a P over all final program configurations from P over a search fringe. To address problems arising from transient preferences, EVAL is implemented using prioritised configuration selection. This prioritisation is based on the following assumption regarding the nature of all choice rules an agent may adopt. Assumption 1 Given a non-empty set of alternatives D, and a dominance graph over those alternatives P, any choice rule Ch adopted by an agent: (i) finds a non-empty subset of D where Ch(D, P) contains (at least) all alternatives γ D for which there is no alternative γ D such that P(γ, γ); and (ii) satisfies the property that an alternative γ D being in Ch(D, P) is independent of any alternative γ D not connected to γ by an undirected path in P. Assumption 1 provides an agent with a stopping condition an indication of when a final configuration in the E MP S of a fringe represents a most preferred execution. The Prioritisation Algorithm Given a search fringe Fringe, a dominance graph P over Fringe, and the set E MP S = Ch(Fringe, P), EVAL(Fringe, P) is defined to: (i) Select a final configuration γ E MP S for which γ Fringe. P(γ, γ), if one exists; (ii) If (i) is not satisfied, select a final configuration γ E MP S which is not connected by an undirected path of preferences in P to a non-final configuration γ Fringe, if one exists; (iii) If (i)-(ii) are not satisfied, select a non-final configuration γ E MP S, if one exists; and (iv) If (i)-(iii) are not satisfied, select any non-final configuration γ Fringe. Steps (i) and (ii) specify the conditions that must be satisfied by a final configuration to demonstrate that it represents an execution in M P S(A). Transient preferences can only exist between configuration pairs in which at least one member is non-final. If a final configuration γ Fringe is not independent of all members of such pairs (according to Assumption 1), an agent cannot predict the status of the execution it represents when a P is constructed over all executions in A. In this case, steps (iii)-(iv) tell the agent to continue its search. Theorem 1 Given an agent a = C, I, using a choice rule satisfying Assumption 1, any execution s r A of a program δ, given an initial situation s 0, found as a result of argumentation-based best first search is in MP S(A). Proof : We demonstrate that a final configuration γ = (δ r, s r ) of δ cannot be selected from any search fringe Fringe, given a dominance graph over its members P, if s r / MP S(A). EVAL, using the prioritisation algorithm, will only select γ Ch(Fringe, P) as a result of search if: (i) there is no configuration γ Fringe such that P(γ, γ); or (ii) there is no non-final configuration γ Fringe where γ and γ are connected by an undirected path in P. If (i) holds, then there is no undefeated argument (based on a speculation or not) that can be formed in support of P(γ, γ)

6 for γ Fringe. As speculation over-estimates what is possible after expanding a configuration, there is no final configuration of δ given s 0 that is preferred to γ. According to part (i) of Assumption 1, s r must be a member of MP S(A). Let U be the set of configurations in Fringe connected to γ by an undirected path in P. If (ii) holds, each configuration in U is final, and no configuration in U is connected to a non-final configuration by an undirected path in P. Thus, there is no argument (based on a speculation) in support of a preference between a configuration γ Fringe and a configuration in U. As speculation over-estimates what is possible, there are no final configurations of δ outside of U that could be connected to γ (or any configuration in U) by an undirected path in P. By part (ii) of Assumption 1, the question of whether s r is in MP S(A) is independent of any execution in A represented by a configuration outside of U. Let U be the set of executions represented by configurations in U. As U A, and s r is in MP S(U ), s r must be in MP S(A). Moreover, the assumption that Ch(Fringe, P) is nonempty for a non-empty Fringe (Assumption 1) ensures that if a program δ has at least one legal execution, argumentationbased best first search is guaranteed to find a result. Example 3 (The Travelling Agent cont.) In the first step of the interpretation of δ in Example 1: Fringe = {γ 1 = (δ, s 1), γ 2 = (δ, s 2)} where s 1 = do(drive(h, u), s 0 ), s 2 = do(walk(h, u), s 0 ), and δ = [eat(cake) ; train(u, c)] [skip(u, c) ; eat(fruit)]. The speculations that can be formed of (δ, s 1) and (δ, s 2) are {sp 1, sp 2 } and {sp 3, sp 4 }, respectively: sp 1 = (nil, do(train(u, c), do(eat(cake), s 1))) sp 2 = (nil, do(eat(fruit), do(skip(u, c), s 1))) sp 3 = (nil, do(train(u, c), do(eat(cake), s 2))) sp 4 = (nil, do(eat(fruit), do(skip(u, c), s 2))) In an ADF defined over Fringe (given C and I as described in Examples 1 and 2), AR = {A i } i=1,2, where A 1 = C 3, P(γ 1, γ 2 ) and A 2 = C 2, P(γ 2, γ 1 ). A 1 is based on principle C 3 (driving is less tiring than walking). A 2 is based on sp 4 and principle C 2 (walking and skipping together is preferred over driving). A 2 defeats A 1 (by Definition 2 and I(C 3, C 2 )), and P = {(γ 2, γ 1 )}. Given the Schwartz set [Schwartz, 1972] as Ch, EVAL selects walk as the first step (E MP S = Ch(Fringe, P) = {γ 2 } and γ Fringe.P(γ, γ 2 )), resulting in the fringe: Fringe = {γ 1 = (δ, s 1), γ 3 = (train(u, c), do(eat(cake), s 2)), γ 4 = (eat(fruit), do(skip(u, c), s 2))} Performing the ARGUE-ADF step on the new fringe results in a new set of arguments AR = {A i } i=1,...,5 where: A 1 = C 1, P(γ 4, γ 3 ) (based on sp 4 ), A 2 = C 2, P(γ 4, γ 1 ) (based on sp 4 ), A 3 = C 3, P(γ 1, γ 3 ) (based on sp 1 and sp 2 ), A 4 = C 3, P(γ 1, γ 4 ) (based on sp 1 and sp 2 ), and A 5 = C 1, P(γ 1, γ 3 ) (based on sp 2 ). The conclusions of undefeated arguments in AR form the dominance graph P = {(γ 4, γ 3 ), (γ 4, γ 1 ), (γ 1, γ 3 )}. EVAL selects the configuration γ 4 to transition to, replacing γ 4 on the fringe with sp 4. Repeating the ARGUE-ADF and EVAL steps given this new fringe results in the execution represented by sp 4 being chosen as the final result. In Example 2, MP S(A = {s 1, s 2, s 3 }) = {s 3 }. In this example, sp 4 = (nil, s 3 ). 6 Conclusions This paper has presented an argumentation-based best first search algorithm for the interpretation of Golog programs. Argumentation allows the flexible specification (and elicitation) of preferences (including comparative preference), while the incorporation of best first search reduces the execution space traversed in finding a most preferred result. Our approach uses the concept of speculation (a tool for predicting future opportunities at each stage of search), and arguments in support of preference between configurations, to guide search toward a most preferred program execution. The proposed argumentation-based interpretation of Golog programs is shown to be admissible. As future work, we intend to consider the implicit speculation of programs, in place of the explicit generation of speculations by a TransR relation. References [Bar-Hillel and Margalit, 1988] M. Bar-Hillel and A. Margalit. Theory and Decision, 24: , [Bench-Capon, 2003] T. J. M. Bench-Capon. Persuasion in practical argument using value based argumentation frameworks. J. of Log. Comp., 13: , [Bienvenu et al., 2006] M. Bienvenu, C. Fritz, and S. McIlraith. Planning with Qualitative Temporal Preferences. In KR, pages , [Boutilier et al., 2000] C. Boutilier, R. Reiter, M. Soutchanski, and S. Thrun. Decision-Theoretic, High-Level Agent Programming in the Situation Calculus. In AAAI, [de Giacomo et al., 2000] G. de Giacomo, Y. Lespérance, and H. J. Levesque. ConGolog, a concurrent programming language based on the situation calculus. Artificial Intelligence, 121(1-2): , [Delgrande et al., 2007] James Delgrande, Torsten Schaub, and Hans Tompits. J. of Log. Comp., 17(5): , [Fritz and McIlraith, 2006] Christian Fritz and Sheila McIlraith. Decision-Theoretic Golog with Qualitative Preferences. In KR, pages , [Levesque et al., 1997] Hector Levesque, Ray Reiter, Yves Lespérance, Fangzhen Lin, and Richard B. Scherl. Golog: A logic programming language for dynamic domains. J. of Log. Prog., 31(1-3):59 83, [Ouerdane et al., 2007] W. Ouerdane, N. Maudet, and A. Tsoukias. Arguing over Actions That Involve Multiple Criteria: A Critical Review. In ECSQARU, [Schwartz, 1972] T. Schwartz. Rationality and the myth of the maximum. Noûs, 6(2):97 117, [Sohrabi et al., 2006] S. Sohrabi, N. Prokoshyna, and S. McIlraith. In ISWC, pages , [Son and Pontelli, 2004] T. Son and E. Pontelli. Planning with preferences using logic programming. In LPNMR, pages , [Tversky, 1969] A. Tversky. Intransitivity of Preferences. Psychological Review, 76(1):31 48, 1969.

Reconciling Situation Calculus and Fluent Calculus

Reconciling Situation Calculus and Fluent Calculus Reconciling Situation Calculus and Fluent Calculus Stephan Schiffel and Michael Thielscher Department of Computer Science Dresden University of Technology {stephan.schiffel,mit}@inf.tu-dresden.de Abstract

More information

On Ability to Autonomously Execute Agent Programs with Sensing

On Ability to Autonomously Execute Agent Programs with Sensing On Ability to Autonomously Execute Agent Programs with Sensing Sebastian Sardiña Dept. of Computer Science University of Toronto ssardina@cs.toronto.edu Yves Lespérance Dept. of Computer Science York University

More information

Interpreting Golog Programs in Flux

Interpreting Golog Programs in Flux Interpreting Golog Programs in Flux Stephan Schiffel and Michael Thielscher Department of Computer Science Dresden University of Technology {stephan.schiffel,mit}@inf.tu-dresden.de Abstract A new semantics

More information

Online Agent Supervision in the Situation Calculus

Online Agent Supervision in the Situation Calculus Online Agent Supervision in the Situation Calculus Bita Banihashemi York University Toronto, ON, Canada bita@cse.yorku.ca Giuseppe De Giacomo Sapienza Università di Roma Roma, Italy degiacomo@dis.uniroma1.it

More information

Resolving Non-Determinism in Programs for Complex Task Planning with Search Control

Resolving Non-Determinism in Programs for Complex Task Planning with Search Control Resolving Non-Determinism in Programs for Complex Task Planning with Search Control Alfredo Gabaldon National ICT Australia, School of CS and Eng., UNSW Sydney, Australia, alfredo@cse.unsw.edu.au Abstract

More information

Tractable Reasoning with Incomplete First-Order Knowledge in Dynamic Systems with Context-Dependent Actions

Tractable Reasoning with Incomplete First-Order Knowledge in Dynamic Systems with Context-Dependent Actions Tractable Reasoning with Incomplete First-Order Knowledge in Dynamic Systems with Context-Dependent Actions Yongmei Liu and Hector J. Levesque Department of Computer Science University of Toronto Toronto,

More information

Argumentation-Based Models of Agent Reasoning and Communication

Argumentation-Based Models of Agent Reasoning and Communication Argumentation-Based Models of Agent Reasoning and Communication Sanjay Modgil Department of Informatics, King s College London Outline Logic and Argumentation - Dung s Theory of Argumentation - The Added

More information

CS 7180: Behavioral Modeling and Decisionmaking

CS 7180: Behavioral Modeling and Decisionmaking CS 7180: Behavioral Modeling and Decisionmaking in AI Markov Decision Processes for Complex Decisionmaking Prof. Amy Sliva October 17, 2012 Decisions are nondeterministic In many situations, behavior and

More information

On the Role of Possibility in Action Execution and Knowledge in the Situation Calculus

On the Role of Possibility in Action Execution and Knowledge in the Situation Calculus On the Role of Possibility in Action Execution and Knowledge in the Situation Calculus Vahid Vaezian (B) and James P. Delgrande School of Computing Science, Simon Fraser University, Burnaby, BC V5A 1S6,

More information

CS 4100 // artificial intelligence. Recap/midterm review!

CS 4100 // artificial intelligence. Recap/midterm review! CS 4100 // artificial intelligence instructor: byron wallace Recap/midterm review! Attribution: many of these slides are modified versions of those distributed with the UC Berkeley CS188 materials Thanks

More information

Lecture 3: High-level Programming in the Situation Calculus: Golog and ConGolog

Lecture 3: High-level Programming in the Situation Calculus: Golog and ConGolog DIS La Sapienza, PhD Course: Reasoning about Action and High-Level Programs Lecture 3: High-level Programming in the Situation Calculus: Golog and ConGolog Yves Lespérance Dept. of Computer Science & Engineering,

More information

Towards the Integration of Programming by Demonstration and Programming by Instruction using Golog

Towards the Integration of Programming by Demonstration and Programming by Instruction using Golog Towards the Integration of Programming by Demonstration and Programming by Instruction using Golog Christian Fritz and Yolanda Gil Information Sciences Institute University of Southern California Marina

More information

Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events

Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events Massimo Franceschet Angelo Montanari Dipartimento di Matematica e Informatica, Università di Udine Via delle

More information

Modelling PRS-Like Agents Mental States

Modelling PRS-Like Agents Mental States Modelling PRS-Like Agents Mental States Wayne Wobcke School of Computer Science and Engineering University of New South Wales Sydney NSW 2052, Australia wobcke@cse.unsw.edu.au Abstract. In recent years,

More information

On the Decidability of Verifying LTL Properties of GOLOG Programs

On the Decidability of Verifying LTL Properties of GOLOG Programs On the Decidability of Verifying LTL Properties of GOLOG Programs Benjamin Zarrieß Theoretical Computer Science TU Dresden, Germany zarriess@tcs.inf.tu-dresden.de Jens Claßen Knowledge-Based Systems Group

More information

ESSENCE 2014: Argumentation-Based Models of Agent Reasoning and Communication

ESSENCE 2014: Argumentation-Based Models of Agent Reasoning and Communication ESSENCE 2014: Argumentation-Based Models of Agent Reasoning and Communication Sanjay Modgil Department of Informatics, King s College London Outline Logic, Argumentation and Reasoning - Dung s Theory of

More information

Planning in Markov Decision Processes

Planning in Markov Decision Processes Carnegie Mellon School of Computer Science Deep Reinforcement Learning and Control Planning in Markov Decision Processes Lecture 3, CMU 10703 Katerina Fragkiadaki Markov Decision Process (MDP) A Markov

More information

Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events

Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events Pairing Transitive Closure and Reduction to Efficiently Reason about Partially Ordered Events Massimo Franceschet Angelo Montanari Dipartimento di Matematica e Informatica, Università di Udine Via delle

More information

DIS La Sapienza PhD Course. Autonomous Agents and Multiagent Systems. Lecture 4: High-Level Programming in the Situation Calculus: Golog and ConGolog

DIS La Sapienza PhD Course. Autonomous Agents and Multiagent Systems. Lecture 4: High-Level Programming in the Situation Calculus: Golog and ConGolog DIS La Sapienza PhD Course Autonomous Agents and Multiagent Systems Lecture 4: High-Level Programming in the Situation Calculus: Golog and ConGolog Yves Lespérance Dept. of Computer Science & Engineering

More information

On the Limits of Planning over Belief States under Strict Uncertainty

On the Limits of Planning over Belief States under Strict Uncertainty On the Limits of Planning over Belief States under Strict Uncertainty Sebastian Sardina Dept. of Comp. Sci. RMIT University Melbourne, AUSTRLIA ssardina@cs.rmit.edu.au Giuseppe De Giacomo Dipart. di Infor.

More information

Predictive Timing Models

Predictive Timing Models Predictive Timing Models Pierre-Luc Bacon McGill University pbacon@cs.mcgill.ca Borja Balle McGill University bballe@cs.mcgill.ca Doina Precup McGill University dprecup@cs.mcgill.ca Abstract We consider

More information

Continuing the KRR Journey

Continuing the KRR Journey 1 Weeks 1 3 Week 4 Week 5 Week 6 Continuing the KRR Journey Propositional and first-order logic (FOL) Expressiveness and Decidability: Propositional logic is decidable, FOL is not. Classical planning problems:

More information

Chapter 16 Planning Based on Markov Decision Processes

Chapter 16 Planning Based on Markov Decision Processes Lecture slides for Automated Planning: Theory and Practice Chapter 16 Planning Based on Markov Decision Processes Dana S. Nau University of Maryland 12:48 PM February 29, 2012 1 Motivation c a b Until

More information

Expressing Transactions with Savepoints as Non-Markovian Theories of Actions

Expressing Transactions with Savepoints as Non-Markovian Theories of Actions Expressing Transactions with Savepoints as Non-Markovian Theories of Actions Iluju Kiringa School of Info. Technology and Eng. University of Ottawa Alfredo Gabaldon Department of Computer Science University

More information

CPP: A Constraint Logic Programming Based Planner with Preferences

CPP: A Constraint Logic Programming Based Planner with Preferences CPP: A Constraint Logic Programming Based Planner with Preferences Phan Huy Tu, Tran Cao Son, and Enrico Pontelli Computer Science Department, New Mexico State University, Las Cruces, New Mexico, USA {tphan

More information

Intention recognition in the Situation Calculus and Probability Theory frameworks

Intention recognition in the Situation Calculus and Probability Theory frameworks Intention recognition in the Situation Calculus and Probability Theory frameworks Robert Demolombe 1 and Ana Mara Otermin Fernandez 2 1 OERA Toulouse France Robert.Demolombe@cert.fr 2 OERA Toulouse France

More information

Abstract Dialectical Frameworks

Abstract Dialectical Frameworks Abstract Dialectical Frameworks Gerhard Brewka Computer Science Institute University of Leipzig brewka@informatik.uni-leipzig.de joint work with Stefan Woltran G. Brewka (Leipzig) KR 2010 1 / 18 Outline

More information

A Closed-Form Solution to the Ramification Problem (Sometimes)

A Closed-Form Solution to the Ramification Problem (Sometimes) A losed-form Solution to the Ramification Problem (Sometimes) Sheila A McIlraith Department of omputer Science University of Toronto Toronto M5S 3G4 anada E-mail: mcilrait@csutorontoca October 19 1997

More information

Designing and Evaluating Generic Ontologies

Designing and Evaluating Generic Ontologies Designing and Evaluating Generic Ontologies Michael Grüninger Department of Industrial Engineering University of Toronto gruninger@ie.utoronto.ca August 28, 2007 1 Introduction One of the many uses of

More information

Applying DTGolog to Large-scale Domains

Applying DTGolog to Large-scale Domains Applying DTGolog to Large-scale Domains by Huy N Pham A thesis presented to Ryerson University in partial fulfillment of the requirements for the degree of Masters of Applied Science in the Program of

More information

A Preference Logic With Four Kinds of Preferences

A Preference Logic With Four Kinds of Preferences A Preference Logic With Four Kinds of Preferences Zhang Zhizheng and Xing Hancheng School of Computer Science and Engineering, Southeast University No.2 Sipailou, Nanjing, China {seu_zzz; xhc}@seu.edu.cn

More information

Non-Markovian Control in the Situation Calculus

Non-Markovian Control in the Situation Calculus Non-Markovian Control in the Situation Calculus An Elaboration Niklas Hoppe Seminar Foundations Of Artificial Intelligence Knowledge-Based Systems Group RWTH Aachen May 3, 2009 1 Contents 1 Introduction

More information

A Non-Monotonic Goal Specification Language for Planning with Preferences

A Non-Monotonic Goal Specification Language for Planning with Preferences A Non-Monotonic Goal Specification Language for Planning with Preferences Tran Cao Son and Enrico Pontelli and Chitta Baral +1 Abstract. This paper introduces a default logic based approach to defining

More information

Projection using Regression and Sensors

Projection using Regression and Sensors Projection using Regression and Sensors Giuseppe De Giacomo Dipartimento di Informatica e Sistemistica Università di Roma La Sapienza Via Salaria 11, 00198 Roma, Italy degiacomo@dis.uniroma1.it Hector

More information

Course 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016

Course 16:198:520: Introduction To Artificial Intelligence Lecture 13. Decision Making. Abdeslam Boularias. Wednesday, December 7, 2016 Course 16:198:520: Introduction To Artificial Intelligence Lecture 13 Decision Making Abdeslam Boularias Wednesday, December 7, 2016 1 / 45 Overview We consider probabilistic temporal models where the

More information

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm

CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm NAME: SID#: Login: Sec: 1 CS 188 Introduction to Fall 2007 Artificial Intelligence Midterm You have 80 minutes. The exam is closed book, closed notes except a one-page crib sheet, basic calculators only.

More information

Composition of ConGolog Programs

Composition of ConGolog Programs Composition of ConGolog Programs Sebastian Sardina 1 Giuseppe De Giacomo 2 1 Department of Computer Science and Information Technology RMIT University, Melbourne, AUSTRALIA 2 Dipartimento di Informatica

More information

An Action Description Language for Iterated Belief Change

An Action Description Language for Iterated Belief Change An Action Description Language for Iterated Belief Change Aaron Hunter and James P. Delgrande Simon Fraser University Burnaby, BC, Canada {amhunter, jim}@cs.sfu.ca Abstract We are interested in the belief

More information

Online agent supervision in the situation calculus - Extended version

Online agent supervision in the situation calculus - Extended version Online agent supervision in the situation calculus - Extended version Bita Banihashemi, Giuseppe De Giacomo, and Yves Lespérance Technical Report EECS-2016-02 July 12 2016 Department of Electrical Engineering

More information

Argumentation with Abduction

Argumentation with Abduction Argumentation with Abduction Neophytos Demetriou Dept. of Computer Science University of Cyprus Nicosia CY-1678, Cyprus nkd@cs.ucy.ac.cy Tel.: (+357) 22892673 Antonis Kakas Dept. of Computer Science University

More information

On the Progression of Situation Calculus Basic Action Theories: Resolving a 10-year-old Conjecture

On the Progression of Situation Calculus Basic Action Theories: Resolving a 10-year-old Conjecture On the Progression of Situation Calculus Basic Action Theories: Resolving a 10-year-old Conjecture Stavros Vassos and Hector J Levesque Department of Computer Science University of Toronto Toronto, ON,

More information

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes.

Final. Introduction to Artificial Intelligence. CS 188 Spring You have approximately 2 hours and 50 minutes. CS 188 Spring 2014 Introduction to Artificial Intelligence Final You have approximately 2 hours and 50 minutes. The exam is closed book, closed notes except your two-page crib sheet. Mark your answers

More information

Iterated Belief Change: A Transition System Approach

Iterated Belief Change: A Transition System Approach Iterated Belief Change: A Transition System Approach Aaron Hunter and James P. Delgrande School of Computing Science Simon Fraser University Burnaby, BC, Canada {amhunter, jim}@cs.sfu.ca Abstract We use

More information

Searching in non-deterministic, partially observable and unknown environments

Searching in non-deterministic, partially observable and unknown environments Searching in non-deterministic, partially observable and unknown environments CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2014 Soleymani Artificial Intelligence:

More information

Design of Distributed Systems Melinda Tóth, Zoltán Horváth

Design of Distributed Systems Melinda Tóth, Zoltán Horváth Design of Distributed Systems Melinda Tóth, Zoltán Horváth Design of Distributed Systems Melinda Tóth, Zoltán Horváth Publication date 2014 Copyright 2014 Melinda Tóth, Zoltán Horváth Supported by TÁMOP-412A/1-11/1-2011-0052

More information

Dialectical Frameworks: Argumentation Beyond Dung

Dialectical Frameworks: Argumentation Beyond Dung Dialectical Frameworks: Argumentation Beyond Dung Gerhard Brewka Computer Science Institute University of Leipzig brewka@informatik.uni-leipzig.de joint work with Stefan Woltran G. Brewka (Leipzig) NMR

More information

Abstraction of Agents Executing Online and their Abilities in the Situation Calculus

Abstraction of Agents Executing Online and their Abilities in the Situation Calculus Abstraction of Agents Executing Online and their Abilities in the Situation Calculus Bita Banihashemi 1, Giuseppe De Giacomo 2, Yves Lespérance 1 1 York University 2 Sapienza Università di Roma bita@cse.yorku.ca,

More information

On Planning with Preferences in HTN

On Planning with Preferences in HTN On Planning with Preferences in HTN Shirin Sohrabi and Sheila A. McIlraith Department of Computer Science University of Toronto Toronto, Canada. {shirin,sheila}@cs.toronto.edu Abstract In this paper, we

More information

Markov Decision Processes Chapter 17. Mausam

Markov Decision Processes Chapter 17. Mausam Markov Decision Processes Chapter 17 Mausam Planning Agent Static vs. Dynamic Fully vs. Partially Observable Environment What action next? Deterministic vs. Stochastic Perfect vs. Noisy Instantaneous vs.

More information

The Situation Calculus and Golog

The Situation Calculus and Golog and A Tutorial Gerhard Lakemeyer Dept. of Computer Science RWTH Aachen University Germany 2nd Hybris Workshop, Freiburg, May 27 28, 2013 Outline 1 2 3 May 28, 2013 2 / 34 Knowledge-Based Agents Example:

More information

Ontologies and Domain Theories

Ontologies and Domain Theories Ontologies and Domain Theories Michael Grüninger Department of Mechanical and Industrial Engineering University of Toronto gruninger@mie.utoronto.ca Abstract Although there is consensus that a formal ontology

More information

Heuristic Search Algorithms

Heuristic Search Algorithms CHAPTER 4 Heuristic Search Algorithms 59 4.1 HEURISTIC SEARCH AND SSP MDPS The methods we explored in the previous chapter have a serious practical drawback the amount of memory they require is proportional

More information

Argumentative Characterisations of Non-monotonic Inference in Preferred Subtheories: Stable Equals Preferred

Argumentative Characterisations of Non-monotonic Inference in Preferred Subtheories: Stable Equals Preferred Argumentative Characterisations of Non-monotonic Inference in Preferred Subtheories: Stable Equals Preferred Sanjay Modgil November 17, 2017 Abstract A number of argumentation formalisms provide dialectical

More information

initial state to some prespecied goal state, the task is to nd a sequence of actions that constitutes a legal execution of some prespecied non-determi

initial state to some prespecied goal state, the task is to nd a sequence of actions that constitutes a legal execution of some prespecied non-determi ConGolog, a concurrent programming language based on the situation calculus: foundations Giuseppe De Giacomo Dipartimento di Informatica e Sistemistica Universita di Roma \La Sapienza" Via Salaria 113,

More information

16.4 Multiattribute Utility Functions

16.4 Multiattribute Utility Functions 285 Normalized utilities The scale of utilities reaches from the best possible prize u to the worst possible catastrophe u Normalized utilities use a scale with u = 0 and u = 1 Utilities of intermediate

More information

Preferences over Objects, Sets and Sequences

Preferences over Objects, Sets and Sequences Preferences over Objects, Sets and Sequences 4 Sandra de Amo and Arnaud Giacometti Universidade Federal de Uberlândia Université de Tours Brazil France 1. Introduction Recently, a lot of interest arose

More information

Denotational Semantics

Denotational Semantics 5 Denotational Semantics In the operational approach, we were interested in how a program is executed. This is contrary to the denotational approach, where we are merely interested in the effect of executing

More information

Preference, Choice and Utility

Preference, Choice and Utility Preference, Choice and Utility Eric Pacuit January 2, 205 Relations Suppose that X is a non-empty set. The set X X is the cross-product of X with itself. That is, it is the set of all pairs of elements

More information

The computational complexity of dominance and consistency in CP-nets

The computational complexity of dominance and consistency in CP-nets The computational complexity of dominance and consistency in CP-nets Judy Goldsmith Dept. of Comp. Sci. University of Kentucky Lexington, KY 40506-0046, USA goldsmit@cs.uky.edu Abstract Jérôme Lang IRIT

More information

Nonmonotonic Tools for Argumentation

Nonmonotonic Tools for Argumentation Nonmonotonic Tools for Argumentation Gerhard Brewka Computer Science Institute University of Leipzig brewka@informatik.uni-leipzig.de joint work with Stefan Woltran G. Brewka (Leipzig) CILC 2010 1 / 38

More information

Contingent Planning with Goal Preferences

Contingent Planning with Goal Preferences Contingent Planning with Goal Preferences Dmitry Shaparau ITC-IRST, Trento, Italy shaparau@irst.itc.it Marco Pistore University of Trento, Italy pistore@dit.unitn.it Paolo Traverso ITC-IRST, Trento, Italy

More information

Searching in non-deterministic, partially observable and unknown environments

Searching in non-deterministic, partially observable and unknown environments Searching in non-deterministic, partially observable and unknown environments CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2017 Soleymani Artificial Intelligence:

More information

, and rewards and transition matrices as shown below:

, and rewards and transition matrices as shown below: CSE 50a. Assignment 7 Out: Tue Nov Due: Thu Dec Reading: Sutton & Barto, Chapters -. 7. Policy improvement Consider the Markov decision process (MDP) with two states s {0, }, two actions a {0, }, discount

More information

Reasoning about Actions with Sensing under Qualitative and Probabilistic Uncertainty

Reasoning about Actions with Sensing under Qualitative and Probabilistic Uncertainty Reasoning about Actions with Sensing under Qualitative and Probabilistic Uncertainty Luca Iocchi and homas Lukasiewicz 2 and Daniele Nardi and Riccardo Rosati Abstract. We focus on the aspect of sensing

More information

Multi-agent actions under uncertainty: situation calculus, discrete time, plans and policies

Multi-agent actions under uncertainty: situation calculus, discrete time, plans and policies Multi-agent actions under uncertainty: situation calculus, discrete time, plans and policies David Poole Department of Computer Science University of British Columbia Vancouver, B.C., Canada V6T 1Z4 poole@cs.ubc.ca

More information

Local Conditional High-Level Robot Programs (extended version)

Local Conditional High-Level Robot Programs (extended version) Local Conditional High-Level Robot Programs (extended version) Sebastian Sardiña Department of Computer Science University of Toronto Toronto, Canada M5S 1A4 ssardina@cs.toronto.edu, WWW home page: http://www.cs.toronto.edu

More information

Introduction to Arti Intelligence

Introduction to Arti Intelligence Introduction to Arti Intelligence cial Lecture 4: Constraint satisfaction problems 1 / 48 Constraint satisfaction problems: Today Exploiting the representation of a state to accelerate search. Backtracking.

More information

Decision making, Markov decision processes

Decision making, Markov decision processes Decision making, Markov decision processes Solved tasks Collected by: Jiří Kléma, klema@fel.cvut.cz Spring 2017 The main goal: The text presents solved tasks to support labs in the A4B33ZUI course. 1 Simple

More information

Coarticulation in Markov Decision Processes Khashayar Rohanimanesh Robert Platt Sridhar Mahadevan Roderic Grupen CMPSCI Technical Report 04-33

Coarticulation in Markov Decision Processes Khashayar Rohanimanesh Robert Platt Sridhar Mahadevan Roderic Grupen CMPSCI Technical Report 04-33 Coarticulation in Markov Decision Processes Khashayar Rohanimanesh Robert Platt Sridhar Mahadevan Roderic Grupen CMPSCI Technical Report 04- June, 2004 Department of Computer Science University of Massachusetts

More information

Cardinal and Ordinal Preferences. Introduction to Logic in Computer Science: Autumn Dinner Plans. Preference Modelling

Cardinal and Ordinal Preferences. Introduction to Logic in Computer Science: Autumn Dinner Plans. Preference Modelling Cardinal and Ordinal Preferences Introduction to Logic in Computer Science: Autumn 2007 Ulle Endriss Institute for Logic, Language and Computation University of Amsterdam A preference structure represents

More information

Automata-based Verification - III

Automata-based Verification - III COMP30172: Advanced Algorithms Automata-based Verification - III Howard Barringer Room KB2.20: email: howard.barringer@manchester.ac.uk March 2009 Third Topic Infinite Word Automata Motivation Büchi Automata

More information

Polarization and Bipolar Probabilistic Argumentation Frameworks

Polarization and Bipolar Probabilistic Argumentation Frameworks Polarization and Bipolar Probabilistic Argumentation Frameworks Carlo Proietti Lund University Abstract. Discussion among individuals about a given issue often induces polarization and bipolarization effects,

More information

CHAPTER 10. Gentzen Style Proof Systems for Classical Logic

CHAPTER 10. Gentzen Style Proof Systems for Classical Logic CHAPTER 10 Gentzen Style Proof Systems for Classical Logic Hilbert style systems are easy to define and admit a simple proof of the Completeness Theorem but they are difficult to use. By humans, not mentioning

More information

Common Sense Reasoning About Complex Actions

Common Sense Reasoning About Complex Actions Common Sense Reasoning About Complex Actions Abstract Reasoning about complex actions in partially observed and dynamic worlds requires imposing restrictions on the behavior of the external world during

More information

Reasoning by Regression: Pre- and Postdiction Procedures for Logics of Action and Change with Nondeterminism*

Reasoning by Regression: Pre- and Postdiction Procedures for Logics of Action and Change with Nondeterminism* Reasoning by Regression: Pre- and Postdiction Procedures for Logics of Action and Change with Nondeterminism* Marcus Bjareland and Lars Karlsson Department of Computer and Information Science Linkopings

More information

Reasoning about Actions in a Multiagent Domain

Reasoning about Actions in a Multiagent Domain Reasoning about Actions in a Multiagent Domain Laura Giordano 1, Alberto Martelli 2, and Camilla Schwind 3 1 DISTA, Università del Piemonte Orientale A. Avogadro laura@mfn.unipmn.it 2 Dipartimento di Informatica

More information

Intention and Rationality for PRS-Like Agents

Intention and Rationality for PRS-Like Agents Intention and Rationality for PRS-Like Agents Wayne Wobcke School of Computer Science and Engineering University of New South Wales Sydney NSW 2052, Australia wobcke@cse.unsw.edu.au Abstract. In this paper,

More information

Motivation for introducing probabilities

Motivation for introducing probabilities for introducing probabilities Reaching the goals is often not sufficient: it is important that the expected costs do not outweigh the benefit of reaching the goals. 1 Objective: maximize benefits - costs.

More information

Intelligent Agents. Formal Characteristics of Planning. Ute Schmid. Cognitive Systems, Applied Computer Science, Bamberg University

Intelligent Agents. Formal Characteristics of Planning. Ute Schmid. Cognitive Systems, Applied Computer Science, Bamberg University Intelligent Agents Formal Characteristics of Planning Ute Schmid Cognitive Systems, Applied Computer Science, Bamberg University Extensions to the slides for chapter 3 of Dana Nau with contributions by

More information

3 Propositional Logic

3 Propositional Logic 3 Propositional Logic 3.1 Syntax 3.2 Semantics 3.3 Equivalence and Normal Forms 3.4 Proof Procedures 3.5 Properties Propositional Logic (25th October 2007) 1 3.1 Syntax Definition 3.0 An alphabet Σ consists

More information

On simulations and bisimulations of general flow systems

On simulations and bisimulations of general flow systems On simulations and bisimulations of general flow systems Jen Davoren Department of Electrical & Electronic Engineering The University of Melbourne, AUSTRALIA and Paulo Tabuada Department of Electrical

More information

Mechanism Design for Argumentation-based Information-seeking and Inquiry

Mechanism Design for Argumentation-based Information-seeking and Inquiry Mechanism Design for Argumentation-based Information-seeking and Inquiry Xiuyi Fan and Francesca Toni Imperial College London, London, United Kingdom, {xf309,ft}@imperial.ac.uk Abstract. Formal argumentation-based

More information

Coarticulation in Markov Decision Processes

Coarticulation in Markov Decision Processes Coarticulation in Markov Decision Processes Khashayar Rohanimanesh Department of Computer Science University of Massachusetts Amherst, MA 01003 khash@cs.umass.edu Sridhar Mahadevan Department of Computer

More information

An Introduction to Description Logics

An Introduction to Description Logics An Introduction to Description Logics Marco Cerami Palacký University in Olomouc Department of Computer Science Olomouc, Czech Republic Olomouc, 21.11.2013 Marco Cerami (UPOL) Description Logics 21.11.2013

More information

Adding Time and Intervals to Procedural and Hierarchical Control Specifications

Adding Time and Intervals to Procedural and Hierarchical Control Specifications Adding Time and Intervals to Procedural and Hierarchical Control Specifications Tran Cao Son Computer Science Department New Mexico State University Las Cruces, NM 88003, USA tson@cs.nmsu.edu Chitta Baral

More information

Reinforcement Learning II

Reinforcement Learning II Reinforcement Learning II Andrea Bonarini Artificial Intelligence and Robotics Lab Department of Electronics and Information Politecnico di Milano E-mail: bonarini@elet.polimi.it URL:http://www.dei.polimi.it/people/bonarini

More information

Progression of Situation Calculus Action Theories with Incomplete Information

Progression of Situation Calculus Action Theories with Incomplete Information Progression of Situation Calculus Action Theories with Incomplete Information Stavros Vassos and Hector Levesque Department of Computer Science University of Toronto Toronto, ON, CANADA {stavros,hector}@cs.toronto.edu

More information

ConGolog, a concurrent programming language based on the situation calculus

ConGolog, a concurrent programming language based on the situation calculus Artificial Intelligence 121 (2000) 109 169 ConGolog, a concurrent programming language based on the situation calculus Giuseppe De Giacomo a,, Yves Lespérance b, Hector J. Levesque c a Dipartimento di

More information

From: AAAI Technical Report FS Compilation copyright 1994, AAAI ( All rights reserved. Forget It!

From: AAAI Technical Report FS Compilation copyright 1994, AAAI (  All rights reserved. Forget It! From: AAAI Technical Report FS-94-02. Compilation copyright 1994, AAAI (www.aaai.org). All rights reserved. Forget It! Fangzhen Lin and Ray Reiter* Department of Computer Science University of Toronto

More information

Axiomatic Semantics. Semantics of Programming Languages course. Joosep Rõõmusaare

Axiomatic Semantics. Semantics of Programming Languages course. Joosep Rõõmusaare Axiomatic Semantics Semantics of Programming Languages course Joosep Rõõmusaare 2014 Direct Proofs of Program Correctness Partial correctness properties are properties expressing that if a given program

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Dynamic Programming Marc Toussaint University of Stuttgart Winter 2018/19 Motivation: So far we focussed on tree search-like solvers for decision problems. There is a second important

More information

Marks. bonus points. } Assignment 1: Should be out this weekend. } Mid-term: Before the last lecture. } Mid-term deferred exam:

Marks. bonus points. } Assignment 1: Should be out this weekend. } Mid-term: Before the last lecture. } Mid-term deferred exam: Marks } Assignment 1: Should be out this weekend } All are marked, I m trying to tally them and perhaps add bonus points } Mid-term: Before the last lecture } Mid-term deferred exam: } This Saturday, 9am-10.30am,

More information

arxiv: v1 [cs.lo] 19 Mar 2019

arxiv: v1 [cs.lo] 19 Mar 2019 Turing-Completeness of Dynamics in Abstract Persuasion Argumentation Ryuta Arisaka arxiv:1903.07837v1 [cs.lo] 19 Mar 2019 ryutaarisaka@gmail.com Abstract. Abstract Persuasion Argumentation (APA) is a dynamic

More information

Automata and Languages

Automata and Languages Automata and Languages Prof. Mohamed Hamada Software Engineering Lab. The University of Aizu Japan Mathematical Background Mathematical Background Sets Relations Functions Graphs Proof techniques Sets

More information

CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability

CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability CS188: Artificial Intelligence, Fall 2009 Written 2: MDPs, RL, and Probability Due: Thursday 10/15 in 283 Soda Drop Box by 11:59pm (no slip days) Policy: Can be solved in groups (acknowledge collaborators)

More information

The Expressivity of Universal Timed CCP: Undecidability of Monadic FLTL and Closure Operators for Security

The Expressivity of Universal Timed CCP: Undecidability of Monadic FLTL and Closure Operators for Security The Expressivity of Universal Timed CCP: Undecidability of Monadic FLTL and Closure Operators for Security Carlos Olarte and Frank D. Valencia INRIA /CNRS and LIX, Ecole Polytechnique Motivation Concurrent

More information

On the Use of an ATMS for Handling Conflicting Desires

On the Use of an ATMS for Handling Conflicting Desires On the Use of an ATMS for Handling Conflicting Desires Leila Amgoud and Claudette Cayrol Institut de Recherche en Informatique de Toulouse 118, route de Narbonne 31062 Toulouse, France {amgoud, ccayrol}@irit.fr

More information

An Adaptive Clustering Method for Model-free Reinforcement Learning

An Adaptive Clustering Method for Model-free Reinforcement Learning An Adaptive Clustering Method for Model-free Reinforcement Learning Andreas Matt and Georg Regensburger Institute of Mathematics University of Innsbruck, Austria {andreas.matt, georg.regensburger}@uibk.ac.at

More information

Logic and Automata I. Wolfgang Thomas. EATCS School, Telc, July 2014

Logic and Automata I. Wolfgang Thomas. EATCS School, Telc, July 2014 Logic and Automata I EATCS School, Telc, July 2014 The Plan We present automata theory as a tool to make logic effective. Four parts: 1. Some history 2. Automata on infinite words First step: MSO-logic

More information

Markov decision processes

Markov decision processes CS 2740 Knowledge representation Lecture 24 Markov decision processes Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Administrative announcements Final exam: Monday, December 8, 2008 In-class Only

More information