Kybernetika. Harish C. Taneja; R. K. Tuteja Characterization of a quantitative-qualitative measure of inaccuracy

Similar documents
54 D. S. HOODA AND U. S. BHAKER Belis and Guiasu [2] observed that a source is not completely specied by the probability distribution P over the sourc

CODING THEOREMS ON NEW ADDITIVE INFORMATION MEASURE OF ORDER

New Generalized Entropy Measure and its Corresponding Code-word Length and Their Characterizations

Kybernetika. Nand Lal Aggarwal; Claude-François Picard Functional equations and information measures with preference

Časopis pro pěstování matematiky

Some Coding Theorems on Fuzzy Entropy Function Depending Upon Parameter R and Ѵ

Commentationes Mathematicae Universitatis Carolinae

On the redundancy of optimum fixed-to-variable length codes

Archivum Mathematicum

EQUADIFF 1. Milan Práger; Emil Vitásek Stability of numerical processes. Terms of use:

Mathematica Bohemica

Archivum Mathematicum

Czechoslovak Mathematical Journal

Czechoslovak Mathematical Journal

Acta Universitatis Carolinae. Mathematica et Physica

Commentationes Mathematicae Universitatis Carolinae

Časopis pro pěstování matematiky

Archivum Mathematicum

Mathematica Bohemica

Czechoslovak Mathematical Journal

Applications of Mathematics

Mathematica Slovaca. Dragan Marušič; Tomaž Pisanski The remarkable generalized Petersen graph G(8, 3) Terms of use:

Kybernetika. Jan Havrda; František Charvát Quantification method of classification processes. Concept of structural a-entropy

Mathematica Slovaca. Antonín Dvořák; David Jedelský; Juraj Kostra The fields of degree seven over rationals with a normal basis generated by a unit

On Measures of" Useful" Information*

A Coding Theorem Connected on R-Norm Entropy

Acta Mathematica Universitatis Ostraviensis

arxiv:math/ v1 [math.st] 19 Jan 2005

Časopis pro pěstování matematiky

EQUADIFF 1. Rudolf Výborný On a certain extension of the maximum principle. Terms of use:

Mathematica Slovaca. Ján Jakubík On the α-completeness of pseudo MV-algebras. Terms of use: Persistent URL:

Toposym 2. Zdeněk Hedrlín; Aleš Pultr; Věra Trnková Concerning a categorial approach to topological and algebraic theories

Applications of Mathematics

Linear Differential Transformations of the Second Order

Entropy and Ergodic Theory Lecture 3: The meaning of entropy in information theory

CSCI 2570 Introduction to Nanocomputing

Mathematica Bohemica

Kybernetika. Andrzej Dziech; Andrzej R. Pach Bounds on the throughput of an unslotted ALOHA channel in the case of a heterogeneous users' population

Časopis pro pěstování matematiky

Information Theory in Intelligent Decision Making

Acta Universitatis Carolinae. Mathematica et Physica

Source Coding. Master Universitario en Ingeniería de Telecomunicación. I. Santamaría Universidad de Cantabria

Nested Inequalities Among Divergence Measures

Upper Bounds on the Capacity of Binary Intermittent Communication

Commentationes Mathematicae Universitatis Carolinae

JOHN THICKSTUN. p x. n sup Ipp y n x np x nq. By the memoryless and stationary conditions respectively, this reduces to just 1 yi x i.

On Some information G. C. PATNI AND K. C. JAIN. Department of Mathematics, University of Rajasthan, Jaipur, India

Applications of Mathematics

Aplikace matematiky. Jiří Neuberg Some limit properties of the best determined terms method. Terms of use:

Chapter 9 Fundamental Limits in Information Theory

Lecture 1: Shannon s Theorem

Acta Mathematica et Informatica Universitatis Ostraviensis

Information and Entropy

Information Theory and Coding Techniques

ON THE SHORTEST PATH THROUGH A NUMBER OF POINTS

Kybernetika. Michał Baczyński; Balasubramaniam Jayaram Yager s classes of fuzzy implications: some properties and intersections

Mathematica Slovaca. Jaroslav Hančl; Péter Kiss On reciprocal sums of terms of linear recurrences. Terms of use:

Czechoslovak Mathematical Journal

The Distribution of Generalized Sum-of-Digits Functions in Residue Classes

Data Compression. Limit of Information Compression. October, Examples of codes 1

Mathematica Bohemica

LINEAR EQUATIONS WITH UNKNOWNS FROM A MULTIPLICATIVE GROUP IN A FUNCTION FIELD. To Professor Wolfgang Schmidt on his 75th birthday

1 Introduction to information theory

APC486/ELE486: Transmission and Compression of Information. Bounds on the Expected Length of Code Words

WSGP 23. Lenka Lakomá; Marek Jukl The decomposition of tensor spaces with almost complex structure

Jensen-Shannon Divergence and Hilbert space embedding

Future Self-Guides. E,.?, :0-..-.,0 Q., 5...q ',D5', 4,] 1-}., d-'.4.., _. ZoltAn Dbrnyei Introduction. u u rt 5,4) ,-,4, a. a aci,, u 4.

Archivum Mathematicum

A Single-letter Upper Bound for the Sum Rate of Multiple Access Channels with Correlated Sources

~,. :'lr. H ~ j. l' ", ...,~l. 0 '" ~ bl '!; 1'1. :<! f'~.., I,," r: t,... r':l G. t r,. 1'1 [<, ."" f'" 1n. t.1 ~- n I'>' 1:1 , I. <1 ~'..

The sequential decoding metric for detection in sensor networks

Acta Universitatis Palackianae Olomucensis. Facultas Rerum Naturalium. Mathematica-Physica-Chemica

Aplikace matematiky. Gene Howard Golub Numerical methods for solving linear least squares problems

Czechoslovak Mathematical Journal

Toposym Kanpur. T. Soundararajan Weakly Hausdorff spaces and the cardinality of topological spaces

Information Theory and Statistics Lecture 2: Source coding

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Chapter 5: Data Compression

Dept. of Linguistics, Indiana University Fall 2015

The Poisson Channel with Side Information

Acta Universitatis Palackianae Olomucensis. Facultas Rerum Naturalium. Mathematica

Acta Universitatis Carolinae. Mathematica et Physica

Časopis pro pěstování matematiky a fysiky

An estimate for the probability of dependent events

The Method of Types and Its Application to Information Hiding

Variable Length Codes for Degraded Broadcast Channels

Mathematica Slovaca. Zbigniew Piotrowski Somewhat continuity on linear topological spaces implies continuity

An introduction to basic information theory. Hampus Wessman

Kybernetika. Margarita Mas; Miquel Monserrat; Joan Torrens QL-implications versus D-implications. Terms of use:

EECS 750. Hypothesis Testing with Communication Constraints

,. WOLFOWITZ. 0 < h < 1. A"code" of length is set l(x,a),i 1,...,t},where CORRECTING CODE THE MAXIMUM ACHIEVABLE LENGTH OF AN ERROR

Convergence of generalized entropy minimizers in sequences of convex problems

INVEX FUNCTIONS AND CONSTRAINED LOCAL MINIMA

The Measure of Information Uniqueness of the Logarithmic Uncertainty Measure

Information Theory, Statistics, and Decision Trees

Czechoslovak Mathematical Journal

Lecture 3. Mathematical methods in communication I. REMINDER. A. Convex Set. A set R is a convex set iff, x 1,x 2 R, θ, 0 θ 1, θx 1 + θx 2 R, (1)

Acta Universitatis Palackianae Olomucensis. Facultas Rerum Naturalium. Mathematica

Generalized hashing and applications to digital fingerprinting

Kybernetika. Sebastian Fuchs Multivariate copulas: Transformations, symmetry, order and measures of concordance

Transcription:

Kybernetika Harish C. Taneja; R. K. Tuteja Characterization of a quantitative-qualitative measure of inaccuracy Kybernetika, Vol. 22 (1986), o. 5, 393--402 Persistent URL: http://dml.cz/dmlcz/124578 Terms of use: Institute of Information Theory and Automation AS CR, 1986 Institute of Mathematics of the Academy of Sciences of the Czech Republic provides access to digitized documents strictly for personal use. Each copy of any part of this document must contain these Terms of use. This paper has been digitized, optimized for electronic delivery and stamped with digital signature within the project DML-CZ: The Czech Digital Mathematics Library http://project.dml.cz

KYBERETfKA - VOLUME 22 (1986), UMBER 5 CHARACTERIZATIO OF A QUATITATIVE-QUALITATIVE MEASURE OF IACCURACY H. C. TAEJA, R. K. TUTEJA A quantitative-qualitative measure of inaccuracy is suggested and is characterized under a set of assumptions. Some properties of the new measure are also discussed. 1. ITRODUCTIO Let P = (pi, p 2,, p ), 0 = Pi = 1, YaPi = 1 be the probability distribution ;= l associated with a finite system of events E = (E l5 E 2,..., E ) representing the realization of some experiment. The different events E,- depend upon the experimenter's goal or upon some qualitative characteristic of the physical system taken into consideration, that is, they have different weights, or utilities. In order to distinguish the events E UE 2,...,E with respect to a given qualitative characteristic of the physical system taken into account, ascribe to each event E t a non-negative number «,- (s^o) directly proportional to its importance and call u ; the utility of the event E ;. Then the weighted entropy [l] of the experiment E is defined as (1.1) I(P;U)= -_Tu lpllog Pl ow let us suppose that the experimenter asserts that the probability of the ith outcome E t is q h whereas the true probability is p h with _P q t = _T Pi = 1. Thus, l = 1 ; = 1 we have two utility information schemes: (1.2) S = E,E2. PÍPI Ulu2 E~ P. u_ 0 ^ Pi g 1, t/j è 0, Щp. = 1 393

of a set of iv events after an experiment, and (1.3) S 0 = EгE 2. E ' ЧiЧг."l«2.. U 0 = <.,=!, и, = 0, <?, = l of the same set of events before the experiment. In both the schemes (1.2) and (1.3) the utility distribution is the same because we assume that the utility u, of an outcome E t is independent of its probability of occurrence p t, or predicted probability q t ; «, is only a "utility" or value of the outcome E, for an observer relative to some specified goal (refer to [5]). The quantitative-qualitative measure of relative information [8, 9], that the scheme (1.2) provides about the scheme (1.3), is (1.4). /(/, e;cl) = «, Pi iog(p I./ fli ). f=il The measure (1.4), in some sense, can be taken as a measure of the extent to which the forecasts q lt q 2 q differ from the corresponding realizations p., p 2 p in a goal oriented experiment E = (E lt E 2 E ). When the utilities are ignored, that is, M, = 1 for each i, the measure (1.4) reduces to the Kullback's measure of relative information [4]. Consider I(P; U) +I(P\Q;U)=-Z u ipi log p, + u ipi log (pjq t ) = i=i i=i = ~ I "(Pi log q t ; i=i and let it be denoted by I(P; Q; 17). Thus (1.5) I(P;Q;U)= - H.p, log <.,. (=i When the utilities are ignored, then (1.5) reduces to Kerridge's inaccuracy [3]. Therefore (1.5) can be viewed as a measure of the inaccuracy associated with the statement of an experimenter made in context with a goal oriented experiment. We can consider (1.5) as a quantitative-qualitative measure of inaccuracy associated with the statement of an experimenter. When p, = q t, for each i, then (1.5) reduces to (1.1), the weighted entropy [1]. In the next section, we derive afresh the measure (1.5) under a set of intuitively reasonable assumptions. =i 2. THE QUAUTATIVE-QUALITATIVE MEASURE OF IACCURACY Let I(Pi> Pi> ; <Zi> <Z2>.; «i> «2> ) be the measure of inaccuracy associated with the goal oriented experiment E = (E u E 2,...). In order to characterize the J function we consider the following 394

A x. The function / is continuous with respect to its arguments p,'s, g:s and M:s. A 2. When JV equally likely alternatives, each having the same utility u, are stated to be equally likely the inaccuracy is a monotonic increasing function of JV. A 3. If a statement is broken down into a number of subsidiary statements, then inaccuracy of the original statement is the weighted sum of the inaccuracies of the subsidiary statements. For example, we must have KPU P2, Ps; q.u 12, <? 3; «i. u 2, u 3) = p 2u 2 + p 3u, = 1 \ Pu P2 + p 3; <iu i2 + q 3; "u P 2 + P3 + (p 2 + P 3 )i(-^-,-^~;~^~,-^-;u 2,u 3 \P2 + Pi P2 + Pi 42 + 4i 42 + 4i A 4. The inaccuracy of a statement is unchanged if two alternatives about which the same assertion is made are combined. For example, I(Pu Pi, Pi, 4\, 42, 42, "1, "2, "3) = P 2U 2 + Pi u i = l[pup2 + Pi-,4u42,uu- P2 + Pi A s. The inaccuracy of a statement is directly proportional to the utilities of the outcomes. For example, for every non-negative A, we must have I(Pu Pz, Pi\ 4i, 42, 4i, h*u &u 2, Xu 3) = = XI(p u p 2, p 3; q u q 2, q 3; u u u 2, u 3). All these A : to A 4 are just modifications of Kerridge's inaccuracy assumptions and A 5 is the monotonicity law expressed by the utilities. In the following theorem we characterize the measure of inaccuracy associated with this system. The proof is on the same lines as in the characterization of Kerridge's inaccuracy [3]. Theorem 1. The only function satisfying the axioms A, to A 5 is (2.1) I(P;Q;U)= -Kjjt^logq,, where K is an arbitrary positive number and the logarithm base is any number greater then one. Proof. It is not difficult to verify that the function (2.1) satisfy axioms A t to A 5. ow we prove that any function satisfying these axioms must be of the form (2.1). Consider the case when there are s m alternatives with utilities u u u 2,..., which are 395

asserted to be equally likely, they may or may not be so. Then by axiom A 4 where u = Y^'iPi- By axiom A 5 /(/>!, p 2,...; s~ m, s~ m,...,; u y, u 2,...) = 1\1; s~ m ; u) (2.2) I([;s~ m ;u) = ul(\;s~ m ;\). Let 7(1; s~ m ; 1) = A's m ). ow A(s m ) is independent of p:s, the true probabilities, therefore, A(s m ) remains unchanged if we replace the true value of p t's by s~ m for all i. By axiom A 3, A(s m ) = m A(s). Using the continuity axiom A t and the monotonic character of A[s), we get, (refer to [7, p. 82]), A(s) = K log s, where K (>0) is an arbitrary constant. Consider the case when all the q,'s are rational. They can be then expressed in the form q ; = n,/jv, where n:s are integers and = Y_jr;. By axiom A 3 I(p u p 2,...; nj, n 2j,...; «u 2,...) + p, J(l; l/«,; «,) = or s/(l;l/iv;5:pi-i). I(p up 2,...; n tj, n 2j,...; u u u 2,...) = = 1(1; ll-^pvd ~ HPJ(U ll'h; Ui) = = V>,«,/(1; Ij; 1) -?,«,/(I ; 1/n,; 1) = = KQjpjU,) log At - K ^p,«, log «j = = -KYJpi«,Aog(«i/At) = -KXP,"ilog«i. By the continuity assumption A,, this holds for all q h not only for rational values. ~~ We shall assume K 1 and take logarithm to the base '2'. We define JV (2.3) I(P; Q;U) = - «,?, log q i7 0 g p ;,?, g 1, w ; ^ 0, ;= l 2>. = l9,= l, i = 1 i = 1 as the quantitative-qualitative measure of inaccuracy associated with the statement of an experimenter who asserts the probabilities of the various outcomes E 1; E 2,...,..., Ejv with utilities u u u 2,..., u, as a l5 <5r 2,, 9 whereas the true probabilities are Pi, p 2,.. p. The absence of a goal implies the absence of a utility measure, that is, the various 396

events are no longer different from a qualitative point of view. The utilities u u u 2,......,u in (2.3) are equal to each other; in order to completely avoid their influence we put u l=u 2 =... = u =\.\n this case, (2.3) becomes /(P;Q;U)= - p t log q t = I(P; Q) i = I which is exactly Kerridge's inaccuracy [3]. When in a goal-directed experiment all events have equal utilities u t = u 2 =...... = u = u. Then (2.3) becomes I(P; Q;U)= -M Pi log q t = u ]{P; Q), which expresses the increase or decrease of the quantitative-qualitative measure of inaccuracy according to the common utility '«' of the event. In particular case when all events have zero utilities with regard to the goal pursued we get a total quantitative-qualitative measure of inaccuracy l(p; Q; U) = 0, even if Kerridge's inaccuracy is not zero. The quantitative-qualitative measure of inaccuracy is also zero if, p t = q t = 1 for one value and consequently zero for all other i, whatever the utilities u t 2: 0, (/ = 1,2,...,) may be. There is an infinite value of I[P; Q; U) if q v = 0, Pi 4= 0, w ; + 0 for any /. 3. PROPERTIES OF THE QUATITATIVE-QUALITATIVE MEASURE OF IACCURACY Following are some of the important properties satisfied by the measure l(p; Q; U): (1) The measure I(P; Q; U) is non-negative, i.e. I(P; Q; U) ^ 0. (2) The measure I(P; Q; U) is a symmetric function of its arguments, that is, I(P; Q; U) remains unchanged if the elements of P, Q and U are arranged in the same way so that one to one correspondence among them is not changed. (3) The measure l(p; Q; U) is a continuous function of its arguments. (4) The measure I(P; Q; U) satisfies the generalized weighted additivity; i.e. where and I(P*P'; Q*Q'; U*U')= V'I(P;Q;U) + U I(P'; Q'; U'), P*P' =(PIPI,~;PIPM',---;PP'I,--,PP'M), Q* Q' = (<ii<i'i> >«I«M; ;iq'i, ^I'M), U *U' = (U1U'1,...,U1U'M;...;UU'1L,...,UU'U) U = Zu ipi, U' = ^u jpj. ;=i j=i 397

(5) The measure I(P; Q; U) satisfies the branching property as follows: I(pup2,... p; quq2,...,q; uuu2,, U) =,- / U \P\ + M 2?2 = If Pi + PI,--;P\ <?. + Hi, -,<?*;, >«* Pi + Pi + ( Pl + P 2 ) l ( - ^ ^, - ^ - ; ^ i -, - ^ - ; " 1,» 2 \Pl + Pl Pl + Pl 1\ +42 Ql + <?2 (6) The minima of the quantitative-qualitative measure of inaccuracy l(p; Q; U) exists at q { = «;/? ;/ u lp i, /=1 i = 1,2,...,. the normalized preferences if the events E h The properties (1) to (5) can be verified very easily, however, to prove the property (6), we give the following theorem: Theorem 2. For fixed u h p h the minima of l(p; Q; U) exists at q t = u^ij u ;p ;, the normalized preferences of the events E ;'s, i = 1, 2,...,.,= ' Proof. We are to find the extreme points for the function I(P; Q; U) = - ";P; lo 8 <7;» ;= l JV IV M ; =» = P;> <?; =s 1 > for fixed u ;, p ;> under the condition a ; = 1. Using the method of Lagrange's multipliers, set (3.1) F(q u q 2,..., q ; X) = - u^ log «, + A( «, - 1), i = 1 i = 1 where X is an arbitrary constant called the Lagrange's constant. ow (3.2) 8 -l=- U^+X, 8q; <?; for i = 1,2,...,; and (3-3) SF f - l ^ - l. ox >=i Equation (3.2) when equated to zero gives (3-4) <?. =» i = l,2,...,. Equating 5F/3A = 0, we get (3-5) «. = -. 398

From (3.4) and (3.5), we get A = u ipi. Thus extreme value for I(P; Q; U) exists at (3.6) <?i = u ipil u ipi, i = 1,2,..., AT. ext, we verify whether (3.6) is a point of maxima or minima for I(P; Q; U). The border matrix for the system under consideration is Чi 2 "ipl «1 n (3.7) ía 0 "2P2 «2 "JVPÍV <7,v J It can be very easily verified that the minors of order 3, 4, 5... etc. of the determinant of the matrix (3.7) are negative. Thus minima occurs at the point (3.6). Q The minimum of I(P; Q;U)=-Y J I'tPi log (u ipij ihpi) = i = 1 i = 1 = u tp, log Pi u log u + u log u, i= 1 where the bar means the mean value with respect to the probability distribution P = ( Pl,p 2,...,p ). Since u log u is a convex U function, therefore, u log u S: u log u, and thus, minimum of I'P; Q; U) g I(P; U), the weighted entropy of the experiment E = = (E UE 2,...,E ). 4. QUATITATIVE-QUALITATIVE MEASURE OF IACCURACY AD CODIG THEORY Consider an information source with output symbols E = (E u E 2,..., E ), and let Q = (q it q 2,..., q ) and P = ( Pl, Pz,.-., P) be respectively the asserted and the realized probability distributions for the source alphabet. Let here each letter E ; be characterized by an additional parameter u ; and thus the cost c f of transmitting E f through the noiseless channel is proportional to the product u.n,, where n i is the length of the codeword associated with E ;. The experimenter constructs code (in fact, personal probability code) keeping in view to minimize the average transmission 399

cost, or equivalently the weighted mean length, (refer to [2]), Z 1i u i n i (4-1) Ku;q) = ^, while the actual weighted mean length is JV Z1j"j J~i (4.2) D«;p) = ^. Rewriting (4.1) and (4.2) as Z Py"y J = I (4-3) L{u; q) - <?;«,. and (4-4) Z/u; p) =!>;«,. respectively, where (4.5),; = «" and i=l i=l Z 9J«J J = l (4-6) Pi = ^ - z PJ U J J=I The distributions (4.5) and (4.6) represent respectively the auxiliary predicted and actual probability distributions over the source alphabet E = (,, E 2,..., E ). We have the following theorem: Theorem 3. If the codeword lengths n u n 2,..., ]T D~ n> g 1, then the weighted mean length is bounded by (4.7) l(p;q;u)-(ulo g u) p + u p lo g u q L, M. < u log > < I(P; Q;U) - (u\o g u) p + - p \o g ü a + j ü log D x n satisfy the Kraft's inequality where Z Pí«<"i L («; p) = ^. I(I>; e; c!) = - z U ÍPÍ iog 9i, v-, 1=1 Z P_i«7 400

and JV (H log U)p = UiPi log U;, Up = Z UiPi, H, = E U-^i. i=i ;=i i=l Proof. Equation (4.7) immediately follows from the Kerridge's inequality [3], JV JV - Z P'i lo g l'i = Z P'i"i < ~ Z P't log «i + 1. ;=i i=i i=i where {pj}f=1 and {<?-}f=1, as defined by (4.6) and (4.5) respectively, are the auxiliary actual and predicted probability distributions over the source alphabet E = = (EUE2,...,E). Particular cases: (I) When P = Q, (4.7) reduces to I'P; U) - u log,u + /7 log u ( _. /(P; U) u log H + il log u u log D (7 log D where /(P; U) = Z M ;P' ' g P. ' s t l ie weighted entropy [l], and the bar means ;=i the mean value with respect to the probability distribution P = (p., p2,..., pa). These were the bounds obtained by Longo [6]. (2) When the utilities are ignored, that is, ut = 1 for each i', then (4.7) reduces to - Z Pi ' g ii = Z P."; < - Z Pi lq g *i + i; ;=i ;=i i=i a result obtained by Kerridge [3]. (Received June 4, 1985.) REFERECES [1] M. Belis and S. Guiasu: A quantitative-qualitative measure of information in cybernetic system. IEEE Trans. Inform. Theory 14 (1968), 593-594. [2] S. Guiasu and C. F. Picard: Borne inferieure de la longueur utile de certains codes. C. R. Acad. Sci. Paris 273 (1971), 248-251. [3] D. F. Kerridge: Inaccuracy and inference. J. Roy. Statist. Soc. Ser. B 23 (1961), 184-194. [4] S. Kullback: Information Theory and Statistics. Wiley, ew York 1959. [5] G. Longo: Quantitative-Qualitative Measures of Information. Springer-Verlag, ew York 1972. [6] G. Longo: A noiseless coding theorem for sources having utilities. SIAM J. Appl. Math. J0(1971), 739 748. [7] C. Shannon and W. Weaver: The Mathematical Theory of Communication. Univ. of Illinois Press, 1949. 401

[8] H. C. Taneja: On measures of relative 'useful' information. Kybernetika 21 (1985), 148 156. [9] H. C. Taneja and R. K. Tuteja: Characterization of a quantitative-qualitative measure of relative information. Inform. Sciences 33 (1984), 217 222. Dr. H. C. Taneja, Department of Mathematics, Maharshi Dayanand University, Rohtak 124 001. India. Dr. R. K. Tuteja, Department of Mathematics, Maharshi Dayanand University, Rohtak 124 001. India. 402