Generalized Fechnerian Scaling

Similar documents
Dissimilarity, Quasimetric, Metric

MEASUREMENT AND REPRESENTATION OF SENSATIONS

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Dissimilarity Cumulation Theory in Smoothly Connected Spaces

Tree sets. Reinhard Diestel

Mathematics-I Prof. S.K. Ray Department of Mathematics and Statistics Indian Institute of Technology, Kanpur. Lecture 1 Real Numbers

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Incompatibility Paradoxes

The Symmetric Space for SL n (R)

chapter 12 MORE MATRIX ALGEBRA 12.1 Systems of Linear Equations GOALS

KRIPKE S THEORY OF TRUTH 1. INTRODUCTION

Part III. 10 Topological Space Basics. Topological Spaces

1 Euclidean geometry. 1.1 The metric on R n

COMPLEXITY OF SHORT RECTANGLES AND PERIODICITY

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

Mathematics 114L Spring 2018 D.A. Martin. Mathematical Logic

Principal Component Analysis

ELEMENTARY LINEAR ALGEBRA

a 11 x 1 + a 12 x a 1n x n = b 1 a 21 x 1 + a 22 x a 2n x n = b 2.

RECOVERING NORMAL NETWORKS FROM SHORTEST INTER-TAXA DISTANCE INFORMATION

Nonlinear Manifold Learning Summary

Theorems. Theorem 1.11: Greatest-Lower-Bound Property. Theorem 1.20: The Archimedean property of. Theorem 1.21: -th Root of Real Numbers

Linear Algebra. The analysis of many models in the social sciences reduces to the study of systems of equations.

A Class of Partially Ordered Sets: III

A matrix over a field F is a rectangular array of elements from F. The symbol

Statistical Pattern Recognition

Another algorithm for nonnegative matrices

The Skorokhod reflection problem for functions with discontinuities (contractive case)

The matrix approach for abstract argumentation frameworks

CALCULUS ON MANIFOLDS. 1. Riemannian manifolds Recall that for any smooth manifold M, dim M = n, the union T M =

Connectedness. Proposition 2.2. The following are equivalent for a topological space (X, T ).

Notes on Ordered Sets

X X (2) X Pr(X = x θ) (3)

Set, functions and Euclidean space. Seungjin Han

LECTURE NOTE #11 PROF. ALAN YUILLE

10. Smooth Varieties. 82 Andreas Gathmann

ALGEBRA. 1. Some elementary number theory 1.1. Primes and divisibility. We denote the collection of integers

Data Mining. Dimensionality reduction. Hamid Beigy. Sharif University of Technology. Fall 1395

Math 239: Discrete Mathematics for the Life Sciences Spring Lecture 14 March 11. Scribe/ Editor: Maria Angelica Cueto/ C.E.

4.6 Bases and Dimension

VC-DENSITY FOR TREES

2. Intersection Multiplicities

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University

Metric Spaces and Topology

Hardy s Paradox. Chapter Introduction

THE STRUCTURE OF 3-CONNECTED MATROIDS OF PATH WIDTH THREE

Isomorphisms between pattern classes

Ultimate approximation and its application in nonmonotonic knowledge representation systems

MATH 521, WEEK 2: Rational and Real Numbers, Ordered Sets, Countable Sets

FINITE ABELIAN GROUPS Amin Witno

October 25, 2013 INNER PRODUCT SPACES

7 Matrix Operations. 7.0 Matrix Multiplication + 3 = 3 = 4

Chapter 3. Riemannian Manifolds - I. The subject of this thesis is to extend the combinatorial curve reconstruction approach to curves

Section Summary. Relations and Functions Properties of Relations. Combining Relations

Lecture 10: Dimension Reduction Techniques

Non-linear Dimensionality Reduction

CS632 Notes on Relational Query Languages I

MA103 Introduction to Abstract Mathematics Second part, Analysis and Algebra

Denotational Semantics

Computational Tasks and Models

The Gauss-Jordan Elimination Algorithm

6 Evolution of Networks

8. Prime Factorization and Primary Decompositions

Recursive definitions on surreal numbers

Foundations of Mathematics MATH 220 FALL 2017 Lecture Notes

Min-Rank Conjecture for Log-Depth Circuits

Boolean Inner-Product Spaces and Boolean Matrices

1 Differentiable manifolds and smooth maps

Equivalence of dynamical systems by bisimulation

Löwenheim-Skolem Theorems, Countable Approximations, and L ω. David W. Kueker (Lecture Notes, Fall 2007)

Lebesgue Measure on R n

Rigid Geometric Transformations

INDUCTION AND RECURSION

1.1 Line Reflections and Point Reflections

Newtonian Mechanics. Chapter Classical space-time

The Integers. Peter J. Kahn

Global (ISOMAP) versus Local (LLE) Methods in Nonlinear Dimensionality Reduction

GALOIS GROUPS OF CUBICS AND QUARTICS (NOT IN CHARACTERISTIC 2)

ISSN: (Online) Volume 3, Issue 5, May 2015 International Journal of Advance Research in Computer Science and Management Studies

Week 2: Sequences and Series

A NEW SET THEORY FOR ANALYSIS

Theorem 5.3. Let E/F, E = F (u), be a simple field extension. Then u is algebraic if and only if E/F is finite. In this case, [E : F ] = deg f u.

Shortest paths with negative lengths

Normal Fans of Polyhedral Convex Sets

The small ball property in Banach spaces (quantitative results)

Lecture 4: Constructing the Integers, Rationals and Reals

Axiomatic set theory. Chapter Why axiomatic set theory?

Metric spaces and metrizability

8.6 Partial Fraction Decomposition

Postulate 2 [Order Axioms] in WRW the usual rules for inequalities

Characterizations of the finite quadric Veroneseans V 2n

1 Lagrange Multiplier Method

4.7 The Levi-Civita connection and parallel transport

Partial cubes: structures, characterizations, and constructions

Part V. 17 Introduction: What are measures and why measurable sets. Lebesgue Integration Theory

Enumeration Schemes for Words Avoiding Permutations

Knowledge spaces from a topological point of view

Some Background Material

A Multiplicative Operation on Matrices with Entries in an Arbitrary Abelian Group

2 Equivalence Relations

Transcription:

Generalized Fechnerian Scaling Ehtibar N. Dzhafarov Purdue University, Indiana Hans Colonius University of Oldenburg, Oldenburg, Germany. INTRODUCTION The problem of reconstructing distances among stimuli from some empirical measures of pairwise dissimilarity is old. The measures of dissimilarity are numerous, including numerical ratings of (dis)similarity, classifications of stimuli, correlations among response variables, errors of substitution, and many others (Everitt & Rabe-Hesketh, 997; Suppes, Krantz, Luce, & Tversky, 989; Semple & Steele, 2). Formal representations of proximity data, like Multidimensional Scaling (org & Groenen, 997; Kruskal & Wish, 978) or Cluster Analysis (Corter, 996; Hartigan, 975), serve to describe and display data structures by embedding them in low-dimensional spatial or graph-theoretical configurations, respectively. In Multidimensional Scaling one embeds data points in a low-dimensional Minkowskian (usually, Euclidean) space so that distances are monotonically (in the metric version, proportionally) related to pairwise dissimilarities. In Cluster Analysis one typically represents proximity relations by a series of partitions of the set of stimuli resulting in a graph-theoretic tree structure with ultrametric or additive-tree metric distances. Discrimination probabilities ψ (x, y) =Pr[x and y are judged to be different] () which we discussed in Chapter occupy a special place among available measures of pairwise dissimilarity. The ability of telling two objects apart or identifying them as being the same (in some respect or overall) is arguably the most basic cognitive ability in biological perceivers and the most basic requirement of intelligent technical systems. At least this seems to be a plausible view, granting it is not self-evident. It is therefore a plausible position that a metric appropriately computed from the values of ψ (x, y)

2 Dzhafarov and Colonius may be viewed as the subjective metric, a network of distances from the point of view of a perceiver. As discussed in Chapter, the notion of a perceiver has a variety of possible meanings, including even cases of paper-and-pencil perceivers, abstract computational procedures assigning to every pair x, y the probability ψ (x, y) (subject to Regular Minimality). The example given in Chapter was that of ψ (x, y) being the probability with which a data set (in a particular format) generated by a statistical model, x, rejects (in accordance with some criterion) a generally different statistical model, y. The pairwise determinations of sameness/difference in this example (meaning, model y is retained/rejected when applied to a data set generated by model x) are usually readily available and simple. It is an attractive possibility therefore to have a general algorithm in which one can use these pairwise determinations to compute distances among conceptual objects (here, statistical models). The alternative, an a priori choice of a distance measure between two statistical models, may be less obvious and more difficult to justify. This chapter provides an informal introduction to Fechnerian Scaling, a metric-from-discriminability theory which ha been gradually developed by the present authors in the recent years (Dzhafarov, 22a, b, c, d; 2a, b; and Dzhafarov & Colonius, 999, 2, 25a, b). Its historical roots, however, can be traced back to the work of G. T. Fechner (8-887). To keep the presentation on a non-technical level, we will provide details for only the mathematically simplest case of Fechnerian Scaling, the case of discrete stimulus sets (such as letters of alphabets or Morse codes); only asimplified and abridged account of the application of Fechnerian Scaling to continuous stimulus spaces will be given. Notation conventions are the same as in Chapter... Example Consider the toy matrix used in Chapter, presented in a canonical form, TOY A C D A..8.8..6 C.6.9.5.7 D.6.9.5 This matrix will be used throughout to illustrate various points. We will describe a computational procedure, Fechnerian Scaling, which when applied

Distances from Discriminability to such matrices produces a matrix of distances we call Fechnerian. Intuitively, they reflect the degree of subjective dissimilarity among the stimuli, from the point of view of the perceiver (organism, group, technical device, or a computational procedure) to whom stimuli x, y {A,, C, D} were presented pairwise and whose responses (interpretable as same and different ) were used to compute the probabilities ψ (x, y) shown as the matrix entries. In addition, when the set of stimuli is finite, Fechnerian Scaling produces a set of what we call geodesic loops, the shortest (in some well-defined sense) chains of stimuli leading from one given object to another given object and back. Thus, when applied to our matrix TOY, Fechnerian Scaling yields the following two matrices: L A C D A A AC CAC DAD ACA CC DCD C ACA C C DCD D ADA DC CDC D G A C D A...9. C.9.7 D..7 We can see in matrix L, for instance, that the shortest (geodesic) loop connecting A and within the 4-element space {A,, C, D} is A C A, whereas the geodesic loop connecting A and C inthesamespaceis A C A. The lengths of these geodesic loops (whose computation will be explained later) is taken to be the Fechnerian distances between A and and between A and C, respectively. As we see in matrix G, the Fechnerian distance between A and is. times the Fechnerian distance between A and C. We should recall some basic facts from Chapter. () The row stimuli and the column stimuli in TOY belong to two distinct observation areas (say, row stimuli are those presented on the left, or chronologically first, the column stimuli are presented on the right, or second). (2) {A,, C, D} are psychologically distinct, that is, no two rows or two columns in the matrix are identical (if they were, they would be merged into a single one). () TOY may be the result of a canonical relabeling of a matrix in which the minima lie outside the main diagonal, such as

4 Dzhafarov and Colonius TOY y a y b y c y d x a.6.6..8 x b.9.9.8. x c.5.6 x d.5.7 The physical identity of the {A,, C, D} in TOY may therefore be different for the row stimuli and the column stimuli..2. Features of Fechnerian Scaling (A) Regular Minimality is the cornerstone of Fechnerian Scaling, and in the case of discrete stimulus sets, it is essentially the only prerequisite for Fechnerian Scaling. Due to Regular Minimality we can assume throughout most of this chapter that our stimulus sets are canonically (re)labeled (as in TOY ), so that or equivalently, x 6= y = ψ (x, y) > max {ψ (x, x),ψ(y, y)}, (2) x 6= y = ψ (x, x) < min {ψ (x, y),ψ(y, x)}. () In accordance with the discussion of the fundamental properties of discrimination probabilities (Chapter ), Fechnerian Scaling does not presuppose that ψ (x, x) is the same for all x (Nonconstant Self-Dissimilarity), or that ψ (x, y) =ψ (y, x) (Asymmetry). () The logic of Fechnerian Scaling is very different from the existing techniques of metrizing stimulus spaces (such as Multidimensional Scaling) in the following respect: Fechnerian distances are computed within rather than across the two observation areas. In other words, the Fechnerian distance between a and b does not mean a distance between a presented first (or on the left) and b presented second (or on the right). Rather, we should logically distinguish G () (a, b), the distance between a and b in the first observation area, from G (2) (a, b), the distance between a and b in the second observation area. This must not come as a surprise if one keeps in mind that a and b in the first observation area are generally perceived differently from a and b in the second observations area. As it turns out, however, if Regular Minimality is satisfied and the stimulus set is put in a canonical form, then it follows from the general theory that G () (a, b) =G (2) (a, b) =G (a, b).

Distances from Discriminability 5 This is illustrated in the diagram below, where the line connecting a stimulus in O with a stimulus in O (O standing for observation area) represents the probability ψ of their discrimination. Note that, for a given a, b, distance G (a, b) is computed, in general, from ψ (x, y) for all x, y, and not just from ψ (a, b). Later all of this will be explained in detail. observation area a G(a, b) b observation area 2 a G(a, b) b (C) In TOY a geodesic loop containing two given stimuli is defined uniquely. In general, however, this need not be the case: there may be more than one loop of the shortest possible length. Moreover, when the set of stimuli is infinitely large, whether discrete or continuous, geodesic loops may not exist at all, and the Fechnerian distance between two stimuli is then defined as the greatest lower bound (rather than minimum) of lengths of all loops that include these two stimuli... Fechnerian Scaling and MDS Multidimensional Scaling (MDS) when applied to discrimination probabilities serves as a convenient reference against which to consider the procedure of Fechnerian Scaling. Assuming that discrimination probabilities ψ (x, y) are known precisely, the classical MDS is based on the assumption that for some metric d (x, y) (distance function) and some increasing transformation β, ψ (x, y) =β (d (x, y)). (4) This is a prominent instance of what is called the probability-distance hypothesis in Dzhafarov (22b). Recall that the defining properties of a metric d are: (A) d (a, b) ; () d (a, b) =if and only if a = b; (C) d (a, c) d (a, b)+d (b, c); (D)d (a, b) =d (b, a). In addition one assumes

6 Dzhafarov and Colonius in MDS that metric d belongstoapredefined class, usually the class of Minkowski metrics with exponents between and 2. It immediately follows from (A), (), (D), and the monotonicity of β that for any distinct x, y, ψ (x, y) =ψ (y, x) ψ (x, x) =ψ ½(y, y) ψ (x, y) ψ (x, x) < ψ (y, x) (Symmetry) (Constant Self-Dissimilarity) (Regular Minimality) (5) We know from Chapter that while the property of Regular Minimality is indeed satisfied in all available experimental data, the property of Constant Self-Dissimilarity is not. The latter can clearly be seen in the table below, a excerpt from Rothkopf s (957) well-known study of Morse code discriminations. In his experiment a large number of respondents made same-different judgments in response to 6 6 auditorily presented pairs of Morse codes for letters of alphabet and digits. Ro 4 5 6 7 8 9 6 95 86 8 55 2 67 77 86 97 88 6 8 95 86 85 78 58 4 5 8 7 74 9 86 7 7 6 74 6 92 8 76 85 84 88 89 4 68 9 95 58 8 88 9 96 95 5 26 92 86 56 58 4 9 4 89 78 6 57 8 8 68 76 5 4 58 8 7 8 68 79 9 9 86 44 48 8 96 4 84 97 94 95 8 9 9 9 9 96 45 89 97 95 86 87 74 22 6 Regular Minimality here is satisfied in the canonical form, and one can see, for example, that the Morse code for digit 6 was judged different from itself by 5% of respondents, but only by 6% for digit 9. Symmetry is clearly violated as well: thus, digits 4 and 5 were discriminated from each other in 8% of cases when 5 was presented first in the two-code sequence, but in only 58% when 5 was presented second. Nonconstant Self-similarity This particular -code subset is chosen so that it forms a self-contained subspace of the 6 codes: a geodesic loop (as explained later) for any two of its elements is contained within the subset.

Distances from Discriminability 7 and Asymmetry are also manifest in the excerpt below from a similar study of Morse-code-like signals by Wish (967). 2 Wi S U W X 2 4 5 S 6 28 44 64 4 84 8 94 89 U 6 6 42 7 55 75 44 85 7 94 W 8 44 4 26 56 22 62 44 26 74 X 45 24 46 7 2 5 59 78 86 6 7 45 85 65 8 7 56 4 5 52 5 9 8 95 2 8 49 79 7 9 69 7 84 67 58 7 5 55 27 69 7 4 2 45 67 4 89 7 48 9 4 5 6 25 5 97 69 8 44 95 97 26 47 49 We can conclude therefore that MDS, or any other data-analytic technique based on the probability-distance hypothesis, is not supported by discrimination probability data. y contrast, Fechnerian Scaling in the case of discrete stimulus sets is only based on Regular Minimality, which is supported by data. Although prior to Dzhafarov (22d) Regular Minimality has not been formulated as a basic property of discrimination, independent of its other properties (such as Constant Self-Dissimilarity), the violations of Symmetry and Constant Self-Dissimilarity have long since been noted. Tversky s (977) contrast model and Krumhansl s (978) distanceand-density scheme are two best known theoretical schemes dealing with these issues. 2. MDFS MDFS (Multidimensional Fechnerian Scaling) is Fechnerian Scaling performed on a stimulus set whose physical description can be represented by an open connected region E of n-dimensional (n ) real-valued vectors, such that ψ (x, y) is continuous with respect to its Euclidean topology. This simply means that as (x k, y k ) (x, y), in the conventional sense, ψ (x k, y k ) ψ (x, y). The theory of Fechnerian Scaling has been developed 2 2 stimuli in this study were 5-element sequences T P T 2 P 2 T, where T stands for a tone (short or long) and P stands for a pause ( or units long). We arbitrarily labeled the stimuli A,,..., Z,,,..., 5, in the order they are presented in Wish s paper. The criterion for choosing this particular subset of stimuli is the same as for matrix Ro.

8 Dzhafarov and Colonius for continuous (arcwise connected) spaces of a much more general structure (Dzhafarov & Colonius, 25a), but a brief overview of MDFS should suffice for understanding the main ideas underlying Fechnerian Scaling. Throughout the entire discussion we tacitly assume that Regular Minimality is satisfied in a canonical form. 2.. Oriented Fechnerian distances in continuous spaces Refer to Fig.. Any a, b E can be connected by a smooth arc x (t), a piecewise continuously differentiable mapping of an interval [α, β] of reals into E, such that x (α) =a, x (β) =b. The main intuitive idea underlying Fechnerian Scaling is that (a) any point x (t) on this arc, t [α, β), can be assigned a local measure of its difference from its immediate neighbors, x (t + dt); (b) by integrating this local difference measure along the arc, from α to β, one can obtain the psychometric length of this arc; and (c) by taking the infimum (the greatest lower bound) of psychometric lengths across all possible smooth arcs connecting a to b one obtains the distance from a to b in space E. α t β E b a x(t) E Figure : The underlying idea of MDFS. [α, β] is a real interval, a x (t) b asmooth arc. The psychometric length of this arc is the integral of local difference of x (t) from x (t + dt), shown by vertical spikes along [α, β]. The inset shows that one should compute the psychometric lengths for all possible smooth arcs leading from a to b. Their infimum is the oriented Fechnerian distance from a to b.

Distances from Discriminability 9 As argued in Dzhafarov and Colonius (999), this intuitive scheme can be viewed as the essence of Fechner s original theory for unidimensional stimulus continua (Fechner, 86). The implementation of this idea in MDFS is as follows (see Fig. 2). ψ(x(t i ), x(t)), i =,2, or ψ(x(t), x(t i )), i =,2, α t 2 t t β x(t ) b E a x(t ) x(t 2 ) Figure 2: The local difference of x (t) from x (t + dt) (as dt +) atagivenpoint, t = t i, is the slope of the tangent line drawn to ψ (x (t i ), x (t)), ortoψ (x (t), x (t i )), at t = t i +. Using ψ (x (t i ), x (t)) yields derivatives of the first kind, using ψ (x (t), x (t i )) yields derivatives of the second kind. Their integration from α to β yields oriented Fechnerian distances (from a to b) of, respectively, first and second kind. As t for a smooth arc x (t) moves from α to β, the value of selfdiscriminability ψ (x (t), x (t)) may vary (Nonconstant Self-Dissimilarity property). Therefore, to see how distinct x (t) is from x (t + dt) it would not suffice to look at ψ (x (t), x (t + dt)) or ψ (x (t + dt), x (t)); oneshould compute instead the increments in discriminability φ () (x (t), x (t + dt)) = ψ (x (t), x (t + dt)) ψ (x (t), x (t)), φ (2) (x (t), x (t + dt)) = ψ (x (t + dt), x (t)) ψ (x (t), x (t)). (6) oth φ () and φ (2) are positive due to the Regular Minimality property (in a canonical form). They are referred to as psychometric differentials of the first kind (or in the first observation area) and second kind (in the second

Dzhafarov and Colonius observation area), respectively. The assumptions of MDFS guarantee that the cumulation of φ () (x (t), x (t + dt)) (i.e., integration of φ () (x (t), x (t + dt)) /dt+) from t = α to t = β always yields a positive quantity. We call this quantity the psychometric length of arc x (t) of the first kind, and denote it L () [a x b], where we use the suggestive notation for arc x connecting a to b: this notation is justified by the fact that the choice of the function x :[α, β] E is irrelevant insofar as the graph of the function (the curve connecting a to b in E) remains invariant. It can further be shown that the infimum of all such psychometric lengths L () [a x b], across all possible smooth arcs connecting a to b, satisfies all properties of a distance except for symmetry. Denoting this infimum by G (a, b), we have (A) G (a, b) ; () G (a, b) = if and only if a = b; (C) G (a,c) G (a, b) +G (b,c); but it is not necessarily true that G (a, b) =G (b, a). Such geometric constructs are called oriented distances. WecallG (a, b) the oriented Fechnerian distance of the first kind from a to b. y repeating the whole construction with φ (2) (x (t), x (t + dt)) in place of φ () (x (t), x (t + dt)) we will get the psychometric lengths L (2) [a x b] of the second kind (for arcs x (t) connecting a to b) and, as their infima, the oriented Fechnerian distances G 2 (a, b) of the second kind (from a to b). 2.2. MDFS and MDS The following observation provides additional justification for computing the oriented Fechnerian distances in the way just outlined. Ametricd (symmetrical or oriented) on some set S is called intrinsic if d (a, b) for any a, b S equals the infimum of the lengths of all allowable arcs connecting a and b (i.e., arcs with some specified properties). The Aside from Regular Minimality and continuity of ψ (x, y), the only other essential assumption of MDFS is that of the existence of a global psychometric transformation Φ which makes the limit ratios Φ lim s + φ (ι) (x (t), x (t + s)) s (ι =, 2) nonvanishing, finite, and continuous in (x (t), ẋ (t)), for all arcs. (Actually, this is the First Main Theorem of Fechnerian Scaling, a consequence of some simpler assumptions.) As it turns out (Dzhafarov, 22d), together with Nonconstant Self-Dissimilarity this implies that Φ (h) /h k> as h +. That is, Φ is a scaling transformation in the small and can therefore be omitted from formulations, on putting k =with no loss of generality. The uniqueness of extending Φ (h) =h to arbitrary values of h [, ] is analyzed in Dzhafarov & Colonius (25b). In the present chapter Φ (h) =h is assumed tacitly.

Distances from Discriminability oriented Fechnerian distances G (a, b) and G 2 (a, b) are intrinsic in this sense, provided the allowable arcs are defined as smooth arcs. In reference to the classical MDS, all Minkowski metrics are (symmetrical) intrinsic metrics, in the same sense. Assume now that the discrimination probabilities ψ (x, y) on E (with thesamemeaningasintheprevioussubsection)canbeobtainedfrom some symmetrical intrinsic distance d on E by means of (4), with β being a continuous increasing function. It is sufficient to assume that (4) holds for small values of d only. Then, as proved in Dzhafarov (22b), d (a, b) =G (a, b) =G 2 (a, b) for all a, b E. In other words, ψ (x, y) cannot monotonically and continuously depend on any (symmetrical) intrinsic metric other than the Fechnerian one. The latter in this case is symmetrical, and its two kinds G and G 2 coincide. 4 The classical MDS, including its modification proposed in Shepard and Carroll (966), falls within this category of models. In the context of continuous stimulus spaces, therefore, Fechnerian Scaling and MDS are not simply compatible, the former is in fact a necessary consequence of the latter (under the assumption of intrinsicality, and without confining the class of metrics d to Minkowski ones). Fechnerian computations, however, are applicable in a much broader class of cases, including those where the probability-distance hypothesis is false (as we know it generally to be). It should be noted for completeness that some non-classical versions of MDS are based on Trversky s (977) or Krumhansl s (978) schemes rather than on the probability-distance hypothesis, and they have the potential of handling nonconstant self-dissimilarity and/or asymmetry (e.g., DeSarbo et al., 992; Weeks & entler, 982). We do not review these approaches here. Certain versions of MDS can be viewed as intermediate between the classical MDS and FSDOS. Shepard and Carroll (966) discussed MDS methods where only sufficiently small distances are monotonically related to pairwise dissimilarities. More recently, this idea was implemented in 4 This account is somewhat simplistic: since the probability-distance hypothesis implies Constant Self-Dissimilarity, the theorem proved in Dzhafarov (22b) is compatible with Fechnerian distances computed with Φ other than identity function (see footnote ). We could avoid mentioning this by positing in the formulation of the probabilitydistance hypothesis that β (h) in (4) has a non-zero finite derivative at h =+. With this assumption, psychometric increments, hence also Fechnerian distances are unique up to multiplication by a positive constant. Equation d G G 2 therefore could more generally be written as d kg kg 2 (k>). Throughout this paper we ignore the trivial distinction between different multiples of Fechnerian metrics. (It should also be noted that in Dzhafarov (22b) intrinsic metrics are called internal, and a single distance G is used in place of G and G 2.)

2 Dzhafarov and Colonius two algorithms where large distances are obtained by cumulating small distances within stimulus sets viewed as manifolds embedded in Euclidean spaces (Tenenbaum, de Silva, & Langford, 2; Roweis & Saul, 2). When applied to discrimination probabilities, these modifications of MDS cannot handle nonconstant self-dissimilarity, but the idea of cumulating small differences can be viewed as the essence of Fechnerian Scaling. 2.. Overall Fechnerian distances in continuous spaces The asymmetry of the oriented Fechnerian distances creates a difficulty in interpretation. It is easy to understand that in general ψ (x, y) 6= ψ (y, x): stimulus x in the two cases belongs to two different observation areas and can therefore be perceived differently (the same being true for y). In G (a, b), however, a and b belong to the same (first) observation area, and the non-coincidence of G (a, b) and G (b, a) prevents one from interpreting either of them as a reasonable measure of perceptual dissimilarity between a and b (in the first observation area, from the point of view of a given perceiver). The same consideration applies, of course, to G 2. In MDFS this difficulty is resolved by taking as a measure of perceptual dissimilarity the overall Fechnerian distances G (a, b)+g (b, a) and G 2 (a, b)+g 2 (b, a). What justifies this particular choice of symmetrization istheremarkablefactthat G (a, b)+g (b, a) =G 2 (a, b)+g 2 (b, a) =G (a, b), (7) where the overall Fechnerian distance G (a, b) (we need not now specify of which kind) can be easily checked to satisfy all properties of a metric (Dzhafarov, 22d; Dzhafarov & Colonius, 25a). On a moment s reflection, (7) makes perfect sense. We wish to obtain a measure of perceptual dissimilarity between a and b, and we use the procedureofpairwisepresentationswithsame-different judgments to achieve this goal. The meaning of (7) is that in speaking of perceptual dissimilarities among stimuli one can abstract away from this particular empirical procedure. Caution should be exercised though: the observation-area-invariance of the overall Fechnerian distance is predicated upon the canonical form of Regular Minimality. In a more general case, as explained in Section.6, G (a, b)+g (b, a) equals G 2 a, b + G 2 (b, a ) if a and a (as well as b and b ) are Points of Subjective Equality, not necessarily physically identical. Equation (7) is an immediate consequence of the following proposition (Dzhafarov, 22d; Dzhafarov & Colonius, 25a): for any smooth arcs

Distances from Discriminability E E observation area observation area 2 b b a y x G(a, b) G (a, b) + G (b, a) ) a y x G(a, b) G 2 (a, b) + G 2 (b, a) ) Figure : Illustration for the Second Main Theorem: the psychometric length of the first kindofaclosedloopfroma to b and back equals the psychometric length of the second kind for the same loop traversed in the opposite direction. This leads to the equality of the overall Fechnerian distances in the two observation areas.

4 Dzhafarov and Colonius a x b and b y a, L () [a x b]+l () [b y a] = L (2) [a y b]+l () [b x a]. (8) Put differently, the psychometric length of the first kind for any closed loop containing a and b equals the psychometric length of the second kind for the same closed loop but traversed in the opposite direction. Together (8) and its corollary (7) constitute what we call the Second Main Theorem of Fechnerian Scaling (see Fig. ). This theorem plays a critical role in extending the continuous theory to discrete and other, more complex object spaces (Dzhafarov & Colonius, 25b).. FSDOS FSDOS stands for Fechnerian Scaling of Discrete Object Sets. The mathematical simplicity of this special case of Fechnerian Scaling allows us to presentitinagreaterdetailthanwedidmdfs... Discrete object spaces Recall that a space of stimuli (or objects) is a set S of all objects of a particular kind endowed with a discrimination probability function ψ (x, y). For any x, y S we define psychometric increments of the first and second kind (or, in the first and second observation areas) as, respectively, φ () (x, y) =ψ (x, y) ψ (x, x), φ (2) (x, y) =ψ (y, x) ψ (x, x). (9) Psychometric increments of both kinds are positive due to (a canonical form of) Regular Minimality, (). A space S is called discrete if, for any x S, h i h i inf φ () (x, y) >, inf φ (2) (x, y) >. y y In other words, the psychometric increments of either kind from x to other stimuli cannot fall below some positive quantity. Intuitively, other stimuli cannot get arbitrarily close to x. Clearly, stimuli in a discrete space cannot be connected by arcs (continuous images of intervals of reals).

Distances from Discriminability 5.2. Main idea To understand how Fechnerian computations can be made in discrete spaces, let us return for a moment to continuous spaces E discussed in the previous section. φ () (x(t i ), x(t i+ )) or φ (2) (x(t i ), x(t i+ )) x(t i ) x(t i+ ) E α t 2 t t t 4 t 5 β b x(t 5 ) a x(t ) x(t 4 ) E x(t ) x(t 2 ) Figure 4: The psychometric length of the first (second) kind of an arc can be approximated by the sum of psychometric increments of the first (second) kind chained along the arc. The right insert shows that if E is represented by a dense grid of points, the Fechnerian computations involve taking all possible chains leading from one point to another through successions of immediately neighboring points. Consider a smooth arc x (t), x :[α, β] E, x (α) =a, x (β) =b, as shown in Fig. 4. We know that its psychometric length L (ι) [a x b] of the ιth kind (ι =, 2) is obtained by cumulating psychometric differentials (6) of the same kind along this arc. It is also possible, however, to approximate L (ι) [a x b] by subdividing [α, β] into α = t,t,..., t k,t k+ = β

6 Dzhafarov and Colonius and computing the sum of the chained psychometric increments L () [x (t ), x (t ),..., x (t k+ )] = kx φ (ι) (x (t i ), x (t i+ )). () As shown in Dzhafarov and Colonius (25a), by progressively refining the partitioning, max i {t i+ t i }, this sum can be made as close as one wishes to the value of L (ι) [a x b]. In practical computations, E (which, we recall, is an open connected region of n-dimensional vectors of reals) can be represented by a sufficiently dense discrete grid of points. In view of the result just mentioned, the oriented Fechnerian distance G ι (a, b) (ι =, 2) between any a and b in this case can be approximated by (a) considering all possible chains of successive neighboring points leading from a to b, (b) computing sums () for each of these chains, and (c) taking the smallest value. This almost immediately leads to the algorithm for Fechnerian computations in discrete spaces. The main difference is that in discrete spaces we have no physical ordering of stimuli to rely on, and the notion of neighboring points is not defined. In a sense, every point in a discrete space can be viewed as a potential neighbor of any other point. Consequently, in place of all possible chains of successive neighboring points leading from a to b onehastoconsidersimplyall possible chains of points leading from a to b (see Fig. 5). i=.. Illustration Returningtoourtoyexample(matrixTOY, reproduced here for reader s convenience together with L and G ), let us compute the Fechnerian distance between, say, objects D and. TOY A C D A..8.8..6 C.6.9.5.7 D.6.9.5

Distances from Discriminability 7 φ () (x i, x i+ ) or φ (2) (x i, x i+ ) x i x i+ a b Figure 5: In a discrete space ( elements whereof are shown in an arbitrary spatial arrangement) Fechnerian computations are performed by taking sums of psychometric increments (of the first or second kind, as shown in the inset) for all possible chains leading from one point to another.

8 Dzhafarov and Colonius L A C D A A AC CAC DAD ACA CC DCD C ACA C C DCD D ADA DC CDC D G A C D A...9. C.9.7 D..7 The whole stimulus space here consists of four stimuli, {A,, C, D}, and we have five different chains in this space which are comprised of distinct (nonrecurring) objects and lead from D to : D, DA, DC, DAC, DCA. We begin by computing their psychometric lengths of the first kind, L () [D], L () [DA], etc. y analogy with (), L () [DCA], for example, is computed as L () [DCA] = φ () (D, C)+φ () (C, A)+φ () (A, ) = [ψ (D, C) ψ (D, D)] + [ψ (C, A) ψ (C, C)] +[ψ (A, ) ψ (A, A)] = [.7.5] + [..5] + [.8.] =.4. We have used here the definition of φ () (x, y) givenin(9).repeatingthis procedure for all our five chains, we will find out that the smallest value is provided by L () [DC] = φ () (D, C)+φ () (C, ) = [ψ (D, C) ψ (D, D)] + [ψ (C, ) ψ (C, C)] = [.7.5] + [.6.5] =.. Note that this value is smaller than the length of the one-link chain ( direct connection ) D: L () [D] =φ () (D, ) =ψ (D, ) ψ (D, D) =..5 =.5. The chain DC canbecalledageodesic chain connecting D to. (Generally there can be more than one geodesic chain, of the same length, for a given pair of stimuli, but in our toy example all geodesics are unique.) Its length is taken to be the oriented Fechnerian distance of the first kind from D to, G (D, ) =..

Distances from Discriminability 9 Consider now the same five chains but viewed in the opposite direction, that is, all chains in {A,, C, D} leading from to D, and compute for these chains the psychometric lengths of the first kind: L () [D], L () [AD], etc. Having done this we will find out that this time the shortest chain is the one-link chain D, with the length L () [D] =φ () (, D) =ψ (, D) ψ (, ) =.9. =.8. The geodesic chain from to D therefore is D, and the oriented Fechnerian distance of the first kind from to D is G (, D) =.8. Using the same logic as for continuous stimulus spaces, we now compute the (symmetrical) overall Fechnerian distance between D and by adding the two oriented distances to and fro, G (D, ) =G (, D) =G (D, )+G (, D) =.+.8 =.. This is the value we find in cells (D, ) and (, D) of matrix G. The concatenation of the two geodesic chains, DC and D, forms the geodesic loop between D and, which we find in cells (D, ) and (, D) of matrix L. This loop, of course, can be written in three different ways depending on which of its three distinct elements we choose to begin and end with. The convention adopted in matrix L istobeginandendwiththerow object: DCD in cell (D, ) and DC in cell (, D). Note that the overall Fechnerian distance G (D, ) and the corresponding geodesic loop could also be found by computing psychometric lengths for all 25 possible closed loops containing objects D and in space {A,, C, D} and finding the smallest. This, however, would be a more wasteful procedure. The reason we do not need to add the qualification of the first kind to the designations of the overall Fechnerian distance G (D, ) and the geodesic loop DCD is that precisely the same value of G (D, ) and the same geodesic loop (only traversed in the opposite direction) are obtained if the computations are performed with psychometric increments of the second kind. For chain DCA, for example, the psychometric length of the second

2 Dzhafarov and Colonius kind, using the definition of φ (2) in (9), is computed as L (2) [DCA] = φ (2) (D, C)+φ (2) (C, A)+φ (2) (A, ) = [ψ (C, D) ψ (D, D)] + [ψ (A, C) ψ (C, C)] +[ψ (, A) ψ (A, A)] = [..5] + [.6.5] + [.8.] =.. Repeating this computation for all our five chains leading from D to, the shortest chain will be found to be D, with the length L (2) [D] =φ (2) (D, ) =ψ (, D) ψ (D, D) =.9.5 =.4, takentobethevalueofg 2 (D, ), the oriented Fechnerian distance form D to of the second kind. For the same five chains but viewed as leading from to D, the shortest chain is CD, with the length L (2) [CD] = φ (2) (, C)+φ (2) (C, D) = [ψ (C, ) ψ (, )] + [ψ (D, C) ψ (C, C)] = [.6.] + [.7.5] =.7 takentobethevalueofg 2 (, D), the oriented Fechnerian distance form to D of the second kind. Their sum is G (D, ) =G (, D) =G 2 (D, )+G 2 (, D) =.4+.7 =., precisely the same value for the overall Fechnerian distance as before (even though the oriented distances are different). The geodesic loop obtained by concatenating the geodesic chains D and CD is also the same as we find in matrix L in cells (D, ) and (, D), but read from right to left: DCD in cell (D, ) and CD in cell (, D). The complete formulation of the convention adopted in L therefore is as follows: the geodesic loop in cell (x, y) begins and ends with x and is read from left to right for the computations of the first kind, and from right to left for the computations of the second kind (yielding one and the same result, the overall Fechnerian distance between x and y)..4. Procedure of FSDOS It is clear that any finite set S ={s, s 2,..., s N } endowed with probabilities p ij = ψ (s i, s j ) forms a discrete space in the sense of our formal definition. As this case is of the greatest interest in empirical applications, in the following we will confine our discussion to finite object spaces. All our

Distances from Discriminability 2 statements, however, unless specifically qualified, apply to discrete object spaces of arbitrary cardinality. The procedure below is described as if one knew the probabilities p ij on the population level. If sample sizes do not warrant this approximation, the procedure should ideally be repeated with a large number of matrices p ij that are statistically retainable given a matrix of frequency estimates ˆp ij. We return to this issue in the concluding section. The computation of Fechnerian distances G ij among {s, s 2,..., s N } proceeds in several steps. The firststepinthecomputationistocheckfor Regular Minimality: for any i and all j 6= i, p ii < min {p ij,p ji }. If Regular Minimality is violated (on the population level), FSDOS will not work. Put differently, given a matrix of frequency estimates ˆψ (s i, s j ), one should use statistically retainable matrices of probabilities p ij that do satisfy Regular Minimality; and if no such matrices can be found, FSDOS is not applicable. The theory of Fechnerian Scaling treats Regular Minimality as the defining property of discrimination. If it is not satisfied, something can be wrong in the procedure: for collective perceivers, e.g., substantially different groups of people could be responding to different pairs of stimuli (violating thereby the requirement of having a single perceiver ), or the semantic meaning of the responses same and different could vary from one pair of stimuli to another. (Alternatively, of course, the theory of Fechnerian Scaling may be wrong itself, which would be a preferable conclusion if regular Minimality was found to be violated systematically, or at least not very rarely.) Having Regular Minimality verified, we compute psychometric increments of the first and second kind, φ () (s i, s j ) = p ij p ii, φ (2) (s i, s j ) = p ji p ii, which are positive for all j 6= i. Consider now a chain of stimuli s i = x, x 2,...,x k = s j leading from s i to s j,withk 2. The psychometric length of the first kind for this chain, L () [x, x 2,..., x k ], is defined as the sum of the psychometric increments φ () (x m, x m+ ) taken along this chain, L () [x, x 2,..., x k ]= kx φ () (x m, x m+ ). m=

22 Dzhafarov and Colonius The set of different psychometric lengths across all possible chains of distinct elements connecting s i to s j being finite, it contains a minimum value L () min (s i, s j ). (The consideration can always be confined to chains (x, x 2,..., x k ) of distinct elements, because if x l = x m (l<m), the length L () cannot increase if the subchain (x l+,...,x m ) is removed.) This value is called the oriented Fechnerian distance of the first kind from object s i to object s j : G (s i, s j )=L () min (s i, s j ). It is easy to prove that the oriented Fechnerian distance satisfies all properties of a metric, except for symmetry: (A) G (s i, s j ) ; () G (s i, s j )= if and only if i = j; (C)G (s i, s j ) G (s i, s m )+G (s m, s j ); but in general, G (s i, s j ) 6= G (s j, s i ). 5 In according with the general logic of Fechnerian Scaling, G (s i, s j ) is interpreted as the oriented Fechnerian distance from s i to s j in the first observation area. Any chain from s i to s j whose elements are distinct and whose length equals G (s i, s j ) is a geodesic chain from s i to s j. Theremaybemorethan onegeodesicchainforgivens i, s j. (Note that in the case of infinite discrete sets mentioned in footnote 5 geodesic chains need not exist.) The oriented Fechnerian distances G 2 (s i, s j ) of the second kind (in the second observation area) and the corresponding geodesic chains are computed analogously, using the chained sums of psychometric increments φ (2) instead of φ (). As argued earlier (Section 2.), the order of two stimuli in a given observation area has no operational meaning, and we add the two oriented distances, to and fro, to obtain the (symmetrical) overall Fechnerian distances G ij = G (s i, s j )+G (s j, s i )=G ji, G ij = G 2 (s i, s j )+G 2 (s j, s i )=G ji. G ij clearly satisfies all the properties of a metric. The validation for this procedure (and for writing G ij without indicating 5 Properties (A) and () trivially follow from the fact that for i 6= j, G (s i, s j ) is the smallest of several positive quantities, L () [x, x 2,...,x k ]. Property (C) follows from the observation that the chains leading from s i to s j through a fixed s k form a proper subset of all chains leading from s i to s j. For an infinite discrete S, L () min (a, b) (a, b S) need not exist and should be replaced with L () inf (a, b), the infimum of L() [x, x 2,..., x k ] for all finite chains of distinct elements with a = x and x k = b (x, x 2,...,x k S). The argument for properties (A) and () then should be modified: for a 6= b, G (a, b) > because L () inf (a, b) inf x φ () (a, x), and by definition of discrete object spaces, inf x φ () (a, x) >.

Distances from Discriminability 2 observation area) is provided by the fact that G (s i, s j )+G (s j, s i )=G 2 (s i, s j )+G 2 (s j, s i ), () i.e., the distance G ij between the ith and the jthobjectsdoesnotdepend on the observation area in which these objects are taken. This fact is a consequence of the following statement, which is of interest on its own sake: for any two chains s i = x, x 2,..., x k = s j and s i = y, y 2,...,y l = s j (connecting s i to s j ), L () [x, x 2,...,x k ]+L () [y l, y l,..., y ] = L (2) [y, y 2,...,y l ]+L (2) [x k, x k,..., x ]. (2) As the proof of this statement is elementary, it may be useful to present it here. Denoting p ij = ψ (x i, x j ) and p ij = ψ (y i, y j ), = = L () [x, x 2,..., x k ]+L () [y l, y l,..., y ] k X m= p m,m+ p Xl ii + p m+,m p m+,m+, m= L (2) [y, y 2,..., y l ]+L (2) [x k, x k,..., x ] Xl m= k p m+,m p X m,m + p m,m+ p m+,m+. m= Subtracting the second equation from the first, ³ L () [x, x 2,..., x k ] L (2) [x k, x k,..., x ] = ³ L () [y l, y l,..., y ] L (2) [y, y 2,...,y l ] + Ã k X + m= Ã l X m= k p m,m+ p X ii m= p m,m+ p! m+,m+ p m+,m p Xl m+,m+ p m+,m p! m,m m= = (p kk p )+(p p kk). ut p = p = p ii and p kk = p kk = p jj, where, we recall, p ij = ψ (s i, s j ). The difference therefore is zero, and (2) is proved. Equation () follows

24 Dzhafarov and Colonius as a corollary, on observing G (s i, s j )+G (s j, s i ) = infl () [x, x 2,..., x k ]+infl () [y l, y l,..., y ] n o = inf L () [x, x 2,..., x k ]+L () [y l, y l,..., y ] n o = inf L (2) [y, y 2,..., y l ]+L (2) [x k, x k,..., x ] = infl (2) [y, y 2,..., y l ]+infl (2) [x k, x k,..., x ] = G 2 (s j, s i )+G 2 (s i, s j ). Together () and (2) provide a simple version of the Second Main Theorem of Fechnerian Scaling, mentioned earlier, when discussing MDFS. An equivalent way of defining the overall Fechnerian distances G ij is to consider all closed loops x, x 2,..., x n, x (n 2) containing two given stimuli s i, s j : G ij is the shortest of the psychometric lengths computed for all such loops. Note that the psychometric length of a loop depends on the direction in which it is traversed: generally, L () (x, x 2,..., x n, x ) 6= L () (x, x n,..., x 2, x ), L (2) (x, x 2,..., x n, x ) 6= L (2) (x, x n,..., x 2, x ). The result just demonstrated tells us, however, that L () (x, x 2,..., x n, x )=L (2) (x, x n,...,x 2, x ), that is, any closed loop in the first observation area has the same length as thesameclosedlooptraversedintheoppositedirectioninthesecondobservation area. In particular, if x, x 2,..., x n, x is a geodesic (i.e., shortest) loop containing the objects s i, s j in the first observation area (obviously, the concatenation of the geodesic chains connecting s i to s j and s j to s i ), then the same loop is a geodesic loop in the second observation area, if traversed in the opposite direction, x, x n,..., x 2, x. The computational procedure of FSDOS is summarized in the form of a detailed algorithm presented in the appendix at the end of this chapter..5. Two examples We used the procedure just described to compute Fechnerian distances and geodesic loops among 6 Morse codes with pairwise discrimination probabilities reported in Rothkopf (957), and among 2 Morse-code-like signals data with discrimination probabilities reported in Wish (967). For typographic reasons only small subsets of these stimulus sets are shown in ma-

Distances from Discriminability 25 trices Ro and Wi in Section., chosen because they form self-contained subspaces: any two elements of such a subset can be connected by a geodesic loop lying entirely within the subset. The Fechnerian distances and geodesic loops are presented here for these subsets only: for matrix Ro they are G Ro 4 5 6 7 8 9 5 42 95 97 6 57 77 4 57 5 48 6 5 47 27 99 6 7 42 48 2 64 47 25 28 6 2 95 6 2 68 95 27 45 65 69 4 97 5 64 68 6 8 6 7 74 5 6 47 47 95 6 4 6 24 4 6 57 27 25 27 8 4 44 92 8 7 77 99 28 45 6 6 44 6 8 8 4 6 6 65 7 24 92 6 26 9 57 7 2 69 74 4 8 8 26 L Ro 4 5 6 7 8 9 5 4 5 565 5675 567875 975 4 5 67 7 8 9 4 5 6 7 8 9 5 4 5 6 67 8 9 4 44 44 44 44 4 454 4564 474 484 494 5 55 55 55 55 545 5 565 5675 567875 5975 6 6556 676 66 66 6456 656 6 676 6786 678986 7 75567 77 77 767 747 7567 767 7 787 7897 8 8755678 88 88 88 848 875678 8678 878 8 898 9 9759 99 99 99 949 9759 986789 9789 989 9 and for matrix Wi they are 6 6 In the complete 2 2 matrix reported in Wish (967) (but outside the submatrix Wi) there are two violations of Regular Minimality, both due to a single

26 Dzhafarov and Colonius G Wi S U W X 2 4 5 S 2 72 89 57 9 2 28 9 8 U 2 76 79 89 7 8 6 7 28 W 72 76 9 55 22 67 58 79 X 89 79 2 67 94 9 45 49 57 89 9 2 7 4 95 2 9 7 55 67 9 72 8 2 2 8 22 94 7 9 6 92 74 28 6 67 9 4 6 84 77 4 9 7 58 45 95 72 92 84 68 5 8 28 79 49 2 8 74 77 68 L Wi S U W X 2 4 5 S S USU WSW XSUX S WSU 2USU2 XSUX 4WSUX4 5XSUX5 U SUS U WUW XWUX SUS WU 2U2 WUX 4WUX4 5XWUX5 W SWS UWU W XWX WS W 2XW2 WX 4WX4 5XWX5 X SUXS UXWU WXW X X WX 2X2 X 4X4 5X5 SS USSU WSW XX 22 44 525 SUWS UWU WW XWX 22 44 5X5 2 SU2US U2U W2XW X2X 2 2 2 2 424 525 SUXXS UXWU WXW XX 22 4XX4 5X5 4 SUX4WS UX4WU WX4W X4X 4 4 242 X4X 4 545 5 SUX5XS UX5XWU WX5XW X5X 25 5X 252 5X 454 5 Recall our convention on presenting geodesic loops. Thus, in matrix L Ro the geodesic chain from letter to digit 8 in the first observation area is 5 6 7 8 and that from 8 to is 8 7 5. In the value, ˆp TV =.: this value is the same as ˆp VV and smaller than ˆp TT =.6 (using the labeling of stimuli described in Section.), footnote 2. As Wish s data are used here for illustration purposes only, we simply replaced ˆp TV =. with p TV =.7, putting p ij = ˆp ij for the rest of the data. Chi-square deviation of thus defined matrix of p ij from the matrix of ˆp ij is negligibly small. A more comprehensive procedure should have involved a repeated generation of statistically retainable p ij matrices subject to Regular Minimality, as discussed in the concluding section.

Distances from Discriminability 27 second observation area the geodesic chains should be read from right to left: 8 7 5 from to 8, and 5 6 7 8 from 8 to. The oriented Fechnerian distances (lengths of the geodesic chains) are G (,8) =.7, G (8,)=.7, G 2 (,8) =.77, andg 2 (8,)=.6. The lengths of the closed loops in both observation areas add up to the same value, G(8,)=.4, as they should. Note that Fechnerian distances G ij are not monotonically related to discrimination probabilities p ij : there is no functional relationship between the two because the computation of G ij for any given (i, j) involves p ij values for all (i, j). Nor are the oriented Fechnerian distances G (s i,s j ) and G 2 (s i,s j ) monotonically related to psychometric increments p ij p ii and p ji p ii, due to the existence of longer-than-one-link geodesic chains. There is, however, a strong positive correlation between p ij and G ij :.94for Rothkopf s data and.89 for Wish s data (the Pearson correlation for the entire matrices, 6 6 and 2 2). This indicates that the probabilitydistance hypothesis, even if known to be false mathematically, may still be acceptable as a crude approximation. We may expect consequently that MDS-distances could provide crude approximations to the Fechnerian distances. That the adjective crude cannot be dispensed with is indicated by the relatively low values of Kendall s correlation between p ij and G ij :.76 for Rothkopf s data and.68 for Wish s data..2..2. 2 2 Dimension 2 2222-22- 22222-2-4 222-8 2-6 2-2222-9 22-7 -5 Dimension A 2222-2-4 22222-22- 2222-9 222-8 2-22-7-5 2-6 Dimension Figure 6: Two-dimensional Euclidean representations for discrimination probabilities (nonmetric MDS, Panel A) and for Fechnerian distances in matrix G Ro (metric MDS, Panel ). The MDS program used is PROXSCAL. in SPSS.5, minimizing raw stress. Sequence of s and 2 s preceding a dash is the Morse code for the symbol following the dash. Insets are scree plots (normalized raw stress versus number of dimensions). MDS can be used in conjunction with FSDOS, as a follow-up analysis once Fechnerian distances have been computed. A nonmetric version of

28 Dzhafarov and Colonius...6.6.2 2 4.2 2 4 LLL SLL LLL SLL SLL LLL LLS LLS SLS LLS LLL SLS Dimension 2 LSL LSS SLS A LSL LSS LLS SLL SLS Dimension Dimension Figure 7: Same as Fig. 6, but for discrimination probabilities (nonmetric MDS, Panel A) and for Fechnerian distances in matrix G Wi (metric MDS, Panel ). L stands for long tone, S for short tone, while digits an show the lengths of the two pauses. MDS can be applied to Fechnerian distances (as opposed to discrimination probabilities directly) simply to provide a rough graphical representation for matrices like Ro and Wi. More interestingly, a metric version of MDS can be applied to Fechnerian distances to test the hypothesis that Fechnerian distances, not restricted a priori to any particular class (except for beingintrinsic),defactobelongtoaclassofeuclideanmetrics(or,more generally, Minkowski ones), at least approximately; the degree of approximation for any given dimensionality is measured by the achieved stress value. Geometrically, metric MDS on Fechnerian distances is an attempt to isometrically embed the discrete object space into a low-dimensional Euclidean (or Minkowskian) space. Isometric embedment (or immersion) means mapping without distorting pairwise distances. Figures 6 and 7 provide a comparison of the metric MDS on Fechnerian distances (matrices Ro, W i) with nonmetric MDS performed on discrimination probabilities directly (matrices G Ro,G Wi ). Using the value of normalized raw stress as our criterion, the two-dimensional solution is almost equally good in both analyses. Therefore, to the extent we consider the traditional MDS solution acceptable, we can view the Fechnerian distances in these two cases as being approximately Euclidean. The configurations of points obtained by performing the metric MDS on Fechnerian distances and nonmetric MDS

Distances from Discriminability 29 on discrimination probabilities are more similar in Fig. 6 than in Fig. 7, indicating that MDS-distances provide a better approximation to Fechnerian distances in the former case. This may reflect the fact that the correlation between the probabilities and Fechnerian distances for Rothkopf s data is higher than for Wish s data (.94 vs.89). A detailed comparison of the configurations provided by the two analyses, as well as such related issues as interpretation of axes are, however, beyond the scope of this chapter..6. General form of Regular Minimality In continuous stimulus spaces it often happens that Regular Minimality does not hold in a canonical form: for a fixed value of x, ψ(x, y) achieves its minimum not at y = x butatsomeothervalueofy. It has been noticed since Fechner (86), for example, that when one and the same stimulus is presented twice in a succession, the second presentation often seems larger (bigger, brighter, etc.) than the first: this is the classical phenomenon of time error. It follows that in a successive pair of unidimensional stimuli, (x, y), the two elements maximally resemble each other when y is physically smaller than x. Other examples were discussed in Chapter. Although it is possible that in discrete stimulus spaces Regular Minimality always holds in a canonical form, it need not be so a priori. Returning once again to our toy example, assume that matrix TOY was the result of a canonical relabeling of matrix TOY, TOY with the correspondence table y a y b y c y d x a.6.6..8 x b.9.9.8. x c.5.6 x d.5.7 O x a x b x c x d O 2 y c y d y b y a common label A C D where O and O 2, as usual, denote the two observation area (row stimuli and column stimuli). Having performed the Fechnerian analysis on TOY and having computed the matrices L and G, it makes sense now to return to the original labeling (using the table of correspondences above) and present the Fechnerian distances and geodesic loops separately for the first and the second observation areas: