PROBLEMS OF CAUSAL ANALYSIS IN THE SOCIAL SCIENCES

Similar documents
WHEN ARE PROBABILISTIC EXPLANATIONS

Scientific Explanation- Causation and Unification

PATRICK SUPPES** Stanford University

Generative Techniques: Bayes Rule and the Axioms of Probability

Evidence and Theory in Physics. Tim Maudlin, NYU Evidence in the Natural Sciences, May 30, 2014

Structure learning in human causal induction

Sociology 6Z03 Topic 10: Probability (Part I)

Lecture 1: Brief Review on Stochastic Processes

Stochastic Processes

2) There should be uncertainty as to which outcome will occur before the procedure takes place.

Patrick Supped and J. Acacicb de Barros1

On Objectivity and Models for Measuring. G. Rasch. Lecture notes edited by Jon Stene.

FACTORIZATION AND THE PRIMES

To Infinity and Beyond. To Infinity and Beyond 1/43

PROOF-THEORETIC REDUCTION AS A PHILOSOPHER S TOOL

ANALYTIC COMPARISON. Pearl and Rubin CAUSAL FRAMEWORKS

Metric spaces and metrizability

CS 237: Probability in Computing

Chapter 4. Basic Set Theory. 4.1 The Language of Set Theory

Measurement Independence, Parameter Independence and Non-locality

To Infinity and Beyond

Structural Uncertainty in Health Economic Decision Models

Introduction to Probability

A General Overview of Parametric Estimation and Inference Techniques.

The Solvability of Probabilistic Regresses. A Reply to Frederik Herzberg

Volume 31, Issue 1. A gender-adjusted measure of literacy. Sreenivasan Subramanian Madras Institute of Development Studies

Introductory notes on stochastic rationality. 1 Stochastic choice and stochastic rationality

What Causality Is (stats for mathematicians)

On the Evolution of the Concept of Time

DD Advanced Machine Learning

1.2 The Role of Variables

PHILOSOPHY OF PHYSICS (Spring 2002) 1 Substantivalism vs. relationism. Lecture 17: Substantivalism vs. relationism

Special Theory Of Relativity Prof. Shiva Prasad Department of Physics Indian Institute of Technology, Bombay

Diary of Mathematical Musings. Patrick Stein

Connectedness. Proposition 2.2. The following are equivalent for a topological space (X, T ).

So, what are special sciences? ones that are particularly dear to the author? ( Oh dear. I am touched. Psychology is just, so, well, special!

Definition of geometric vectors

Causal Inference. Prediction and causation are very different. Typical questions are:

A SYSTEM VIEW TO URBAN PLANNING: AN INTRODUCTION

Avoiding the Block Universe: A Reply to Petkov Peter Bokulich Boston University Draft: 3 Feb. 2006

Section 3.1: Direct Proof and Counterexample 1

Gaussian Quiz. Preamble to The Humble Gaussian Distribution. David MacKay 1

Flexible Estimation of Treatment Effect Parameters

Semester I BASIC STATISTICS AND PROBABILITY STS1C01

Confidence Intervals

Advanced Statistical Methods for Observational Studies L E C T U R E 0 1

Discrete Probability Refresher

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

ALBERT EINSTEIN AND THE FABRIC OF TIME by Gevin Giorbran

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Causal Discovery. Beware of the DAG! OK??? Seeing and Doing SEEING. Properties of CI. Association. Conditional Independence

15 Skepticism of quantum computing

Bayesian Inference. 2 CS295-7 cfl Michael J. Black,

Computability Theory

TOOLING UP MATHEMATICS FOR ENGINEERING*

Paul D. Thorn. HHU Düsseldorf, DCLPS, DFG SPP 1516

Conditional probabilities and graphical models

Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi

A Detailed Derivation of the Distribution of Class Identification in a Chance Society :

Proving Completeness for Nested Sequent Calculi 1

Introductory Quantum Chemistry Prof. K. L. Sebastian Department of Inorganic and Physical Chemistry Indian Institute of Science, Bangalore

UMASS AMHERST MATH 300 SP 05, F. HAJIR HOMEWORK 8: (EQUIVALENCE) RELATIONS AND PARTITIONS

Lecture - 30 Stationary Processes

Lecture 22: Quantum computational complexity

Probability theory basics

FERMAT S TEST KEITH CONRAD

Rapid Introduction to Machine Learning/ Deep Learning

Intuitive infinitesimals in the calculus

External validity, causal interaction and randomised trials

Predicting the Treatment Status

Deep Metaphysical Indeterminacy

Section 7.1: Functions Defined on General Sets

Basic Probability. Introduction

2. Introduction to commutative rings (continued)

DEEP METAPHYSICAL INDETERMINACY

Chapter ML:IV. IV. Statistical Learning. Probability Basics Bayes Classification Maximum a-posteriori Hypotheses

35 Chapter CHAPTER 4: Mathematical Proof

An Absorbing Markov Chain Model for Problem-Solving

A Crucial Mistake in the Free Will Debate

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Do not copy, post, or distribute

2.1 Lecture 5: Probability spaces, Interpretation of probabilities, Random variables

Introduction to Metalogic

where Female = 0 for males, = 1 for females Age is measured in years (22, 23, ) GPA is measured in units on a four-point scale (0, 1.22, 3.45, etc.

Mathematics and Language

Probabilistic Reasoning. (Mostly using Bayesian Networks)

Introduction to Algebra: The First Week

A Note on the Existence of Ratifiable Acts

On the Triangle Test with Replications

Advanced Statistical Methods for Observational Studies L E C T U R E 0 1

ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS

Energy Transformations IDS 101

Mathematics-I Prof. S.K. Ray Department of Mathematics and Statistics Indian Institute of Technology, Kanpur. Lecture 1 Real Numbers

The World According to Wolfram

Tree sets. Reinhard Diestel

Machine Learning. Bayes Basics. Marc Toussaint U Stuttgart. Bayes, probabilities, Bayes theorem & examples

Agency and Interaction in Formal Epistemology

University of Regina. Lecture Notes. Michael Kozdron

A Little Deductive Logic

Transcription:

Patrick Suppes PROBLEMS OF CAUSAL ANALYSIS IN THE SOCIAL SCIENCES This article is concerned with the prospects and problems of causal analysis in the social sciences. On the one hand, over the past 40 years there has developed an elaborate and powerful statistical methodology. The depth of these developments far exceeds that of any other time in the history of statistics. On the other hand, only very recently have very explicit causal analyses been the focus of extensive discussion in the social sciences. It is my feeling that the various developments in economics, philosophy, and sociology are coming together to give a new and highly applicable synthesis of ideas about causality. I shall not provide a detailed bibliography of the developments I refer to, because that has been done rather recently in a complete way by Paul Humphreys [unpublished], but what I shall concentrate on is the problem of introducing more structure into causal analysis. 1. FORMAL DEFINITION The philosophical literature has tended to emphasize discussion of causality in terms of events, but the statistical and social science literature almost uniformly formulates causal concepts in terms of Epistemologia V (19821, Numero Speciale - Special Issue, pp. 239-250.

2

Problems of Causal Analysis 24 1 P If Y,I is a prima facie quadrant cause of X, and g the variances of YtI and X, exist as <well as their covariance, and if neither variance is zero, then the correlation of X, and Y,! is nonnegative. Obviously this definition is too simple for any complete account. It is apparent that prima facie causes can be spurious. A typical and familiar example would be the falling of a barometer preceding a storm. Given our modern knowledge of meteorology, we do not believe that the falling barometer is a genuine cause of the storm. For this reason I introduce the notion of a spurious cause. Spurious Cause. The definition is an obvious extension of the first one. A spurious cause must be a prima facie cause. (D2) A property Y,, is a spurious quadrant cause Of x, if and only if there is a t < t and a property Ztll such that (i) Y,, is a weak prima facie quadrant cause of X,, (ii) For all x, y and z if P( Ut, Z y, Zp 2 z) > O, then P(X, 2 x I Y,, 2 y, Z, 2 z) = P(X, Z x 1 Z, 2 z). In the literature of applied statistics, the notion most closely corresponding to that of spurious quadrant causality is that of spurious correlation. Roughly speaking, two random variables are said to be spuriously correlated if the correlation between them can be shown to vanish when a third variable is introduced and held constant. The necessity of investigating the possibility of spurious correlation before using the existence of a correlation to make an inference about causal relations has long been recognized in statistics. A good detailed discussion of the causal Significance of spurious correlations is to be found in Simon [ 19541. Corresponding to the preceding theorem about correlations, we

Problems of Causal Analysis 243 The models I consider apply to an experimental situation which consists of a sequence of trials. On each trial the subject of the experiment makes a response, which is followed by a reinforcing event. Thus an experiment may be represented by a sequence (A, El,A2, E2,...,A,,E,,...) of random variables, where the choice of letters follows conventions established in the literature: the value of the random variable A, is a number j representing the actual response on trial n, and the value of E, is a number k rep- resenting the reinforcing event on trial n. The relevant data on each trial may then be represented by an ordered pair (j, k) of integers with l j r, and O < k t, that is, we envisage in general r responses and t + l reinforcing events. Any sequence of these pairs of integers is a sequence of values of the random variables and thus represents a possible experimental outcome. The general aim of the theory is to predict the probability distribution of the response random variable when a particular distribution, or class of distributions, is imposed on the reinforcement random variable. The theory is formulated for the probability of a response on trial n + l given the entire preceding sequence of responses and reinforcements. For this preceding sequence I use the notation x,. Thus (It is convenient to write these sequences in this order, but note that the numbering here is from past to present, not the reverse.) Our single axiom is the following linearity assumption: I also define here certain moments which are of experimental interest. The moments ax, of the response probabilities at trial n are:

Problems of Causal Analysis 245 and it is in terms of di) that we define moments $')j,, exactly analogous to (2). We shall also be interested in the joint moments and their asymptotes $'l,...,-s if they exist. To work with these latter moments in terms of Axiom M we need the additional reasonable assumption that when all the n - l preceding responses and reinforcements are given, the s responses on trial n are statistically independent: Axiom I. I ~P(X,-~ ) > O, then The experimental restriction implied by Axiom I has been satisfied in the multiperson studies employing the linear model. From a causal standpoint the interesting thing about these multiperson situations is that one person's response is a prima facie cause of a later response by someone else, though we may prove, using Axiom I, that these responses are spurious. In other words, what we can prove using Axioms L, M, and I is that for an individual's response in a multiperson interactive situation only the sequence of preceding reinforcements is genuinely causal. For further discussion of this point, see Suppes [ 19701. Notice that what happens in this learning-theory example is typical. We start with a surface interaction but then go deeper to eliminate the direct interaction. In the present instance we eliminate causal interactions in terms of responses by going directly to the information obtained from the stochastic reinforcement schedule. I do not mean to suggest that we can do this in all kinds of social interaction. It would, in fact, be my own view that, especially in the case of language acquisition, it is exactly the responses of the mother that serve as the most important causal influence on language acquisition by the child - a view that is hardly news. But I have chosen the present limited example because it illustrates the general principle very well. This general principle is also

er tes.

Problems of Causal Analysis 247 social sciences. To avoid any confusion on this point, let me be clear that the issue is not one of general philosophical belief. It is reasonable to believe that a person s actions at a given time are probabilistically determined by the encoding of his past experience in his central nervous system and by the current state of the many chemical substances, such as hormones, enzymes, etc., in his body at the present instant. We do not have to accept a philosophical view of direct action,at a distance across time so that an event that occurred in childhood directly affects an action in adulthood without benefit of intervening internal states. Most reasonable people would probably deny belief in such remote action at a distance across time. The difficulty is scientific rather than philosophical, but the difficulty is so profound scientifically that it must affect our general philosophical view of what is possible in the social sciences. The problem is simply that of being able to postulate detailed internal states which we have some hope of being able to identify by actual empirical methods. The great success of physics and chemistry has depended upon the structural identity of substances, modulo at least the phenomenological properties we have as yet investigated with any thoroughness. It is a plausible thesis that we do not have in the case of persons or even other animals anything like such uniformity of structure; rather, one person s internal structure at a given moment is in no interesting way isomorphic with the internal structure of another person. By interesting way I mean of course in terms of psychological properties and not gross physical properties. If the situation is as hopeless as I am inclined to think it is, this means that the methodology of the social sciences and the development of causal theories must take quite a different direction than that which has been so successful in the physical sciences. Referring to the learning example considered earlier, there is a concise technical way of putting the point. The kind of learning model considered is, from a stochastic viewpoint, a stochastic process that is a chain of infinite order. The probability of a present response depends upon the complete past of the organism. In contrast, an internal-state theory of such matters would postulate an

Problems of Causal Analysis 249 developments, by and large, have taken place with concern for practical applications in a proper statistical setting and not with deterministic theories of causality reminiscent of classical physics of the nineteenth century. I have not really said anything, of course, about the statistical theory associated with the ideas developed in general outline. That is another story and far too complex even to sketch in the present article. I do want to emphasize, however, that it is important to bring any causal theory to maturity by showing how it relates to detailed statistical theory and practice. Department of Philosophy Stanford University REFERENCES Estes, WK. and Suppes, P. [l9591 Foundations of Linear Models, Studies in Mathematicul Learning Theory, R.R. Bush and W.K. Estes (eds.), Stanford University Press, Stanford, 1959, pp. 137-179. Lamperti, J. and Suppes, P. [l9591 Chains of Infinite Order and Their Application to Learning Theory, Pacific Journal of Mathematics 9 (1959), pp. 739-754. Lehmann, E.L. [l9661 Some Concepts of Dependence, The Annals of Mathematical Statistics 37 (1966), pp. 1137-1153. Simon, H.A. [l9541 Spurious Correlation: A Causal Interpretation, Journal ofamerìcan Statistical Association 49 (1954), pp. 467-492. Suppes, P. [l9701 A Probabilistic Theory of Causality (Acta Phàlosophica Fennica 24), North-Holland, Amsterdam, 1970. Suppes, P. and Atkinson, R.C. [l9601 Markov Learning Models for Multz3erson Interactions, Stanford university Press, Stanford, 1960.