Achilles: Now I know how powerful computers are going to become!

Similar documents
Solution to Proof Questions from September 1st

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Chapter 26: Comparing Counts (Chi Square)

Section 2: Estimation, Confidence Intervals and Testing Hypothesis

appstats27.notebook April 06, 2017

Introduction to Algebra: The First Week

MITOCW ocw f07-lec39_300k

Notes 11: OLS Theorems ECO 231W - Undergraduate Econometrics

Solving Quadratic & Higher Degree Equations

Confidence Intervals

Physics 509: Bootstrap and Robust Parameter Estimation

MITOCW watch?v=hdyabia-dny

Volume vs. Diameter. Teacher Lab Discussion. Overview. Picture, Data Table, and Graph

Uncertainty and Bias UIUC, 403 Advanced Physics Laboratory, Fall 2014

Communication Engineering Prof. Surendra Prasad Department of Electrical Engineering Indian Institute of Technology, Delhi

Section 3: Simple Linear Regression

Introduction to Error Analysis

HOLLOMAN S AP STATISTICS BVD CHAPTER 08, PAGE 1 OF 11. Figure 1 - Variation in the Response Variable

Vocabulary: Samples and Populations

The following are generally referred to as the laws or rules of exponents. x a x b = x a+b (5.1) 1 x b a (5.2) (x a ) b = x ab (5.

I started to think that maybe I could just distribute the log so that I get:

Predicting AGI: What can we say when we know so little?

Physics 6720 Introduction to Statistics April 4, 2017

Overview. Confidence Intervals Sampling and Opinion Polls Error Correcting Codes Number of Pet Unicorns in Ireland

Warm-up Using the given data Create a scatterplot Find the regression line

Probability Distributions

Lecture 4: Training a Classifier

MITOCW watch?v=poho4pztw78

Lecture 12: Quality Control I: Control of Location

Solving Quadratic & Higher Degree Equations

A CHEMISTRY PARTY. A puppet show

Chapter 27 Summary Inferences for Regression

MITOCW Investigation 3, Part 1

Lecture 27: Theory of Computation. Marvin Zhang 08/08/2016

Math 111, Introduction to the Calculus, Fall 2011 Midterm I Practice Exam 1 Solutions

Solving Quadratic & Higher Degree Equations

PHYSICS 15a, Fall 2006 SPEED OF SOUND LAB Due: Tuesday, November 14

Bootstrap Confidence Intervals for the Correlation Coefficient

#29: Logarithm review May 16, 2009

I started to think that maybe I could just distribute the log so that I get:

Line Broadening. φ(ν) = Γ/4π 2 (ν ν 0 ) 2 + (Γ/4π) 2, (3) where now Γ = γ +2ν col includes contributions from both natural broadening and collisions.

Q520: Answers to the Homework on Hopfield Networks. 1. For each of the following, answer true or false with an explanation:

MITOCW MITRES18_005S10_DiffEqnsGrowth_300k_512kb-mp4

Chapter 1 Review of Equations and Inequalities

How Measurement Error Affects the Four Ways We Use Data

MITOCW ocw f07-lec37_300k

A story about anxiety and mindfulness. For a workbook on Mindfulness to go with this story. Please click this link BEE-LINDA S MINDFULNESS WORKBOOK

*Karle Laska s Sections: There is no class tomorrow and Friday! Have a good weekend! Scores will be posted in Compass early Friday morning

Metric Prefixes UNITS & MEASUREMENT 10/6/2015 WHY DO UNITS AND MEASUREMENT MATTER?

Chapter 8: Estimating with Confidence

SECONDARY MATH THREE. An Integrated Approach. MODULE 2 Logarithmic Functions

Biostatistics and Design of Experiments Prof. Mukesh Doble Department of Biotechnology Indian Institute of Technology, Madras

MITOCW ocw f99-lec01_300k

Design and Optimization of Energy Systems Prof. C. Balaji Department of Mechanical Engineering Indian Institute of Technology, Madras

Quadratic Equations Part I

Let V be a vector space, and let X be a subset. We say X is a Basis if it is both linearly independent and a generating set.

Section 5.4. Ken Ueda

15 Skepticism of quantum computing

Fitting a Straight Line to Data

MITOCW ocw f99-lec09_300k

MITOCW MITRES18_005S10_DiffEqnsMotion_300k_512kb-mp4

Lecture 5. 1 Review (Pairwise Independence and Derandomization)

Squaring and Unsquaring

Statistical inference

Lecture 14, Thurs March 2: Nonlocal Games

TheFourierTransformAndItsApplications-Lecture28

Linear classifiers: Overfitting and regularization

VBM683 Machine Learning

EQ: How do I convert between standard form and scientific notation?

EXPERIMENTAL UNCERTAINTY

Ask. Don t Tell. Annotated Examples

When they compared their results, they had an interesting discussion:

PS2: Two Variable Statistics

Lesson 21 Not So Dramatic Quadratics

MITOCW ocw-18_02-f07-lec02_220k

Getting Started with Communications Engineering

Good morning everyone, and welcome again to MSLs World Metrology Day celebrations.

MITOCW ocw-18_02-f07-lec17_220k

Instructor (Brad Osgood)

Zeno s Paradox #1. The Achilles

Why is the field of statistics still an active one?

Lesson 11-1: Parabolas

Study and research skills 2009 Duncan Golicher. and Adrian Newton. Last draft 11/24/2008

f rot (Hz) L x (max)(erg s 1 )

Solve Systems of Equations Algebraically

Section 0.6: Factoring from Precalculus Prerequisites a.k.a. Chapter 0 by Carl Stitz, PhD, and Jeff Zeager, PhD, is available under a Creative

( )( b + c) = ab + ac, but it can also be ( )( a) = ba + ca. Let s use the distributive property on a couple of

MITOCW watch?v=dztkqqy2jn4

Expectation is linear. So far we saw that E(X + Y ) = E(X) + E(Y ). Let α R. Then,

PS2 Lab 1: Measurement and Uncertainty Supplemental Fall 2013

A particularly nasty aspect of this is that it is often difficult or impossible to tell if a model fails to satisfy these steps.

The State of the Universe [2010] There is only data and the interpretation of data (green text = assumptions)

2.4 Log-Arithm-etic. A Practice Understanding Task

Math101, Sections 2 and 3, Spring 2008 Review Sheet for Exam #2:

CHAPTER 1. Introduction

CSC236 Week 3. Larry Zhang

Experimental Design and Graphical Analysis of Data

2 Analogies between addition and multiplication

Big Data Analysis with Apache Spark UC#BERKELEY

- a value calculated or derived from the data.

Transcription:

A Sigmoid Dialogue By Anders Sandberg Achilles: Now I know how powerful computers are going to become! Tortoise: How? Achilles: I did curve fitting to Moore s law. I know you are going to object that technological progress cannot be exponential forever, so be assured I did a conservative analysis by fitting a logistic curve. Tortoise: So you assume that over time, like many other technologies, computer performance will first accelerate until some limitations make it slow down and eventually approach some ultimate limit? Achilles: Exactly. And that limit is my estimate of how powerful computers will ever become. I used a least squares fit using Matlab s fmincon function to get the most likely sigmoid. The data is from the performance curve database http://pcdb.santafe.edu/. Here is my graph: And as you can see, we will get 10 7.5 flops per dollar in the long run. That is more than 300,000 times our current best performance! Tortoise: Hmm. Achilles: You are not convinced, my friend?

Tortoise: No. How can you be certain of this result, since the data points appear to be a bit random? Achilles: There is always noise in data; we have to do the best we can with the numbers we got. In the future we can always update this forecast. Tortoise: As more data arrives your estimate will change, but will it change a little or a lot? Achilles: OK, that is a problem. But we can do a bootstrapping confidence interval. I resample (with replacement) data points, fit the curve to this proxy dataset, repeat this a lot of times and get a probability distribution of the parameters. Including the ultimate limit. Then I just find the interval containing 90% of the results, and I will have a good estimate of the confidence interval. Here it is: And here is a histogram of the parameter:

: The 95% confidence interval for the ultimate level is from 10 2.96 to 10 13.99. Happy? Tortoise: Hmm. 11 orders of magnitude of uncertainty in eventual computer power. Achilles: Well, no forecast is perfect. Still, this gives us some useful information about the ultimate limits. Tortoise: Maybe. If I were to make a forecast, I would be most concerned with what we could see in the near future. That would allow me to check whether I was likely right or wrong. Achilles: Growth sigmoids have three parameters: how far up they will go, how sharply they grow when they grow, and when the growth is fastest (i.e. its inflection point). You are suspicious about the first, the second might be more to your liking and I think the third is exactly what you look for. Tortoise: Yes, knowing how far we have to go has a great deal of practical importance. And when previously accelerating growth starts to slow we can actually notice it a short while afterwards. Could you tell me when Moore s law will have its inflection point? Achilles: Sure. It will occur in 1986. Well, with 95% confidence between 1975 and 2002.

Tortoise: Interesting prediction. Achilles: Well, I don t mind being called conservative. Tortoise: So you trust these estimates? Achilles: Sure. Although it was a bit worrying to see how much uncertainty is hidden in the estimates. But I guess this is natural, it is still early days even if we have just passed the inflection point. Tortoise: I wonder how reliable predictions like this are before the inflection point. In fact, can you reliably predict when inflection points occur before they happen? Achilles: I can create some artificial data and test it. Here is a sigmoid with inflection point at time 5, asymptote 1 and sharpness 1, where there are 100 data points with N(0,0.1) noise added. The red curve is the fit, the blue curve is the real curve.

OK, let s try this kind of fit when the 100 data points lie in the interval 0 to T for increasing T. Also, I will use 1000 bootstrap samples to estimate the asymptote and when the inflection point will happen. Here are the estimates with confidence intervals:

Oh dear Tortoise: So we should not trust the estimates at all until well after the inflection point. Achilles: Including our estimates of when it will happen or has happened. Tortoise: This makes me doubt the claim we have passed the inflection point. Either we have, and we have an accurate estimate, or we have not, and we just got a random estimate. Achilles: Yes. But these curves are a bit unnecessarily nasty, since each plotted point corresponds to a different trial with different random noise in the given data. Sometimes the data forms a really misleading pattern, but not always. Tortoise: But we cannot tell the difference when we look at our data. Also, I suspect that for sharper transitions the predictions are even worse before they happen. By the way, I wonder whether there is a bias towards overestimating the eventual level and when the inflection point happens in the middle of the curve? Achilles: Let me run a lot of independent trials and see. Each dot is a trial, the red line is the median estimate they make.

OK, the median is not bad at estimating how the curve looks. At least it is not biased.

Tortoise: But in real life we cannot get it, since we will only have one of those blue points. And at least the eventual level has a very skew distribution. Achilles: It is interesting to see that the uncertainty at time 1, 3, 4, and 5 is a bit larger than the one at time 2. Tortoise: I guess that is because it is really hard to predict any shape when the stretch of data is shorter than the spread due to noise, and tricky to do it while the change is maximal. But at time 2 there is at least a detectable straight line. All in all, it shows just how unpredictable this kind of forecasting is. Achilles: So you think computers are not going to develop as far as I do? Tortoise: Not at all. I am very optimistic about how far things can go. I just think we shouldn t trust this kind of extrapolation too much. Even if it is true technology progresses like a sigmoid, we cannot reliably use it to predict how far we are going until it is obvious anyway. And who knows, maybe the true curve is really an infinite exponential or an asymmetric sigmoid-like curve (like the solution to Y (t)= Y^2(t) Y^3(t) ). But these models have the same problem. The more complex the curve, the more uncertain the predictions will be. Achilles: So we should use straight lines instead? Tortoise: They don t make much sense for the problem given what we actually know about technology growth. Prior information is always important. You used constraints on the sigmoid curve parameters to keep them positive, and this no doubt improved reliability. We should do what we can to make plausible and robust predictions but we shouldn t think they are going to be correct even if we have the right model. Achilles: Want to bet that that people will prefer objective-sounding numbers anyway? Tortoise: That sounds like a robust and plausible prediction.