Küchenhoff: An exact algorithm for estimating breakpoints in segmented generalized linear models

Similar documents
Toutenburg, Nittner: The Classical Linear Regression Model with one Incomplete Binary Variable

Augustin: Some Basic Results on the Extension of Quasi-Likelihood Based Measurement Error Correction to Multivariate and Flexible Structural Models

Pruscha: Semiparametric Estimation in Regression Models for Point Processes based on One Realization

Toutenburg, Fieger: Using diagnostic measures to detect non-mcar processes in linear regression models with missing covariates

Pruscha, Göttlein: Regression Analysis for Forest Inventory Data with Time and Space Dependencies

Pruscha: Semiparametric Point Process and Time Series Models for Series of Events

Fahrmeir, Pritscher: Regression analysis of forest damage by marginal models for correlated ordinal responses

Linear Regression and Its Applications

Fahrmeir: Discrete failure time models

Analysing geoadditive regression data: a mixed model approach

On estimation of the Poisson parameter in zero-modied Poisson models

p(z)

Fahrmeir: Recent Advances in Semiparametric Bayesian Function Estimation

Regression and Statistical Inference

Support Vector Machines

Novel Approach to Analysis of Nonlinear Recursions. 1 Department of Physics, Bar-Ilan University, Ramat-Gan, ISRAEL

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses

CS 195-5: Machine Learning Problem Set 1

Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"

Prediction of ordinal outcomes when the association between predictors and outcome diers between outcome levels

STA 216: GENERALIZED LINEAR MODELS. Lecture 1. Review and Introduction. Much of statistics is based on the assumption that random

INRIA Rh^one-Alpes. Abstract. Friedman (1989) has proposed a regularization technique (RDA) of discriminant analysis

Mathematics Research Report No. MRR 003{96, HIGH RESOLUTION POTENTIAL FLOW METHODS IN OIL EXPLORATION Stephen Roberts 1 and Stephan Matthai 2 3rd Febr

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA

1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa

Projected partial likelihood and its application to longitudinal data SUSAN MURPHY AND BING LI Department of Statistics, Pennsylvania State University

Introduction to the Generalized Linear Model: Logistic regression and Poisson regression

The iterative convex minorant algorithm for nonparametric estimation

Linear Methods for Prediction

Tutz: Modelling of repeated ordered measurements by isotonic sequential regression

Estimating the Marginal Odds Ratio in Observational Studies

Models, Testing, and Correction of Heteroskedasticity. James L. Powell Department of Economics University of California, Berkeley

Outline. Supervised Learning. Hong Chang. Institute of Computing Technology, Chinese Academy of Sciences. Machine Learning Methods (Fall 2012)

Machine Learning Linear Classification. Prof. Matteo Matteucci

Discrete Dependent Variable Models

Likelihood ratio testing for zero variance components in linear mixed models

Payer, Küchenhoff: Modelling extreme wind speeds in the context of risk analysis for high speed trains

Linear Methods for Prediction

The OSCAR for Generalized Linear Models

Statistical Methods for SVM

Managing Uncertainty

Ability Bias, Errors in Variables and Sibling Methods. James J. Heckman University of Chicago Econ 312 This draft, May 26, 2006

Gaussian Process Regression: Active Data Selection and Test Point. Rejection. Sambu Seo Marko Wallat Thore Graepel Klaus Obermayer

STA216: Generalized Linear Models. Lecture 1. Review and Introduction

Combining estimates in regression and. classication. Michael LeBlanc. and. Robert Tibshirani. and. Department of Statistics. cuniversity of Toronto

OPTIMAL DESIGNS FOR COMPLETE INTERACTION. By Rainer Schwabe. Freie Universitat Berlin

International Journal of Pure and Applied Mathematics Volume 19 No , A NOTE ON BETWEEN-GROUP PCA

An explicit dynamic model of segregation

Least Squares Estimation-Finite-Sample Properties

The best expert versus the smartest algorithm

ECONOMETRICS II (ECO 2401S) University of Toronto. Department of Economics. Spring 2013 Instructor: Victor Aguirregabiria

BIAS OF MAXIMUM-LIKELIHOOD ESTIMATES IN LOGISTIC AND COX REGRESSION MODELS: A COMPARATIVE SIMULATION STUDY

Average Reward Parameters

(1.) For any subset P S we denote by L(P ) the abelian group of integral relations between elements of P, i.e. L(P ) := ker Z P! span Z P S S : For ea

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

The assumptions are needed to give us... valid standard errors valid confidence intervals valid hypothesis tests and p-values

Bridging the two cultures: Latent variable statistical modeling with boosted regression trees

A general mixed model approach for spatio-temporal regression data

Part III. Hypothesis Testing. III.1. Log-rank Test for Right-censored Failure Time Data

Modelling geoadditive survival data

Multilevel Analysis, with Extensions

Rough Sets, Rough Relations and Rough Functions. Zdzislaw Pawlak. Warsaw University of Technology. ul. Nowowiejska 15/19, Warsaw, Poland.

2 Tikhonov Regularization and ERM


Calculation of maximum entropy densities with application to income distribution

Projektpartner. Sonderforschungsbereich 386, Paper 163 (1999) Online unter:

Gerhard Rein. Mathematisches Institut der Universitat Munchen, Alan D. Rendall. Max-Planck-Institut fur Astrophysik, and

CSC 411: Lecture 09: Naive Bayes

1 Introduction A common problem in categorical data analysis is to determine the effect of explanatory variables V on a binary outcome D of interest.

STAT5044: Regression and Anova

Part 8: GLMs and Hierarchical LMs and GLMs

RLRsim: Testing for Random Effects or Nonparametric Regression Functions in Additive Mixed Models

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

Gibbs Sampling in Linear Models #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Gaussian process for nonstationary time series prediction

Iterative procedure for multidimesional Euler equations Abstracts A numerical iterative scheme is suggested to solve the Euler equations in two and th

Beyond Mean Regression

University of California, Berkeley

Generalized Linear Models (GLZ)

type of the at; and an investigation of duration of unemployment in West Germany from 1980 to 1995, with spatio-temporal data from the German Federal

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Weighted tests of homogeneity for testing the number of components in a mixture

G. Larry Bretthorst. Washington University, Department of Chemistry. and. C. Ray Smith

290 J.M. Carnicer, J.M. Pe~na basis (u 1 ; : : : ; u n ) consisting of minimally supported elements, yet also has a basis (v 1 ; : : : ; v n ) which f

Testing Restrictions and Comparing Models

Linear Models for Classification

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing

Statistical Methods for Data Mining

Sample size determination for logistic regression: A simulation study

Estimators for the binomial distribution that dominate the MLE in terms of Kullback Leibler risk

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics

Asymptotic theory for bent-cable regression the basic case

Machine Learning Lecture Notes

BANA 7046 Data Mining I Lecture 4. Logistic Regression and Classications 1

Widths. Center Fluctuations. Centers. Centers. Widths

The Bias-Variance dilemma of the Monte Carlo. method. Technion - Israel Institute of Technology, Technion City, Haifa 32000, Israel

ON THE DIAMETER OF THE ATTRACTOR OF AN IFS Serge Dubuc Raouf Hamzaoui Abstract We investigate methods for the evaluation of the diameter of the attrac

A Stable Finite Dierence Ansatz for Higher Order Dierentiation of Non-Exact. Data. Bob Anderssen and Frank de Hoog,

Transcription:

Küchenhoff: An exact algorithm for estimating breakpoints in segmented generalized linear models Sonderforschungsbereich 386, Paper 27 (996) Online unter: http://epububuni-muenchende/ Projektpartner

An exact algorithm for estimating breakpoints in segmented generalized linear models Helmut Kuchenho University of Munich, Institute of Statistics, Akademiestrasse, D-8799 Munchen Summary We consider the problem of estimating the unknown breakpoints in segmented generalized linear models Exact algorithms for calculating maximum likelihood estimators are derived for dierent types of models After discussing the case of a GLM with a single covariate having one breakpoint a new algorithm is presented when further covariates are included in the model The essential idea of this approach is then used for the case of more than one breakpoint As further extension an algorithm for the situation of two regressors each having a breakpoint is proposed These techniques are applied for analysing the data of the Munich rental table It can be seen that these algorithms are easy to handle without too much computational eort The algorithms are available as GAUSS{programs Keywords: Breakpoint, generalized linear model, segmented regression Introduction In many practical regression{type problems we cannot t one uniform regression function to the data, since the functional relationship between the response Y and the regressor X changes at certain points of the domain of X These points are usually called breakpoints or changepoints One important example is the threshold model used in epidemiology (see Ulm, 99, Kuchenho and Carroll, 996), where the covariate X, typically an exposure, has no inuence on Y, eg the occurrence of a certain disease, up to a certain level Thus the relationship between X and Y is described by a constant up to this level and for values of X greater than this level it is given by an increasing function In such situations, we apply segmented or multiphase regression models which are obtained by a piecewise denition of the regression function E(Y jx = x) on intervals of the domain of X An overview concerning this topic can be found in Chapter 9 of Seber and Wild (989) Assuming a generalized linear model and a known number of segments we have

2 E(Y jx = x) = 8 >< >: G( + x) if x G( 2 + 2 x) if <x 2 G( K + K x) if K; <x () and y# ; b(#) f(yj# ) = exp + c(y ) : (2) Here is the nuisance{parameter and b (#) =E(Y jx = x), see Fahrmeir and Tutz (994), Seber and Wild (989), G is the link{function, eg logistic, identity etc,and f denotes the density function of Y given X = x For the threshold model mentioned above, for instance, there are two segments, where G is the logistic link and = The endpoints i of the intervals denote the breakpoints Since they are typically unknown, they have to be estimated For theoretical, but also practical reasons the breakpoints are assumed to ly between the smallest and the largest sample value x i ::: n We further assume that the regression function is continuous, ie i + i i = i+ + i+ i i K ; : Thus the model can be stated in another parameterisation: E(Y jx = x) =G( + x + where t + = KX i=2 t if t if t<: i (x ; i; ) + ) (3) From this representation it can be seen, that (3) is a usual generalized linear model, if the breakpoints i are known Therefore the ML{estimation can be performed by a grid{search{type algorithm in case of two segments, see Stasinopoulos and Rigby (992) For the linear model an exact algorithm for the least squares estimator was given by Hudson (966) (see also Schulze, 987 or Hawkins, 976) In Section 2 it is shown, that this algorithm also works for the GLM with one breakpoint In Section 3 the algorithm is extended to models with further covariates In Section 4 the ideas of Section 3 are used to derive the algorithm for fairly general models with more than one breakpoint and more than one covariate with breakpoints Giving an algorithm for such general models we ll a gap existing so far in the literature In Section 5 an example is considered We investigate the relationship between the net rent of ats in

3 Munich and the at size as well as the age of the ats based on data of the Munich rental table Finally, problems concerning the computing time are discussed and some interesting additional aspects are pointed out 2 Exact ML{estimation for models with one breakpoint We consider a GLM with one breakpoint and density (2) The regression function can then be written as E(Y jx = x) =G( + (x ; ) ; + 2 (x ; ) + ) with t ; = ;(;t) + : (4) Here is the slope parameter of the rst segment and 2 is the slope in segment 2 The log{likelihood function of one observation, conditioned on X, isgiven by y# ; b(#) G(y + (x{ ) ; + 2 (x{ ) + )= + c(y ) where the nuisance parameter is assumed to be constant over the segments If G is the natural link function, then # = + (x{ ) ; + 2 (x{ ) + : Having iid observations (x i y i ) ::: n, the log{likelihood function to be maximized in ( 2 ) is G(y i + (x i { ) ; + 2 (x i { ) + ): (5) Since this function is not dierentiable in at x i,we rst calculate the prole likelihood That is, we maximize (5) with respect to all other parameters and get P() = max 2 G(y i + (x i { ) ; + 2 (x i { ) + ): (6) Obviously model (4) is a GLM for xed Thus, the calculation of the prole likelihood corresponds to the ML{estimation of a GLM Since (6) is continuous in, it can be maximized by a grid search, see Ulm (99) For a GLIM{macro see Stasinopoulos and Rigby (992) Though these algorithms give reliable results, if the grid is appropriately chosen, it would be desirable to have an exact algorithm at one's disposal

4 We derive such an exact algorithm for maximizing (5) following the ideas of Hudson (966) Since we have assumed, that the nuisance parameter is constant for the two segments, it can be neglected in maximizing (5) Let the observations (x i y i ) ::: n be ordered with respect to x i such thatx i x j for i<j The log{likelihood is dierentiable with respect to =( 2 ) everywhere except for those values of with = x i for one if ::: ng: Therefore the algorithm has to be divided into roughly two steps according to the dierentiability of the log{likelihood In the rst step the points of dierentiability, ie the case ^(x k x k+ ) for some k, are considered Denoting the partial derivative ofg( ) with respect to the second argument by G 2 ( ) we therefore get G(y i glp i ) = [G 2 (y i glp i )] where glp is the \broken linear predictor" B glp i = + (x i { ) ; + 2 (x i { ) + (x i { ) ; (x i { ) + { I fxig{ 2 I fxi>g C A (7) and I denotes the indicator function Let (^ ^ ^ 2 ^) be a zero of (7) with ^(x k x k+ )and ^ 6= ^ 2 The system of equations obtained by equating (7) to results after some algebra in kx kx i=k+ i=k+ G 2 (y i ^ + ^ x i ) = (8) G 2 (y i ^ + ^ x i )x i = (9) G 2 (y i ^ 2 + ^2 x i ) = () G 2 (y i ^ 2 + ^ 2 x i )x i = () with ^ j = ^ ; ^ j ^ j = 2: From equations (8) and (9), () and () we conclude that (^ ^ ) and (^ 2 ^ 2 ), respectively, are ML{solutions of the regressions in the two segments Since is uniquely determined by the continuity condition, possible zeros with ^(x k x k+ ) can be determined by estimating the parameters separately in the two segments based on (x i y i ) ::: k and (x i y i ) i=k+ ::: n,

5 which yields (^ ^ )and(^ 2 ^ 2 ), respectively The estimator for is then obtained from ^ = ^ 2 ; ^ ^ ; ^ 2 : If ^(x k x k+ ), it is a zero of (7) In this case, the estimator ^ is given by ^ =^ + ^ ^: If ^ 62 (x k x k+ ), we deduce from (8) { () that there is no local maximum with (x k x k+ ) The above mentioned procedure is performed for the nite number of intervals (x k x k+ ), where these intervals have tobechosen such that the ML{estimators exist in the corresponding segments In the second step we calculate the prole likelihood P(x i ) i = ::: n, obtaining the maximum of the log{likelihood at all points of non{dierentiability Finally the global maximum of the log{likelihood is given by the maximum of this nite number of local maxima Conducting this algorithm the estimation of at most m +2m GLMs is needed, if there are m observations with dierent values of x 3 Models with covariates In many practical situations there will be further covariates in the regression model, which leads to the following extension of model (4): E(Y jx = x Z = z) =G( + (x ; ) ; + 2 (x ; ) + + z ) (2) where Z is the vector of covariates with vector of parameters As in Section 2 (x i y i z i ) ::: n denote the corresponding observations with x i x j for i<j Toderive the ML{estimator, we rst consider again the case ^(x k x k+ ) Then the derivative of the log{likelihood with respect to =( 2 ) is G(y i glp i ) = [G 2 (y i glp i )] B (x i { ) ; (x i { ) + { I fxig{ 2 I fxi>g z i C A (3) with glp i = + (x i { ) ; + 2 (x i { ) + + z i :

6 Analogously to (8) { () we get the following system of equations with (^ ^ ^2 ^ ^) denoting a zero of (3): kx G 2 (y i ^ + ^ x i + z i^)z i + with ^ j =^ ; ^j ^ j = 2 kx kx i=k+ i=k+ i=k+ G 2 (y i ^ + ^ x i + z i^) = (4) G 2 (y i ^ + ^ x i + z i^)x i = (5) G 2 (y i ^ 2 + ^ 2 x i + zi^) = (6) G 2 (y i ^ 2 + ^ 2 x i + z i^)x i = (7) G 2 (y i ^ 2 + ^2 x i + z i^)z i = (8) Equations (4) { (8) correspond to a generalized linear model with an analysis of covariance{type design matrix D k = B x x k x k+ x n z z k z k+ Therefore we obtain zeros of (3) by tting a generalized linear model with design matrix D k,which again yields because of the continuity condition z n C A : ^ = ^ 2 ; ^ 2 ^ ; ^ 2 : If ^(x k x k+ ), we have found a local maximum, otherwise there is no maximum with ^(x k x k+ ) The remaining part of the algorithm is now completely analogous to that presented in Section 2

4 Further extensions 4 Models with more than two segments 7 Let us now consider the case of K>2 segments, ie Model (3) In practical problems the number of segments will be typically not greater than three Williams (97), for instance, restricts his investigations to this special case But even in the situation of a linear regression with a normally distributed Y no complete algorithm for calculating the exact ML-estimator can be found in the literature Williams (97) states explicitly that his algorithm for three segments shows certain gaps We describe the complete algorithm for K = 3, where it will be formulated such that it can be directly extended to the case of K>3 We start with the generalized linear model with two breakpoints in the following parametrization: E(Y jx = x) =G( + x + 2 (x ; ) + + 3 (x ; 2 ) + ) (9) with < 2 and i 6= for i =2 3 Let (x i y i ) ::: n again denote the ordered observations with x i x j for i<j Then, we get the derivative of the log-likelihood with respect to =( 2 3 2 ) as G 2 (y i glp i )( x i (x i ; ) + (x i ; 2 ) + ; 2 I fxi>g ; 3 I fxi>2g) (2) with glp i = + x i + 2 (x i ; ) + + 3 (x i ; 2 ) + : Since the log-likelihood is not dierentiable at points with = x i or 2 = x i for some x i, the domain of ( 2 ) is divided into rectangles R kl with R kl =[x k x k+ ] [x l x l+ ] k+<l: In the interior of R kl the log-likelihood is dierentiable Equating (2) to we get after some algebra G 2 (y i + x i )( x i ) I fxig = G 2 (y i ; 2 +( + 2 )x i )( x i ) I fxi>gi fxi2g = G 2 (y i ; 2 ; 3 2 +( + 2 + 3 )x i )( x i ) I fxi>2g = : Obviously all maxima of the log-likelihood correspond to the maxima of the separate regressions in the three segments As in the preceding sections we

8 therefore derive the ML-solutions separately for the dierent segments which are denoted by ~ ~ ~ 2 ~ 2 ~ 3 ~ 3 The continuity assumption yields ~ = ~ ; ~ 2 ~ 2 ; ~ and ~ 2 = ~ 2 ; ~ 3 ~ 3 ; ~ 2 : If ~ (x k x k+ )and~ 2 (x l x l+ ) then a local maximum has been found Otherwise, the log-likelihood function has its maximum value at the boundary of the rectangle, ie fx k x k+ g or 2 fx l x l+ g Let for instance = x i then the model can be rewritten introducing a new covariate z =(x ; ) + as E(Y jx = x) =G( + x + 2 z + 3 (x ; 2 ) + ): Thus, we are in the situation of a model with one breakpoint and an additional covariate as discussed in Section 3 To obtain all maxima at the boundary of R kl wecheck successively all points = x k+ and 2 = x l as well as 2 = x l+ This yields the maximizer ^ kl for the rectangle R kl Finally, the global maximum is obtained as ^ ML = arg max L(^ kl ): k l 42 Models with two regressors both having a breakpoint As a further extension we allow for two regressors with each of them having a breakpoint which is modelled as E(Y jx Z) =G( + (x{ ) ; + 2 (x{ ) + + (z{ 2 ) ; + 2 (z{ 2 ) + ): Even in this case we can essentially proceed as above Let (x i y i z i ) ::: n denote the ordered observations with x i x j for i < j The observations of the second regressor Z are ordered by a second index z ir with z ir z is r < s For deriving the ML-estimator of the parameter vector =( 2 2 2 ) we again divide the domain of ( 2 ) into rectangles R kr =[x k x k+ ] [z lr z lr+] and consider rst the case where ( 2 ) lies in the interior of R kr Using ideas of Section 2 we t a generalized linear model with parameter vector ( 2 2 2 2 ) and design matrix D = ; ( x i ) I fxig ( x i ) I fxi>g ( z i ) I fzi2g ( z i ) I fzi>2g The corresponding model equation is given by ::: n : E(Y jx = x Z = z) = G(( + x)i fxg +( 2 + 2 x)i fx>g +( + z)i fz2 g +( 2 + 2 z)i fz>2g ):

9 From the ML-estimator (^ ^ 2 ^ ^ 2 ^ ^ 2 ^ ^ 2 ) we obtain as estimators for ( 2 ) using the continuity assumption ^ = ^ ; ^ 2 ^ 2 ; ^ and ^ 2 = ^ ; ^ 2 ^ 2 ; ^ : If (^ ^ 2 ) R kr wehave found a local maximum of the log-likelihood since the score equations can be transformed similarly to equations (4) - (8) At the boundary of the rectangle the above model reduces to one with only a single regressor, one breakpoint and a covariate Thus, the approach of Section 3 can be applied This leads to the maximizer ^ kr for each rectangle R kr Finally, the global maximum results from ^ ML = arg max k r L(^ kr ): Further extensions to more than two covariates or more than two breakpoints are straightforward, but the corresponding algorithms become considerably more complicated and thus require an increasing computational eort 5 An example: the Munich rental table Rental tables are built up based on surveys in larger cities or communities in Germany They serve as a formal instrument for rating rents depending on year of construction, at size, and other covariates For a detailed description of the data material and the statistical methods used we refer to Fahrmeir, Gieger, Mathes, and Schneewei(995) and Fahrmeir, Gieger, and Klinger (995) As a rst approach we model the relationship between net rent (Y ) and at size (X), where the assumption of a breakpoint is justied because smaller ats are more expensive relative to bigger ats Thus, we consider the following model equation E(Y jx = x) = + (x{ ) ; + 2 (x{ ) + : (2) Besides the presence of a breakpoint an additional problem occurs when analysing this data caused by heteroscedasticity Following Fahrmeir et al (995) we apply a weighted regression using the weights proposed there The results are obtained from the algorithm presented in Section 2 and are given in Table, where the estimated variances are calculated by the asymptotic theory derived in Kuchenho (995) As it can be seen from these results the breakpoint takesavalue of about 44 m 2 For smaller ats the estimated slope of 54 is below theonefor bigger ats ( ^ 2 =9:5) Comparing our results with those gained from a linear regression no essential dierences can be stated regarding the t of the

Table : Parameter estimates of the weighted broken linear regression model (2) for the Munich rental table In the column ^ the estimated standard deviations are listed parameter estimate ^ ^ 44 46 ^ 6 36 ^ 54 3 ^ 2 95 43 data For a more detailed analysis and a comparison with the results from Fahrmeir, Gieger, Mathes, and Schneewei(995) see Kuchenho (995) In a second step we take into account the age of the at (Z) as additional regressor which is dened as 994 minus year of construction Possible changes in the way of building ats are reected in the model equation by allowing for a breakpoint in the second regressor, age: E(Y jx Z) = + (x{ ) ; + 2 (x{ ) + + (z{ 2 ) ; + 2 (z{ 2 ) + : (22) Here, the application of the algorithm proposed in Section 42 yields the results given in Table 2 Table 2: Parameter estimates of the weighted broken linear regression model (22) for the Munich rental table and 2 denote the breakpoints belonging to at size and age In the column ^ the estimated standard deviations are listed parameter estimate ^ ^ 37 37 ^ 2 36 97 ^ 5 29 ^ 354 22 ^ 2 88 35 ^ -29 4 ^ 2-39 28 In this case, the application of a segmented model yields a much better t than the use of a linear model without breakpoints The estimated breakpoint of 36 for the variable age corresponds to the year 963 Since 2 is not signicantly dierent from, the age of ats built before 963 is not essential for their net rent For ats built after 963 the net rent increases with the

year of construction (^ = ;2:9) As already indicated, this eect is possibly due to a substantial change in the way ofhow houses were built 6 Discussion From our experience when analysing data by applying the above algorithms as for instance in the example presented in Section 5 we can state a good practicability of these algorithms For the Munich rental table, where we have about 2 data points, the computing time for the model with one breakpoint was about 5 minutes on a sun-sparc work station For the model with two regressors each having a breakpoint it took about minutes for calculating the ML-estimators using the algorithm presented in Section 42 Thus, the computing times do not seem to constitute a real problem In addition, algorithms for linear, logistic, and weighted linear models can be obtained on request from the author The algorithms are written in Gauss and can easily be rewritten for other generalized linear models Finally,we should address two important aspects related to the proposed algorithms When deriving the above algorithm we assumed that the nuisance parameter is constant over the segments Note that this assumption fails eg in the linear regression when there are dierent variances in the segments In case of logistic regression this problem does not occur since there is no nuisance parameter ( = ) This case of the logistic regression where can be treated as known can be generalized to other models That means, alternatively to the assumption of a constant nuisance parameter the derivation of the algorithm remains valid if the nuisance parameter, itself, or the ratio of the nuisance parameters of the dierent segments is known One of the most interesting aspects of our idea for deriving an exact algorithm concerns the models with covariates The approach proposed in this paper is not only useful for such models, for which this approach was designed, but it is also an essential part of the derivation of the corresponding algorithm for the case of more than two segments Thus, we were also able to solve the problem pointed out by Williams (97) References Fahrmeir, L, Gieger, C, and Klinger, A (995) Additive, dynamic and multiplicative regression, Discussion Paper, SFB 386, Munich Fahrmeir, L, Gieger, C, Mathes, H, and Schneewei, H (995) Gutachten zur Erstellung des Mietspiegels fur Munchen 994, Teil B: Statistische Analyse der Nettomieten, Institute of Statistics, Ludwig-Maximilians- University ofmunich

2 Fahrmeir, L and Tutz, G (994) Multivariate statistical modelling based on generalized linear models, Springer-Verlag, New York Hawkins, D M (976) Point estimation of the parameters of piecewise regression models, Applied Statistics 25: 5{57 Hudson, D J (966) Fitting segmented curves whose join points have to be estimated, Journal of the American Statistical Association 6: 97{ 29 Kuchenho, H (995) Schatzmethoden in mehrphasigen Regressionsmodellen, Habilitationsschrift, Institute of Statistics, Ludwig-Maximilians- University ofmunich Kuchenho, H and Carroll, R J (996) Biases in segmented regression with errors in predictors, Statistics in Medicine 5: to appear Schulze, U (987) Mehrphasenregression: Stabilitatsprufung, Schatzung, Hypothesenprufung, Akademie-Verlag, Berlin Seber, G A F and Wild, C J (989) Nonlinear regression, John Wiley & Sons, New York Stasinopoulos, D M and Rigby, R A (992) Detecting break points in generalised linear models, Computational Statistics & Data Analysis 3: 46{47 Ulm, K (99) A statistical method for assessing a threshold in epidemiological studies, Statistics in Medicine : 34{349 Williams, D A (97) Discrimination between regression models to determine the pattern of enzyme synthesis in synchronous cell cultures, Biometrics 26: 23{32