12 Symbolic Regression

Size: px
Start display at page:

Download "12 Symbolic Regression"

Transcription

1 Computational Geosciences PART I METHODS 12 Symbolic Regression 12-1 Introduction Symbolic regression is the process of determining the symbolic function which describes a data set effectively developing an analytic model which summarizes the data and is useful for predicting response behaviors as well as facilitating human insight and understanding. The symbolic regression approach adopted herein is based upon genetic programming wherein a population of functions are allowed to breed and mutate with the genetic propagation into subsequent generations based upon a survival-of-the-fittest criteria. Amazingly, this works and, although computationally intensive, summary solutions may be reasonably discovered using current laptop and desktop computers. It should be noted that symbolic regression is not a silver bullet and is, in fact, complementary to its nonlinear analysis brethren of neural networks and statistical learning theory (support vector machines) as well as classical linear analysis and data mining tools and techniques. In all cases, an understanding of the problem, the data variables, the data quality, and definitions of success is a critical aspect since context free analysis can easily lead to confidently wrong answers i.e., is very dangerous. The symbolic regression capability provides a complete and flexible foundation for evolving multivariate datadriven models. The decision to develop this symbolic regression capability within the Mathematica environment was validated during the development of this package as the ability to seamlessly blend symbolic and numeric computing has enabled new features which would have been onerous to implement in a strictly procedural and numeric environment. Additional benefits include facilitating analysis documentation Symbolic Regression (SR) Method Unlike traditional linear and nonlinear regression methods that fit parameters to an expression (equation/relation) of a given form, symbolic regression (SR) simultaneously searches for both the parameters as well as the form of expression. Although the discipline of SR has matured significantly in the last few years (e.g., Davidson et al. 2003), its applications to geodesy are very rare, exemplified only in the works of Wu et al. (2007, 2008) who employed it for transforming GPS coordinates into two-dimensional coordinates. In a more recent work, Wu and Su (2013) developed a lattice-based clustering method and integrated it with a genetic programming to build a better regression model for coordinate transformation. In natural and technical sciences, and in finance, however, the method has been applied for quite a long time efficiently, (e.g., Santini and Tettamanzi (2001); Banks (2002); Kwon and Moon (2005); Babu and Karthik (2007); Schmidt and Lipson (2009); Langdon and Gustafson (2010); and Garg and Tai (2011)). In hydrological sciences, for example, Parasuraman et al. (2007) applied it to model the dynamics of evapotranspiration, where their results performed better than the traditional Penman-Montieith method, and were comparable to those of artificial neural network (ANN). The proposed method could be of use to geodesy, where regression analysis and functional approximation are often handled. For instance, SR could be used for gravimetric corrections where they have traditionally been carried out using a wide variety of parametric and non-parametric surfaces such as polynomial models (e.g., Fotopoulos and Sideris 2005), spline interpolation

2 2 SymbolicRegression_12.nb Fotopoulos and Sideris 2005), spline interpolation (e.g., Featherstone 2000), least squares collocation (LSC) (e.g., Iliffe et al. 2003), Kriging (e.g., Nahavandchi and Soltanpour 2004), combined least squares adjustments (e.g., Fotopoulos 2005) and thin plate spline (TPS) surface of solving the problem via finite elements method (e.g., Zaletnyik et al. 2007). Applying soft computing technique, Kavzoglu and Saka (2005) and Lao-Sheng Lin (2007) employed artificial neural network (ANN) for approximating the GPS/leveling geoidal heights instead of the corrector surface itself. Zaletnyik et al. (2007) also used ANN but with radial bases activation functions (RBF) and regularization in the training phase. Soltanpour et al. (2006) used second generation wavelets to approximate corrector surface directly. Another soft computing technique represented by the support vector machines (SVM) was employed by Zaletnyik et al. (2008). Some of these models are global types of regression, (e.g., linear parametric models and ANN), while some are interpolating type of local models, (e.g., thin plate spline and SVM). Generally speaking, local methods are more precise than global on the training set, but their complexity is very high since they involve all of the measured data points in the model structure. In addition, they are less effective on the validation set since the overlearning affect the training set. The foregoing discussions support the need for the proposed SR method, particularly in geodesy where it is of utmost need but rarely used. SR can be considered as a broad generalization of the class of Generalized Linear Models (GLM). GLM is a linear combination of n basis functions β i, i = 1,2,...n with a dependent variable y and an independent vector variable x. n y (x) = c 0 + c i β i (x) + ϵ i=1 where ci are the coefficients and ε the error term. SR will search for a set of basic functions (building blocks) and coefficients (weights) in order to minimize the error ε in case of given y and x. The standard basic functions are constant, addition, subtraction, multiplication, division, sine, cosine tangent, exponential, power, square root, etc. To select the optimal set of basic functions, Koza (1992) suggested employment of genetic programming (GP). GP is a biologically inspired machine learning method that evolves computer programs to perform a task. In order to carry out genetic programming, the individuals (competing functions) should be represented by a binary tree. In standard GP, the leaves of the binary tree are called terminal nodes represented by variables and constants, while the other nodes, the so called non-terminal nodes are represented by functions. Let us see a simple example. Consider Its binary tree representation can be seen in Fig 12.1, β i (x) = x 1 x x 3 Fig The binary tree representation of a basic function βi In this example, there are three variables (x1, x2, x3), two constants (1, 2), and three elementary functions (plus, times, rational). The binary tree of y(x) can be built up from such trees as sub-trees. Mathematically

3 SymbolicRegression_12.nb 3 y(x) = c 0 + c 1 tree 1 + c 2 tree GP randomly generates a population of individuals ( y k (x), k = 1, 2...n ) represented by tree structures to find the best performing trees. There are two important features of the function represented by a binary tree: complexity and fitness. We define complexity as the number of nodes in a binary tree needed to represent the function, Morales (2004). The fitness qualifies how good a model (y = y(x)) is. Basically, there are two types of measures used in SR; the root mean squared error (RMSE) and the Rsquare. The later returns the square of the Pearson product moment correlation coefficient (R) describing the correlation between the predicted values and the target values, (e.g., Ferreira (2006)). Then, the goodness of the model, the fitness function can be defined as, f = 1 1+ RMSE or f = R2 where 0 f 1 GP tries to minimize this error to improve the fitness of the population consisting of individuals (competing functions) from generation to generation by mutation and cross-over procedure. Mutation is an eligible random change in the structure of the binary tree, which is applied to a randomly chosen sub-tree in the individual. This sub-tree is removed from the individual and replaced by a new randomly created sub-tree. This operation leads to a slightly (or even substantially) different basic function. Let us consider the binary tree on Fig. 12.2a, where the sub-tree of y 2 is replaced by y+x 2. Then, the mutated binary tree can be seen in Fig. 12.2b Fig Binary tree representations of the mutation. y 2 in a) is replaced by y+x 2 in b) The operation cross-over representing sexuality can accelerate the improvement of the fitness of a function more effectively than mutation alone can do. It is a random combination of two different basic functions (parents), based on their fitness, in order to create a new generation of functions, more fitter than the original functions. To carry out cross-over, crossing points (nonterminal nodes) in tree of both parents should be randomly selected, as it can be seen in Fig. 3. Then subtrees belonging to these nodes will be exchanged creating offsprings. Let us consider the parents before the cross-over. The first parent (x-y)/3, with its crossing point (x) is shown in Fig. 12.3a, the second parent 3x+y 2 /5 with its crossing point (see the sub-tree of y 2 ) is presented in Fig b.

4 4 SymbolicRegression_12.nb Fig The parents before the cross over The children produced by the cross-over are given on Fig The first child (y 2 -y)/3 is shown in Fig a while the second child (16/5)x is shown in Fig. 12.4b. Fig The children produced by cross-over The generalization of GP was invented by Cramer (1985) and further developed by Koza (1992). GP is a class of evolutionary algorithms working on executable tree structures (parse trees). Koza (1992) showed that GP is capable of doing symbolic regression (or function identification) by generating mathematical expressions approximating a given sample set very closely or in some cases even perfectly. Therefore, GP finds the entire approximation model and its (numerical) parameters simultaneously. An important goal in symbolic regression is to get a solution, which is numerically robust and does not require high levels of complexity to give accurate output values for given input parameters. Small mean errors may lead to wrong assumptions about the real quality of the found expressions. To be on the safer side, a worst case absolute error should be determined. Sometimes, alternating the worst case absolute error and RMSE or R 2 as target to be minimized during GP process is the best strategy. Complexity and fitness are conflicting features leading to a multiobjective problem, (e.g., Smits and Kotanchek (2004)). Useful expression is both predictive and parsimonious. Some expressions may be more accurate but overfit the data, whereas others may be more parsimonious but oversimplify. The prediction error versus complexity or 1-fitness versus complexity Pareto front represents the optimal solutions as they vary over expression complexity and maximum prediction error, (e.g., Paláncz and Awange (2012)). As Fig shows, functions representing the Pareto front have the following features: In case of fixed complexity, there is no such a solution (function), which could provide less error as the Pareto solution, Reversely, in case of fixed error, there is no such a solution (function), which would

5 SymbolicRegression_12.nb 5 Reversely, in case of fixed error, there is no such a solution (function), which would have smaller complexity than the Pareto solution. Fig The Pareto - front In order to illustrate the method, let us start with a simple, didactic example: Kepler s Third Law. This problem was considered and rediscovered by Langley et al. (1987) by a family of heuristic techniques as well as by Koza (1992) by genetic programming Didactic example- Kepler third law Kepler s problem The third law of Kepler states: "The square of the orbital period of a planet is directly proportional to the cube of the semi-major axis of its orbit (average distance from the Sun)." P 2 a 3 where P is the orbital period of the planet and a is the semi-major axis of the orbit. For example, suppose planet A is 4 times as far from the Sun as planet B. Then planet A must traverse 4 times the distance of planet B each orbit, and moreover it turns out that planet A travels at half the speed of planet B, in order to maintain equilibrium with the reduced gravitational centripetal force due to being 4 times further from the Sun. In total it takes 4 2 = 8 times as long for planet A to travel an orbit, in agreement with the law (8 2 = 4 3 ). The third law currently receives additional attention as it can be used to estimate the distance from an exoplanet to its central star, and help to decide if this distance is inside the habitable zone of that star. The exact relation, which is the same for both elliptical and circular orbits, is given by the equation This third law used to be known as the harmonic law, because Kepler enunciated it in a laborious attempt to determine what he viewed as the "music of the spheres" according to precise laws, and express it in terms of musical notation. His result based on the Rudolphine table containing the observations of Tycho Brache 1605, see Table 12.1 Table 12.1 Normalized observation planetary data Planet where a is give in units of Earth s semi-major axis. Period Semimajor axis P [yr] a Mercury Venus Earth Mars Jupiter Saturn Uranus Neptune Pluto

6 6 SymbolicRegression_12.nb Let us assume that Kepler could have employed one of the function approximation techniques like polynomial regression, artificial neural networks, support vector machine, thin plate spline. Could he find this simple relation with these sophisticated methods? Polynomial regression Let us employ a fifth order algebraic polynomial. Then the resulted polynomial is P = a a a a a Neural network Employing a single-layer feedforward network (Kecman 2001) with two nodes where sigmoid activation function is implemented, we obtained P = a a a Support vector machine Employing wavelet - kernel (Palancz et al. 2005) the result is the following, P = (0.39-a)2 Cos[ ( a)] (0.72-a)2 Cos[ ( a)] (1.-a)2 Cos[ (1. - a)] (1.52-a)2 Cos[ ( a)] (5.2-a)2 Cos[ (5.2 - a)] (9.54-a)2 Cos[ ( a)] (19.19-a)2 Cos[ ( a)] (30.06-a)2 Cos[ ( a)] (39.53-a)2 Cos[ ( a)] Thin plate - spline Employing polyharmonic interpolation (Fasshauer 2007) the approximation is, P = φ φ φ φ φ φ φ φ φ 9 ] where φ i = r 2 log (r) and alternatively r = a - a i for a a i r = 0 for a = a i As we can see, in this way one can hardly find a simple relation, although all of these methods give quite satisfactory solution for the problem, see Table Table 12.2 Errors of the different methods of approximation of the observed planetary data Model 1 - R 2 Standard Deviation Polynomial Regression Neural Network Support Vector Machine Thin plate spline interpolation where R stands for the Root Square Mean Error (RMSE). Remarks. 1 fifth order algebraic polynomial. 2 feedforward neural network with two nodes and linear tail

7 SymbolicRegression_12.nb 7. feedforward neural network with two nodes and linear tail. 3 with Gaussian- kernel. 4 basis function: x 2 Log( x ) Now let us see, how can we solve this problem with symbolic regression Symbolic regression Fig shows the Pareto - front of the generated models via DataModeler. The points represent the generated models. The red points stand for the models belong to the Pareto-front. Fig The Pareto-front (red points) and the evaluated models in case of the Kepler s problem In Table 12.3 some of the models of the Pareto front can be seen. Table Model selection riport Model Complexity 1 - R 2 Function x x ( x) x x x a 3/ x 3/2 + 2 x x 3/2 + 2 x x x 3/2 + 2 x * x x 3/ x x + 5 x x 3/ a +a x x x + x 12+2 x+x 2 It goes without saying that our candidate is the 4th model, since it has a small error and at the same time its complexity is low. In practice, it is useful to carry out an additional nonlinear regression with the candidate model in order to refine the value of its parameters. This means, we use symbolic regression to find the best model structure and then its

8 8 SymbolicRegression_12.nb the value of its parameters. This means, we use symbolic regression to find the best model structure and then its parameters will be improved via standard nonlinear regression, in order to avoid a long run of the symbolic regression procedure. Now let us compute the parameters p i of the candidate model, P = p 0 + p 1 a p 2 Table 12.4 indicates that the estimation of the parameter p 0 is unreliable, since its confidence interval is very wide and the statistic P-Value is high, For this model 1 - R 2 = Table The statistics of the estimated parameters via nonlinear regression Parameter Estimated Value Confidencia Interval P- Value p { , } p { , } p { , } Now let us repeat the regression for the model P = p 1 a p2 we get the following values of the two parameters, see Table 12.5 Table The statistics of the estimated parameters via nonlinear regression Parameter Estimated Value Confidence Interval P- Value p { , p { , From practical point of view one considers p 1 = 1 and a = p 2 = 1.5. Although the error of this simplified model is somewhat higher, since 1 - R 2 = < Now let us see, how we compute the symbolic result via Mathematica Application Mathematica Mathematica itself has not a built-in function for symbolic regression, however a third party product DataModeler package developed by Evolved Analytics ( can be employed. Let us load the package << DataModeler` First we should provide the input - output corresponding data, see Table 12.1 samplepts = SetPrecision[{0.39, 0.72, 1., 1.52, 5.2, 9.54, 19.19, 30.06, 39.53}, 10]; observedresponse = SetPrecision[{0.24, 0.61, 1., 1.88, 11.86, 29.46, 84.01, , }, 10]; Then we compute the Pareto front. The computation is carried out parallel in order to decrease the running time.

9 SymbolicRegression_12.nb 9 Timing@ ParetoFrontLogPlot[ quickmodels1 = SymbolicRegression[ samplepts, observedresponse, DataVariables {x}, AlignModel True, RobustModels True, TimeConstraint 50, FitnessPrecision 10 ] ] , Fig The Pareto-front (red points) computed by Mathematica. The models represented by the Pareto front are in the following table. Here we listed out models which have less error than err and their complexity is lower than comp err = 0.01; comp = 150; ModelSelectionReport[quickModels1, QualityBox -> {comp, err}] Model Selection Report Complexity 1-R 2 Function x ( x) x x x x x x 3/ x 3/ x x 3/ x (10 +x) x x x 3/ x x x 3/ x x 3/ x (4.93+x) x x 3/ x x x 3/2 - (7+x) x2 (11.82+x) x x 3/ x 2 (( x) ( x)) x x x 5/ x +x x x x x + x 2 Table The some good models representing the Pareto- front Kepler considered the 5th model, and leaved out the small constant: x 3! In order to use the selected model for

10 10 SymbolicRegression_12.nb further computations we should do the following operations Here we select the model having less error than and complexity is not greater than 30! kp = SelectModels[quickModels1, QualityBox -> {30, }]; This already a Mathematica expression kpm = ModelPhenotype[kp] // First x 3 and, since u = kpm[[2]] x 3 then the Kepler model is KP = Round Coefficient u, x 3 u[[2]] x Teaching Example Let us try to approximate the following function on the basis of some discrete points. The function is z[{x_, y_}] := (x 2 - y 2 ) Sin[0.5 x] p1 = Plot3D[z[{x, y}], {x, -10, 10}, {y, -10, 10}, PlotPoints {30, 30}, BoxRatios {1, 1, 0.5}] Fig The function to be approximated Let the discrete data (x i, y i, f (x i, y i ) data = Flatten[Table[N[{i - 10, j - 10, z[{i - 10, j - 10}]}], {i, 0, 30, 4}, {j, 0, 30, 4}], 1];

11 SymbolicRegression_12.nb 11 Show[{p1, ListPointPlot3D[data, PlotStyle Directive[Blue, PointSize[0.015]], BoxRatios {1, 1, 0.5}]}] Fig The given points We employ symbolic regression, since we should like to have an explicit expression for discribing the surface. Now we separate the data: the inputs (x i, y i ) and the outputs f (x i, y i ) inputs = Transpose[Take[Transpose[data], {1, 2}]] {{-10., -10.}, {-10., -6.}, {-10., -2.}, {-10., 2.}, {-10., 6.}, {-10., 10.}, {-10., 14.}, {-10., 18.}, {-6., -10.}, {-6., -6.}, {-6., -2.}, {-6., 2.}, {-6., 6.}, {-6., 10.}, {-6., 14.}, {-6., 18.}, {-2., -10.}, {-2., -6.}, {-2., -2.}, {-2., 2.}, {-2., 6.}, {-2., 10.}, {-2., 14.}, {-2., 18.}, {2., -10.}, {2., -6.}, {2., -2.}, {2., 2.}, {2., 6.}, {2., 10.}, {2., 14.}, {2., 18.}, {6., -10.}, {6., -6.}, {6., -2.}, {6., 2.}, {6., 6.}, {6., 10.}, {6., 14.}, {6., 18.}, {10., -10.}, {10., -6.}, {10., -2.}, {10., 2.}, {10., 6.}, {10., 10.}, {10., 14.}, {10., 18.}, {14., -10.}, {14., -6.}, {14., -2.}, {14., 2.}, {14., 6.}, {14., 10.}, {14., 14.}, {14., 18.}, {18., -10.}, {18., -6.}, {18., -2.}, {18., 2.}, {18., 6.}, {18., 10.}, {18., 14.}, {18., 18.}} outputs = Transpose[Take[Transpose[data], {3, 3}]] // Flatten {0., , , , , 0., , , , 0., , , 0., , , , , , 0., 0., , , , , , , 0., 0., , , , , , 0., , , 0., , , , 0., , , , , 0., , , , , , , , , 0., , , , , , , , , 0.} Then we compute the Pareto front, which now is not convex, see Fig ,

12 12 SymbolicRegression_12.nb ParetoFrontLogPlot[ model = SymbolicRegression[inputs, outputs, AlignModel True, RobustModels True, FitnessPrecision 10, TimeConstraint 500 ] ] , Fig The Pareto front now is not convex The selected models can be seen in Table 12.7 ModelSelectionReport[model, QualityBox {500, 0.01}] Model Selection Report Complexity 1-R 2 Function x x x x x x x x 1 +x x x 1 x x 1 3 x x 1 5 x x 1 3 x 1 2 x x x x (7.78- x 1) x x x 5 1 x x x 1 x x 3 1 x x 1 5 x x 1 x 1 2 x x x x 1 3 x x x x x x x x x 1 + x x x x 1 x x 1 3 x x 1 5 x x 1 7 x x 1 3 x 1 2 x Model selection Table 12.7 The eminent models representing the Pareto- front Now we select the model with complexity 262 and error kp = SelectModels[model, QualityBox -> {262, 0.005}]; kpm = ModelPhenotype[kp] // Flatten x x x x x x x x1 + x1 2 + x x1 x x1 3 x x1 5 x x1 3 x1 2 x2 2

13 SymbolicRegression_12.nb 13 p2 = Plot3D[kpm, {x1, -10, 10}, {x2, -10, 10}, PlotPoints {30, 30}, BoxRatios {1, 1, 0.5}] Fig The selected model Now let us see, how the data points fit to the surface of the model, see Fig Show[{p2, ListPointPlot3D[data, PlotStyle Directive[Blue, PointSize[0.015]], BoxRatios {1, 1, 0.5}]}] Fig The selected model and the data points The fitting looks quite good. However, comparing the surfaces of the original model and the selected one, we can see the differences, see Fig left side,

14 14 SymbolicRegression_12.nb Show[{p2, p1}] Fig The surfaces of the original and the selected model In the data points, we have good fitting but where were no points, the fitting is less efficient.this fact shows that we should not select the model with the smallest error this would mean in general overlearning. Let us select the best model kp = SelectModels[model, QualityBox -> {302, 0.000}]; kpm = ModelPhenotype[kp] // Flatten x x x x x x x x1 + x1 2 + x x x1 x x1 3 x x1 5 x x1 7 x x1 3 x1 2 x2 2 p2 = Plot3D[kpm, {x1, -10, 10}, {x2, -10, 10}, PlotPoints {30, 30}, BoxRatios {1, 1, 0.5}]

15 SymbolicRegression_12.nb 15 Show[{p2, ListPointPlot3D[data, Fig The best selected model PlotStyle Directive[Blue, PointSize[0.015]], BoxRatios {1, 1, 0.5}]}] Show[{p2, p1}] Fig The best selected model and the data points 12-7 Extention of function space Fig The surfaces of the original and the best selected model In the basic option, the choice of the type basic functions, which can be employed by the method is quite poor. To gettin better results these function space can be extended, see Help DataModeler SymbolicRegression FunctionPatterns The Pareto - front can be seen in Fig Table? shows that we have got back the original model!

16 16 SymbolicRegression_12.nb ParetoFrontLogPlot[ model = SymbolicRegression[inputs, outputs, AlignModel True, RobustModels True, FitnessPrecision 10, FunctionPatterns BuildFunctionPatterns["ExtendedMath", Sin], MemoryLimit 5 "GB", TimeConstraint 500] ] , Fig The Pareto front in case of extended function space ModelSelectionReport[model, QualityBox {500, }] Model Selection Report Complexity 1-R 2 Function Sin x1 Table 12.8 The bets model of the Pareto- front kp = SelectModels[model, QualityBox -> {52, 0.2}]; kpm = ModelPhenotype[kp] // Flatten x1 2 + x2 2 Sin x x12 +x 2 2

17 SymbolicRegression_12.nb 17 p2 = Plot3D[kpm, {x1, -10, 10}, {x2, -10, 10}, PlotPoints {30, 30}, BoxRatios {1, 1, 0.5}] Show[{p2, ListPointPlot3D[data, Fig The selected model PlotStyle Directive[Blue, PointSize[0.015]], BoxRatios {1, 1, 0.5}]}] Now the two surfaces fit perfectly! Fig The selected model with the data points

18 18 SymbolicRegression_12.nb Show[{p2, p1}] 12-8 Application example Geometric transformation Problem definition Fig The surface of the original and the selected models Transformation of coordinates is important in computer vision, photogrammetry as well as in geodesy. In this example we consider some standard 2D transformations as similarity, affine and projective transformations between the coordinates of the fiducial marks on the comparator plate and that of the corresponding points on the reseau plate. The 16 observed (x, y) and calibrated reseau (X, Y) coordinates are given in Table 12.9 Similarity transformation Table 12.9 The coordinates of the corresponding observation points on the comparator and reseau planes Points x [mm] [mm] [mm] [mm] The advantage of this model is that it is linear in the coordinates as well as in the parameters. Consequently an iterative solution is not required and the inverse transformation is easy. This transformation can be parametrized in the following form, y X Y

19 SymbolicRegression_12.nb 19 x y = a b -b a X Y + c d For each observed ith point the following pair of observations equation can be written for the residuals (r xi, r yi ), x i y i 1 0 y i -x i 0 1 a b c d + -x i -y i = r x i r yi A minimum of two fiducial marks or reseau crosses are required for a unique solution. Having more observation points linear least square can be directly applied. In our example the result can be seen in Table Table The estimated parameters of the similarity transformation model Parameter Value a b c d The norm of the residual vectors in the observation points are shown in Table Affine transformation Table The norm of the residual vectors at the observation points in case of similarity transformation model Points r xy [mm] This is also a linear model although it needs six parameters, therefore minimum 3 fiducial marks or reseau crosses are required for a unique solution. Iterative solution is not required and the inverse transformation is easy. This model can be parametrized in the following form x y = a d b e X Y + c f For each observed ith point the following pair of observations equation can be written for the residuals (r xi, r yi ), y i x i x i 1 y i a b c d e f + -x i -y i = r x i r yi Having more than 3 observation points linear least square can be directly applied. In our example the result can be seen in Table 12.12

20 20 SymbolicRegression_12.nb Table The estimated parameters of the affine transformation model Parameters Values a b c d e f The norm of the residual vectors in the observation points are shown in Table Projective transformation Table The norm of the residual vectors at the observation points in case of affine transformation model Points r xy [mm] The projective transformation can be expressed by the following two equations, and x = a 1 X + a 2 Y + a 3 c 1 X + c 2 Y + 1 y = b 1 X + b 2 Y + a 3 c 1 X + c 2 Y + 1 These equations are a special case of the collinearity condition for mapping of 2 D points from one plane onto another. There are 8 unknown parameters, so if only four fiducial marks are available, the solution is not unique. This is perhaps the main reason why it is not frequently used. However, it is relevant for systems that have been retro-fitted with a reseau plate, such as the Hasselblad camera used to acquire the assignment imagery. The least square solution is nonlinear due to the rational nature of the functions. However, an approximate linear solution can be implemented if both sides of the equations are multiplied by the denominator and partial derivates with respect to the observables (as in the combined adjutment model) are ignored. The inverse of the transformation can be also computed by solving the equation system in symbolic form for X and Y, we obtain

21 SymbolicRegression_12.nb 21 and X = (a 2 (-y + a 3 ) + (x - a 3 ) b 2 + (-x + y) a 3 c 2 )/(-x b 2 c 1 + a 2 (-b 1 + y c 1 ) + x b 1 c 2 + a 1 (b 2 - y c 2 )) Y = (a 1 (y - a 3 ) + (-x + a 3 ) b 1 + (x - y) a 3 c 1 ) /(-x b 2 c 1 + a 2 (-b 1 + y c 1 ) + x b 1 c 2 + a 1 (b 2 - y c 2 )) Having more than 4 observation points nonlinear least square can be applied. In our example the result can be seen in Table Table The estimated parameters of the projective transformation model Parameters Values a a a b b b c 1 0 c The norm of the residual vectors in the observation points are shown in Table Table The norm of the residual vectors at the observation points in case of projective transformation model Application of symbolic regression Points r xy [mm] {{-110, -110}, {-40, -110}, {40, -110}, {110, -110}, {110, -40}, {40, -40}, {-40, -40}, {-100, -40}, {-110, 40}, {-40, 40}, {40, 40}, {110, 40}, {110, 110}, {40, 110}, {-40, 110}, {-110, 110}} outputx = { , , , , , , , , , , , , , , , }; outputy = {-107.4, , , , , , , , , , , , , , , };

22 22 SymbolicRegression_12.nb inputxy = {{-110, -110}, {-40, -110}, {40, -110}, {110, -110}, {110, -40}, {40, -40}, {-40, -40}, {-100, -40}, {-110, 40}, {-40, 40}, {40, 40}, {110, 40}, {110, 110}, {40, 110}, {-40, 110}, {-110, 110}}; Per definition neither the model nor the parameters of the model are not known. However in order to have a chance for computing the inverse of the transformation trigonometrical functions are excluded from the functionset. Now, we create linear and nolinear models, but linear models will be optimized, Timing@ ParetoFrontLogPlot[ ] model = SymbolicRegression[inputXY, outputx, OptimizeLinearModel True, ] TimeConstraint , Fig The Pareto fron in case of Optimal Linear model for x = (X, Y)

23 SymbolicRegression_12.nb 23 ModelSelectionReport model, QualityBox 250, 10-8 Page 1 Page 2 Model Selection Report Complexity 1-R 2 Function x x x x x x x 1 x x x x 1 x x x x 1 x x x1 x1 x1 x2 x x x 1 x x x x 1 x x x x1 x2 x x x 1 x x x x 1 x x x x x1 x2 x2 2 x x x 1 x x 2 2 x x x 1 x x x 2 2 x x x 1 x x x 2 2 x x x x x 1 x (3 -x x1 2) x x x x x x 1 x x (7 - x 2 ) x 3 2 x x x x x x2 x1 x1 2 2 x x x (x 1 - x 2) 2 + Table Linear and nonlinear models for x = (X, Y) The best linear model among the linear and nonlinear models is kp = SelectModels model, QualityBox -> 30, ; kpm = ModelPhenotype[kp] // Flatten x x x1 x2 Now, only linear models will be created

24 24 SymbolicRegression_12.nb ParetoFrontLogPlot[model = CreateLinearModel[ inputxy, outputx, PolynomialOrder 4, TimeConstraint 10]] , Fig The Pareto fron in case of direct linear modelling for x = (X, Y) ModelSelectionReport model, QualityBox 230, 10-8 Model Selection Report Complexity 1-R 2 Function x x 1 x x 2 1 x x 3 1 x x x x x x x 1 x x 2 1 x x x 1 x x 4 2 Table The linear model for x = (X, Y) kp = SelectModels model, QualityBox -> 205, 10-9 ; kpm = ModelPhenotype[kp] // Flatten x x x x x x1 x x1 2 x x1 3 x x x1 x x1 2 x x x1 x x2 4 Surprisingly, the best linear model is among the linear and nonlinear models is better than direct linear model! Now we select both xv1 = ` `X `Y `*^-7X Y X Y X Y and

25 SymbolicRegression_12.nb 25 xv2 = `*^ `x `*^-6x `*^-9x `*^-10x `x `*^-6x1 x `*^-10x1 2 x `*^-11x1 3 x `x `*^-9x1 x `*^-13x1 2 x `*^-9x `*^-11x1 x `*^-7x2 4 /. {x1 X, x2 Y} X X X X Y X Y X 2 Y X 3 Y Y X Y X 2 Y Y X Y Y 4 Similarly let us compute the y (X, Y) function, too. Timing@ ParetoFrontLogPlot[ model = SymbolicRegression[inputXY, outputy, OptimizeLinearModel True, TimeConstraint 150 ] ] , Fig The Pareto fron in case of Optimal Linear model for y = (X, Y)

26 26 SymbolicRegression_12.nb ModelSelectionReport model, QualityBox 250, 10-8 Page 1 Page 2 Model Selection Report Complexity 1-R 2 Function x x x x x x x 1 x x x x x x x x x x 1 x x x1 x x x x x x x x2 x x x x x x 1 2 x / ( x 1 + x 2) x x x x x / ( x 1 +x 2) x x x x x 1 2 x / ( x 1 + x 2) x x x x x 2 1 x x x 2 x x x x x 1 x x 1 +x / ( x 1 + x 2 ) x x x x 1 x x1 +x / ( x 1 + x 2) x x x x x x x (4+ x 1 +x 2) x x /( x 1 -x 2) x / ( x 1 +x 2) + x2 The best linear model is Table Linear and nonlinear models for y = (X, Y) kp = SelectModels model, QualityBox -> 42, ; kpm = ModelPhenotype[kp] // Flatten x x x1 x x2 2 Now, only linear models will be created

27 SymbolicRegression_12.nb 27 ParetoFrontLogPlot[model = CreateLinearModel[ inputxy, outputy, PolynomialOrder 4, TimeConstraint 10]] , Fig The Pareto fron in case of direct linear modelling for y = (X, Y) ModelSelectionReport model, QualityBox 230, 10-8 Model Selection Report Complexity 1-R 2 Function x x 1 x x 2 1 x x 3 1 x x x x x x x 1 x x 2 1 x x x 1 x x 4 2 Table The linear model for y = (X, Y) kp = SelectModels model, QualityBox -> 205, ; kpm = ModelPhenotype[kp] // Flatten x x x x x x1 x x1 2 x x1 3 x x x1 x x1 2 x x x1 x x2 4 Now the best linear model is among the linear and nonlinear models has smaller complexity but higher error level than the direct linear model! Therefore we select both yv1 = ` `x `x `*^-7x1 x `*^-7x2 2 /. {x1 X, x2 Y} X Y X Y Y 2

28 28 SymbolicRegression_12.nb yv2 = `*^ `x `x `*^-9x `*^-10x `x `*^-8x1 x `*^-9x1 2 x `*^-12x1 3 x `x `*^-9x1 x `*^-11x1 2 x `*^-9x `*^-11x1 x `*^-8x2 4 /. {x1 X, x2 Y} X X X X Y X Y X 2 Y X 3 Y Y X Y X 2 Y Y X Y Y 4 Now we compute the sum of residuals rxy1 = MapThread[Norm[{(#2 - xv1), (#3 - yv1)}] /. {X #1[[1]], Y -> #1[[2]]} &, {inputxy, outputx, outputy}]; TableForm[rxy1] Apply[Plus, rxy1] RMSE % / rxy2 = MapThread[Norm[{(#2 - xv2), (#3 - yv2)}] /. {X #1[[1]], Y -> #1[[2]]} &, {inputxy, outputx, outputy}]; TableForm[rxy2] Apply[Plus, rxy2]

29 SymbolicRegression_12.nb 29 RMSE % / So the higher order pair give better result. Table The norm of the residual vectors at the observation points in case of the nonlinear transformation model obtained by SR Points r xy [mm] The Fig illustrates how this nonlinear transformation warpes (deformates) the original (X, Y) plane into the (x, y) plane

30 30 SymbolicRegression_12.nb Fig The original plane and below the plane after warping via the nonlinear transformation Comparison of the different transformation models Max[rxy2] StandardDeviation[rxy2] It is not surprising that transformation models having more parameters and nonlinear form can provide less error than others. Let us summarize some statistic values of the different methods in Table Transformation model Table Some statistical values of the different transformation models Max of Absolute Errors Standard Deviation of the Absolute Errors RMSE of the residual Errors [mm] [mm] [mm] Similarity Affine Projective Symbolic Regression Both SR models are better than the other models. Although the nonlinear transformation obtained by the symbolic regression gives the best fitting, to compute the inverse of the transformation is not an easy task. However it can be solved via Groebner basis.

Physics Unit 7: Circular Motion, Universal Gravitation, and Satellite Orbits. Planetary Motion

Physics Unit 7: Circular Motion, Universal Gravitation, and Satellite Orbits. Planetary Motion Physics Unit 7: Circular Motion, Universal Gravitation, and Satellite Orbits Planetary Motion Geocentric Models --Many people prior to the 1500 s viewed the! Earth and the solar system using a! geocentric

More information

Koza s Algorithm. Choose a set of possible functions and terminals for the program.

Koza s Algorithm. Choose a set of possible functions and terminals for the program. Step 1 Koza s Algorithm Choose a set of possible functions and terminals for the program. You don t know ahead of time which functions and terminals will be needed. User needs to make intelligent choices

More information

Kepler's constant in Earth-Moon System is gradually. decreasing

Kepler's constant in Earth-Moon System is gradually. decreasing Kepler's constant in Earth-Moon System is gradually decreasing Hu Wang Email:vortex.wh@hotmail.com Abstract Studies on nautiloids, coral fossils, rotation of the Earth, and Earth-Moon distance variation

More information

Investigating the Solar System

Investigating the Solar System Investigating the Solar System This Workbook belongs to: Our Local Star: The Sun Location in The Solar System Interesting Facts 1. 2. 3. 4. Name of Star: THE SUN 5. Draw and Color your own Sun in the blank

More information

Observational Astronomy - Lecture 4 Orbits, Motions, Kepler s and Newton s Laws

Observational Astronomy - Lecture 4 Orbits, Motions, Kepler s and Newton s Laws Observational Astronomy - Lecture 4 Orbits, Motions, Kepler s and Newton s Laws Craig Lage New York University - Department of Physics craig.lage@nyu.edu February 24, 2014 1 / 21 Tycho Brahe s Equatorial

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.

Kernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt. SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University

More information

Discriminative Direction for Kernel Classifiers

Discriminative Direction for Kernel Classifiers Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering

More information

Holdout and Cross-Validation Methods Overfitting Avoidance

Holdout and Cross-Validation Methods Overfitting Avoidance Holdout and Cross-Validation Methods Overfitting Avoidance Decision Trees Reduce error pruning Cost-complexity pruning Neural Networks Early stopping Adjusting Regularizers via Cross-Validation Nearest

More information

Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators

Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators Andrea Mambrini 1 University of Birmingham, Birmingham UK 6th June 2013 1 / 33 Andrea Mambrini GSGP: theory-laden

More information

Yes, inner planets tend to be and outer planets tend to be.

Yes, inner planets tend to be and outer planets tend to be. 1. Planet Density Make some general comments about inner and outer planets density Inner Planets Density Outer Planets Density Is there a pattern or a trend in planet density? Yes, inner planets tend to

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels

SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels Karl Stratos June 21, 2018 1 / 33 Tangent: Some Loose Ends in Logistic Regression Polynomial feature expansion in logistic

More information

Chapter 4 Thrills and Chills +Math +Depth Acceleration of the Moon +Concepts The Moon is 60 times further away from the center of Earth than objects on the surface of Earth, and moves about Earth in an

More information

Support Vector Machine (SVM) and Kernel Methods

Support Vector Machine (SVM) and Kernel Methods Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin

More information

HSC General Maths. Name: Tracey Hughes-Butters 2011 Mathtastic Learning GENERAL MATHEMATICS NOTES HSC ALGEBRAIC MODELLING 3

HSC General Maths. Name: Tracey Hughes-Butters 2011 Mathtastic Learning GENERAL MATHEMATICS NOTES HSC ALGEBRAIC MODELLING 3 Allgebraic Modelllling 3 Notes... HSC General Maths Name: Tracey Hughes-Butters 011 Mathtastic Learning 1 HSC CAPACITY MATRIX GENERAL MATHEMATICS TOPIC: Algebraic Modelling 3 & 4a Algebra skills & Linear

More information

Modeling Data with Functions

Modeling Data with Functions Chapter 11 Modeling Data with Functions 11.1 Data Modeling Concepts 1 11.1.1 Conceptual Explanations: Modeling Data with Functions In school, you generally start with a function and work from there to

More information

Unit: Planetary Science

Unit: Planetary Science Orbital Motion Kepler s Laws GETTING AN ACCOUNT: 1) go to www.explorelearning.com 2) click on Enroll in a class (top right hand area of screen). 3) Where it says Enter class Code enter the number: MLTWD2YAZH

More information

Lecture 9 Evolutionary Computation: Genetic algorithms

Lecture 9 Evolutionary Computation: Genetic algorithms Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Simulation of natural evolution Genetic algorithms Case study: maintenance scheduling with genetic

More information

Name. Satellite Motion Lab

Name. Satellite Motion Lab Name Satellite Motion Lab Purpose To experiment with satellite motion using an interactive simulation in order to gain an understanding of Kepler s Laws of Planetary Motion and Newton s Law of Universal

More information

Importance of the study of extrasolar planets. Exoplanets Introduction. Importance of the study of extrasolar planets

Importance of the study of extrasolar planets. Exoplanets Introduction. Importance of the study of extrasolar planets Importance of the study of extrasolar planets Exoplanets Introduction Planets and Astrobiology (2017-2018) G. Vladilo Technological and scientific spin-offs Exoplanet observations are driving huge technological

More information

Support Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM

Support Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM 1 Support Vector Machines (SVM) in bioinformatics Day 1: Introduction to SVM Jean-Philippe Vert Bioinformatics Center, Kyoto University, Japan Jean-Philippe.Vert@mines.org Human Genome Center, University

More information

CSC 4510 Machine Learning

CSC 4510 Machine Learning 10: Gene(c Algorithms CSC 4510 Machine Learning Dr. Mary Angela Papalaskari Department of CompuBng Sciences Villanova University Course website: www.csc.villanova.edu/~map/4510/ Slides of this presenta(on

More information

1. The two triangles shown below are similar. This means that all the angles are equal and the sides are proportional.

1. The two triangles shown below are similar. This means that all the angles are equal and the sides are proportional. 1. The two triangles shown below are similar. This means that all the angles are equal and the sides are proportional. a. How many times bigger is the big triangle in comparison to the little triangle?

More information

1 The Solar System. 1.1 a journey into our galaxy

1 The Solar System. 1.1 a journey into our galaxy 1 The Solar System Though Pluto, and the far-flung depths of the Solar System, is the focus of this book, it is essential that Pluto is placed in the context of the planetary system that it inhabits our

More information

Forecasting & Futurism

Forecasting & Futurism Article from: Forecasting & Futurism December 2013 Issue 8 A NEAT Approach to Neural Network Structure By Jeff Heaton Jeff Heaton Neural networks are a mainstay of artificial intelligence. These machine-learning

More information

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 BIOLOGICAL INSPIRATIONS Some numbers The human brain contains about 10 billion nerve cells (neurons) Each neuron is connected to the others through 10000

More information

Conceptual Explanations: Modeling Data with Functions

Conceptual Explanations: Modeling Data with Functions Conceptual Explanations: Modeling Data with Functions In school, you generally start with a function and work from there to numbers. Newton s Law tells us that F=ma. So if you push on a 3kg object with

More information

Related Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM

Related Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM Lecture 9 SEM, Statistical Modeling, AI, and Data Mining I. Terminology of SEM Related Concepts: Causal Modeling Path Analysis Structural Equation Modeling Latent variables (Factors measurable, but thru

More information

NSTA Web Seminar: Discover the Universe from Galileo to Today

NSTA Web Seminar: Discover the Universe from Galileo to Today LIVE INTERACTIVE LEARNING @ YOUR DESKTOP NSTA Web Seminar: Discover the Universe from Galileo to Today Presented by: Dr. Natalie Batalha Tuesday, January 20, 2009 International Year of Astronomy: Advances

More information

ARTIFICIAL NEURAL NETWORK WITH HYBRID TAGUCHI-GENETIC ALGORITHM FOR NONLINEAR MIMO MODEL OF MACHINING PROCESSES

ARTIFICIAL NEURAL NETWORK WITH HYBRID TAGUCHI-GENETIC ALGORITHM FOR NONLINEAR MIMO MODEL OF MACHINING PROCESSES International Journal of Innovative Computing, Information and Control ICIC International c 2013 ISSN 1349-4198 Volume 9, Number 4, April 2013 pp. 1455 1475 ARTIFICIAL NEURAL NETWORK WITH HYBRID TAGUCHI-GENETIC

More information

Physics 12. Unit 5 Circular Motion and Gravitation Part 2

Physics 12. Unit 5 Circular Motion and Gravitation Part 2 Physics 12 Unit 5 Circular Motion and Gravitation Part 2 1. Newton s law of gravitation We have seen in Physics 11 that the force acting on an object due to gravity is given by a well known formula: F

More information

Large and small planets

Large and small planets Large and small planets Journey through the Solar System C 41 time 50 minutes. Preparation For the activity Planets show the planets under each other on the board, as shown in the table. learning outcomes

More information

9/12/2010. The Four Fundamental Forces of Nature. 1. Gravity 2. Electromagnetism 3. The Strong Nuclear Force 4. The Weak Nuclear Force

9/12/2010. The Four Fundamental Forces of Nature. 1. Gravity 2. Electromagnetism 3. The Strong Nuclear Force 4. The Weak Nuclear Force The Four Fundamental Forces of Nature 1. Gravity 2. Electromagnetism 3. The Strong Nuclear Force 4. The Weak Nuclear Force The Universe is made of matter Gravity the force of attraction between matter

More information

Lecture 13. Gravity in the Solar System

Lecture 13. Gravity in the Solar System Lecture 13 Gravity in the Solar System Guiding Questions 1. How was the heliocentric model established? What are monumental steps in the history of the heliocentric model? 2. How do Kepler s three laws

More information

GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS

GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS A genetic algorithm is a random search technique for global optimisation in a complex search space. It was originally inspired by an

More information

Introduction. Chapter 1

Introduction. Chapter 1 Chapter 1 Introduction In this book we will be concerned with supervised learning, which is the problem of learning input-output mappings from empirical data (the training dataset). Depending on the characteristics

More information

Support Vector Machines and Kernel Methods

Support Vector Machines and Kernel Methods 2018 CS420 Machine Learning, Lecture 3 Hangout from Prof. Andrew Ng. http://cs229.stanford.edu/notes/cs229-notes3.pdf Support Vector Machines and Kernel Methods Weinan Zhang Shanghai Jiao Tong University

More information

CS145: INTRODUCTION TO DATA MINING

CS145: INTRODUCTION TO DATA MINING CS145: INTRODUCTION TO DATA MINING 5: Vector Data: Support Vector Machine Instructor: Yizhou Sun yzsun@cs.ucla.edu October 18, 2017 Homework 1 Announcements Due end of the day of this Thursday (11:59pm)

More information

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD WHAT IS A NEURAL NETWORK? The simplest definition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided

More information

COMS 4771 Introduction to Machine Learning. Nakul Verma

COMS 4771 Introduction to Machine Learning. Nakul Verma COMS 4771 Introduction to Machine Learning Nakul Verma Announcements HW1 due next lecture Project details are available decide on the group and topic by Thursday Last time Generative vs. Discriminative

More information

18 An Eclipsing Extrasolar Planet

18 An Eclipsing Extrasolar Planet Name: Date: 18 An Eclipsing Extrasolar Planet 18.1 Introduction One of the more recent new fields in astronomy is the search for (and discovery of) planets orbiting around stars other than our Sun, or

More information

Early Theories. Early astronomers believed that the sun, planets and stars orbited Earth (geocentric model) Developed by Aristotle

Early Theories. Early astronomers believed that the sun, planets and stars orbited Earth (geocentric model) Developed by Aristotle Planetary Motion Early Theories Early astronomers believed that the sun, planets and stars orbited Earth (geocentric model) Developed by Aristotle Stars appear to move around Earth Observations showed

More information

APPENDIX III - PLANETARY SYMBOLS DECODED CDEFGHIJ VENUS MARS JUPITER SATURN URANUS NEPTUNE PLUTO

APPENDIX III - PLANETARY SYMBOLS DECODED CDEFGHIJ VENUS MARS JUPITER SATURN URANUS NEPTUNE PLUTO PAGE 1 APPENDIX III - PLANETARY SYMBOLS DECODED There have been many attempts to decode the symbols of the planets. Most all of these decodings are either very complicated or somewhat mystical in nature.

More information

State Vector Reference Body Transformations Throughout The Solar System

State Vector Reference Body Transformations Throughout The Solar System Introduction One of the more mind-boggling programming challenges encountered during interplanetary trajectory design software development is how to transform a specified inertial position and velocity

More information

Pedagogical information

Pedagogical information SHOOTING STAR Shooting Star, an interactive computer simulation using calculation power of super computers. Students should investigate and become familiar with Kepler's laws, Newton's theory of gravitation,

More information

Characterization of Jet Charge at the LHC

Characterization of Jet Charge at the LHC Characterization of Jet Charge at the LHC Thomas Dylan Rueter, Krishna Soni Abstract The Large Hadron Collider (LHC) produces a staggering amount of data - about 30 petabytes annually. One of the largest

More information

PHYSICS. Chapter 13 Lecture FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E RANDALL D. KNIGHT Pearson Education, Inc.

PHYSICS. Chapter 13 Lecture FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E RANDALL D. KNIGHT Pearson Education, Inc. PHYSICS FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E Chapter 13 Lecture RANDALL D. KNIGHT Chapter 13 Newton s Theory of Gravity IN THIS CHAPTER, you will learn to understand the motion of satellites

More information

Unit 1: The Earth in the Universe

Unit 1: The Earth in the Universe Unit 1: The Earth in the Universe 1. The Universe 1.1. First ideas about the Universe 1.2. Components and origin 1.3. Sizes and distances 2. The Solar System 3. The planet Earth 3.1. Movements of the Earth

More information

Support Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar

Support Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar Data Mining Support Vector Machines Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar 02/03/2018 Introduction to Data Mining 1 Support Vector Machines Find a linear hyperplane

More information

Introduction to Computer Graphics (Lecture No 07) Ellipse and Other Curves

Introduction to Computer Graphics (Lecture No 07) Ellipse and Other Curves Introduction to Computer Graphics (Lecture No 07) Ellipse and Other Curves 7.1 Ellipse An ellipse is a curve that is the locus of all points in the plane the sum of whose distances r1 and r from two fixed

More information

Lecture 6. Regression

Lecture 6. Regression Lecture 6. Regression Prof. Alan Yuille Summer 2014 Outline 1. Introduction to Regression 2. Binary Regression 3. Linear Regression; Polynomial Regression 4. Non-linear Regression; Multilayer Perceptron

More information

Copyright 2008 Pearson Education, Inc., publishing as Pearson Addison-Wesley.

Copyright 2008 Pearson Education, Inc., publishing as Pearson Addison-Wesley. Chapter 13. Newton s Theory of Gravity The beautiful rings of Saturn consist of countless centimeter-sized ice crystals, all orbiting the planet under the influence of gravity. Chapter Goal: To use Newton

More information

A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE

A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE Li Sheng Institute of intelligent information engineering Zheiang University Hangzhou, 3007, P. R. China ABSTRACT In this paper, a neural network-driven

More information

VISUAL PHYSICS ONLINE

VISUAL PHYSICS ONLINE VISUAL PHYSICS ONLINE PRACTICAL ACTIVITY HOW DO THE PANETS MOVE? One of the most important questions historically in Physics was how the planets move. Many historians consider the field of Physics to date

More information

Portable Assisted Study Sequence ALGEBRA IIB

Portable Assisted Study Sequence ALGEBRA IIB SCOPE This course is divided into two semesters of study (A & B) comprised of five units each. Each unit teaches concepts and strategies recommended for intermediate algebra students. The second half of

More information

Finding Other Earths. Jason H. Steffen. Asset Earth Waubonsee Community College October 1, 2009

Finding Other Earths. Jason H. Steffen. Asset Earth Waubonsee Community College October 1, 2009 Finding Other Earths Jason H. Steffen Asset Earth Waubonsee Community College October 1, 2009 True Earth Analog Necessities: 1) Main Sequence Star 2) Within the Stellar Habitable Zone 3) Roughly Earth

More information

Lecture 5: Linear Genetic Programming

Lecture 5: Linear Genetic Programming Lecture 5: Linear Genetic Programming CIU036 Artificial Intelligence 2, 2010 Krister Wolff, Ph.D. Department of Applied Mechanics Chalmers University of Technology 41296 Göteborg, Sweden krister.wolff@chalmers.se

More information

Given a planet s radius, its orbital period in days is given by:

Given a planet s radius, its orbital period in days is given by: Sidereal Messenger by Daniel Browning, April 22 2013 Introduction A mathematician s job is never done. This brief item will attempt to improve on an equation that forges a relationship between the orbital

More information

The Law of Ellipses (Kepler s First Law): all planets orbit the sun in a

The Law of Ellipses (Kepler s First Law): all planets orbit the sun in a Team Number Team Members Present Learning Objectives 1. Practice the Engineering Process a series of steps to follow to design a solution to a problem. 2. Practice the Five Dimensions of Being a Good Team

More information

Transit Tracks. Activity G14. What s This Activity About? Tips and Suggestions. What Will Students Do? What Will Students Learn?

Transit Tracks. Activity G14. What s This Activity About? Tips and Suggestions. What Will Students Do? What Will Students Learn? G14 Transit Tracks Activity G14 Grade Level: 7 12 Source: This is a first version of an activity intended to be part of the Full Option Science System (FOSS) Planetary Science Course. Transit Tracks is

More information

Bioinformatics: Network Analysis

Bioinformatics: Network Analysis Bioinformatics: Network Analysis Model Fitting COMP 572 (BIOS 572 / BIOE 564) - Fall 2013 Luay Nakhleh, Rice University 1 Outline Parameter estimation Model selection 2 Parameter Estimation 3 Generally

More information

Chapter 14 Satellite Motion

Chapter 14 Satellite Motion 1 Academic Physics Mechanics Chapter 14 Satellite Motion The Mechanical Universe Kepler's Three Laws (Episode 21) The Kepler Problem (Episode 22) Energy and Eccentricity (Episode 23) Navigating in Space

More information

ASTR 200 : Lecture 6 Introduction to the Solar System Pearson Education Inc., publishing as Addison-Wesley

ASTR 200 : Lecture 6 Introduction to the Solar System Pearson Education Inc., publishing as Addison-Wesley ASTR 200 : Lecture 6 Introduction to the Solar System 1 2004 Pearson Education Inc., publishing as Addison-Wesley Comparative Planetology Studying the similarities among and differences between the planets

More information

FIFTH MIDTERM -- REVIEW PROBLEMS

FIFTH MIDTERM -- REVIEW PROBLEMS Physics 2210 Fall 2005 Paolo Gondolor FIFTH MIDTERM -- REVIEW PROBLEMS A solution set is available on the course web page in pdf format (no solutions for 27-29). 7 1. Assume that the planet Uranus has

More information

Gravity and the Orbits of Planets

Gravity and the Orbits of Planets Gravity and the Orbits of Planets 1. Gravity Galileo Newton Earth s Gravity Mass v. Weight Einstein and General Relativity Round and irregular shaped objects 2. Orbits and Kepler s Laws ESO Galileo, Gravity,

More information

7.4 Universal Gravitation

7.4 Universal Gravitation Circular Motion Velocity is a vector quantity, which means that it involves both speed (magnitude) and direction. Therefore an object traveling at a constant speed can still accelerate if the direction

More information

Evolution Strategies for Constants Optimization in Genetic Programming

Evolution Strategies for Constants Optimization in Genetic Programming Evolution Strategies for Constants Optimization in Genetic Programming César L. Alonso Centro de Inteligencia Artificial Universidad de Oviedo Campus de Viesques 33271 Gijón calonso@uniovi.es José Luis

More information

Understanding Generalization Error: Bounds and Decompositions

Understanding Generalization Error: Bounds and Decompositions CIS 520: Machine Learning Spring 2018: Lecture 11 Understanding Generalization Error: Bounds and Decompositions Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the

More information

Graphical Analysis and Errors MBL

Graphical Analysis and Errors MBL Graphical Analysis and Errors MBL I Graphical Analysis Graphs are vital tools for analyzing and displaying data Graphs allow us to explore the relationship between two quantities -- an independent variable

More information

Evidence that the Earth does not move: Greek Astronomy. Aristotelian Cosmology: Motions of the Planets. Ptolemy s Geocentric Model 2-1

Evidence that the Earth does not move: Greek Astronomy. Aristotelian Cosmology: Motions of the Planets. Ptolemy s Geocentric Model 2-1 Greek Astronomy Aristotelian Cosmology: Evidence that the Earth does not move: 1. Stars do not exhibit parallax: 2-1 At the center of the universe is the Earth: Changeable and imperfect. Above the Earth

More information

Scaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.).

Scaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.). Local Search Scaling Up So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.). The current best such algorithms (RBFS / SMA*)

More information

Decision Tree Learning

Decision Tree Learning Decision Tree Learning Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University References: 1. Machine Learning, Chapter 3 2. Data Mining: Concepts, Models,

More information

Algebra II Unit Breakdown (Curriculum Map Outline)

Algebra II Unit Breakdown (Curriculum Map Outline) QUARTER 1 Unit 1 (Arithmetic and Geometric Sequences) A. Sequences as Functions 1. Identify finite and infinite sequences 2. Identify an arithmetic sequence and its parts 3. Identify a geometric sequence

More information

Data Warehousing & Data Mining

Data Warehousing & Data Mining 13. Meta-Algorithms for Classification Data Warehousing & Data Mining Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 13.

More information

Tycho Brahe

Tycho Brahe Tycho Brahe 1546-1601 At the time of Shakespeare and Elizabeth I and Champlain Lost part of his nose in a duel over who was the best mathematician At 27 he measured the distance of a supernova and a comet

More information

Johannes Kepler ( ) German Mathematician and Astronomer Passionately convinced of the rightness of the Copernican view. Set out to prove it!

Johannes Kepler ( ) German Mathematician and Astronomer Passionately convinced of the rightness of the Copernican view. Set out to prove it! Johannes Kepler (1571-1630) German Mathematician and Astronomer Passionately convinced of the rightness of the Copernican view. Set out to prove it! Kepler s Life Work Kepler sought a unifying principle

More information

Learning Outcomes in Focus

Learning Outcomes in Focus Contextual strand: E&S 3 Learning Outcomes in Focus Students should be able to interpret data to compare the Earth with other planets and moons in the solar system, with respect to properties including

More information

ASTR 200 Reminder Center section = 'no laptop' zone

ASTR 200 Reminder Center section = 'no laptop' zone ASTR 200 Reminder Center section = 'no laptop' zone Laptops OK 1 NO Laptops Laptops OK ASTR 200 Frontiers of Astrophysics www.phas.ubc.ca/~gladman/a200 2 Homework in ASTR 200 The clarity and logic and

More information

IN THE ALMIGHTY GOD NAME Through the Mother of God mediation I do this research

IN THE ALMIGHTY GOD NAME Through the Mother of God mediation I do this research Jupiter Uranus Distance (Metonic Cycle Reason) By Gerges Francis Twadrous 2 nd course student Physics Department - Physics & Math Faculty Peoples' Friendship University Moscow Russia -2010-2013 TEL +201022532292

More information

Jeff Howbert Introduction to Machine Learning Winter

Jeff Howbert Introduction to Machine Learning Winter Classification / Regression Support Vector Machines Jeff Howbert Introduction to Machine Learning Winter 2012 1 Topics SVM classifiers for linearly separable classes SVM classifiers for non-linearly separable

More information

Detecting Dark Matter Halos Sam Beder, Brie Bunge, Adriana Diakite December 14, 2012

Detecting Dark Matter Halos Sam Beder, Brie Bunge, Adriana Diakite December 14, 2012 Detecting Dark Matter Halos Sam Beder, Brie Bunge, Adriana Diakite December 14, 2012 Introduction Dark matter s gravitational pull affects the positions of galaxies in the universe. If galaxies were circular,

More information

The Doppler Method, or Radial Velocity Detection of Planets: I. Technique

The Doppler Method, or Radial Velocity Detection of Planets: I. Technique ASTs309L The Doppler Method, or Radial Velocity Detection of Planets: I. Technique 1. Keplerian Orbits 2. Spectrographs/Doppler shifts 3. Precise Radial Velocity measurements ASTs309L The Doppler Effect:

More information

Introduction to Support Vector Machines

Introduction to Support Vector Machines Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction

More information

Linear & nonlinear classifiers

Linear & nonlinear classifiers Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table

More information

Geometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators

Geometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators Geometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators Andrea Mambrini University of Birmingham, UK NICaiA Exchange Programme LaMDA group, Nanjing University, China 7th

More information

Assignment 1. Due Jan. 31, 2017

Assignment 1. Due Jan. 31, 2017 Assignment 1 Due Jan. 31, 2017 Show all work and turn in answers on separate pages, not on these pages. Circle your final answers for clarity. Be sure to show/explain all of your reasoning and that your

More information

Mathematics - High School Algebra II

Mathematics - High School Algebra II Mathematics - High School Algebra II All West Virginia teachers are responsible for classroom instruction that integrates content standards and mathematical habits of mind. Students in this course will

More information

Which of the following planets are all made up of gas? When a planets orbit around the Sun looks like an oval, it s called a(n)

Which of the following planets are all made up of gas? When a planets orbit around the Sun looks like an oval, it s called a(n) When a planets orbit around the Sun looks like an oval, it s called a(n) - ellipse - circle - axis - rotation Which of the following planets are all made up of gas? - Venus, Mars, Saturn and Pluto - Jupiter,

More information

Evolutionary computation

Evolutionary computation Evolutionary computation Andrea Roli andrea.roli@unibo.it DEIS Alma Mater Studiorum Università di Bologna Evolutionary computation p. 1 Evolutionary Computation Evolutionary computation p. 2 Evolutionary

More information

Chapter 2. Altitude Measurement

Chapter 2. Altitude Measurement Chapter Altitude Measurement Although altitudes and zenith distances are equally suitable for navigational calculations, most formulas are traditionally based upon altitudes which are easily accessible

More information

Lecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University

Lecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University Lecture 18: Kernels Risk and Loss Support Vector Regression Aykut Erdem December 2016 Hacettepe University Administrative We will have a make-up lecture on next Saturday December 24, 2016 Presentations

More information

Patterns in the Solar System (Chapter 18)

Patterns in the Solar System (Chapter 18) GEOLOGY 306 Laboratory Instructor: TERRY J. BOROUGHS NAME: Patterns in the Solar System (Chapter 18) For this assignment you will require: a calculator, colored pencils, a metric ruler, and meter stick.

More information

Data Mining Part 5. Prediction

Data Mining Part 5. Prediction Data Mining Part 5. Prediction 5.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline How the Brain Works Artificial Neural Networks Simple Computing Elements Feed-Forward Networks Perceptrons (Single-layer,

More information

Putting Earth In Its Place

Putting Earth In Its Place Teacher Instructions Overview: During this activity, students build a model of our Solar System to gain insight into the relative sizes and distances involved. ives: The student will: create a scale model

More information

TAKEN FROM HORIZONS 7TH EDITION CHAPTER 1 TUTORIAL QUIZ

TAKEN FROM HORIZONS 7TH EDITION CHAPTER 1 TUTORIAL QUIZ TAKEN FROM HORIZONS 7TH EDITION CHAPTER 1 TUTORIAL QUIZ 1. If the solar system is scaled down so that the Sun is represented by a basketball, a. a ping-pong ball located 500 feet away would properly represent

More information

Indirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina

Indirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina Indirect Rule Learning: Support Vector Machines Indirect learning: loss optimization It doesn t estimate the prediction rule f (x) directly, since most loss functions do not have explicit optimizers. Indirection

More information

Lecture 4: Kepler and Galileo. Astronomy 111 Wednesday September 6, 2017

Lecture 4: Kepler and Galileo. Astronomy 111 Wednesday September 6, 2017 Lecture 4: Kepler and Galileo Astronomy 111 Wednesday September 6, 2017 Reminders Online homework #2 due Monday at 3pm Johannes Kepler (1571-1630): German Was Tycho s assistant Used Tycho s data to discover

More information

Kernel Methods & Support Vector Machines

Kernel Methods & Support Vector Machines Kernel Methods & Support Vector Machines Mahdi pakdaman Naeini PhD Candidate, University of Tehran Senior Researcher, TOSAN Intelligent Data Miners Outline Motivation Introduction to pattern recognition

More information