12 Symbolic Regression
|
|
- Kristina Craig
- 5 years ago
- Views:
Transcription
1 Computational Geosciences PART I METHODS 12 Symbolic Regression 12-1 Introduction Symbolic regression is the process of determining the symbolic function which describes a data set effectively developing an analytic model which summarizes the data and is useful for predicting response behaviors as well as facilitating human insight and understanding. The symbolic regression approach adopted herein is based upon genetic programming wherein a population of functions are allowed to breed and mutate with the genetic propagation into subsequent generations based upon a survival-of-the-fittest criteria. Amazingly, this works and, although computationally intensive, summary solutions may be reasonably discovered using current laptop and desktop computers. It should be noted that symbolic regression is not a silver bullet and is, in fact, complementary to its nonlinear analysis brethren of neural networks and statistical learning theory (support vector machines) as well as classical linear analysis and data mining tools and techniques. In all cases, an understanding of the problem, the data variables, the data quality, and definitions of success is a critical aspect since context free analysis can easily lead to confidently wrong answers i.e., is very dangerous. The symbolic regression capability provides a complete and flexible foundation for evolving multivariate datadriven models. The decision to develop this symbolic regression capability within the Mathematica environment was validated during the development of this package as the ability to seamlessly blend symbolic and numeric computing has enabled new features which would have been onerous to implement in a strictly procedural and numeric environment. Additional benefits include facilitating analysis documentation Symbolic Regression (SR) Method Unlike traditional linear and nonlinear regression methods that fit parameters to an expression (equation/relation) of a given form, symbolic regression (SR) simultaneously searches for both the parameters as well as the form of expression. Although the discipline of SR has matured significantly in the last few years (e.g., Davidson et al. 2003), its applications to geodesy are very rare, exemplified only in the works of Wu et al. (2007, 2008) who employed it for transforming GPS coordinates into two-dimensional coordinates. In a more recent work, Wu and Su (2013) developed a lattice-based clustering method and integrated it with a genetic programming to build a better regression model for coordinate transformation. In natural and technical sciences, and in finance, however, the method has been applied for quite a long time efficiently, (e.g., Santini and Tettamanzi (2001); Banks (2002); Kwon and Moon (2005); Babu and Karthik (2007); Schmidt and Lipson (2009); Langdon and Gustafson (2010); and Garg and Tai (2011)). In hydrological sciences, for example, Parasuraman et al. (2007) applied it to model the dynamics of evapotranspiration, where their results performed better than the traditional Penman-Montieith method, and were comparable to those of artificial neural network (ANN). The proposed method could be of use to geodesy, where regression analysis and functional approximation are often handled. For instance, SR could be used for gravimetric corrections where they have traditionally been carried out using a wide variety of parametric and non-parametric surfaces such as polynomial models (e.g., Fotopoulos and Sideris 2005), spline interpolation
2 2 SymbolicRegression_12.nb Fotopoulos and Sideris 2005), spline interpolation (e.g., Featherstone 2000), least squares collocation (LSC) (e.g., Iliffe et al. 2003), Kriging (e.g., Nahavandchi and Soltanpour 2004), combined least squares adjustments (e.g., Fotopoulos 2005) and thin plate spline (TPS) surface of solving the problem via finite elements method (e.g., Zaletnyik et al. 2007). Applying soft computing technique, Kavzoglu and Saka (2005) and Lao-Sheng Lin (2007) employed artificial neural network (ANN) for approximating the GPS/leveling geoidal heights instead of the corrector surface itself. Zaletnyik et al. (2007) also used ANN but with radial bases activation functions (RBF) and regularization in the training phase. Soltanpour et al. (2006) used second generation wavelets to approximate corrector surface directly. Another soft computing technique represented by the support vector machines (SVM) was employed by Zaletnyik et al. (2008). Some of these models are global types of regression, (e.g., linear parametric models and ANN), while some are interpolating type of local models, (e.g., thin plate spline and SVM). Generally speaking, local methods are more precise than global on the training set, but their complexity is very high since they involve all of the measured data points in the model structure. In addition, they are less effective on the validation set since the overlearning affect the training set. The foregoing discussions support the need for the proposed SR method, particularly in geodesy where it is of utmost need but rarely used. SR can be considered as a broad generalization of the class of Generalized Linear Models (GLM). GLM is a linear combination of n basis functions β i, i = 1,2,...n with a dependent variable y and an independent vector variable x. n y (x) = c 0 + c i β i (x) + ϵ i=1 where ci are the coefficients and ε the error term. SR will search for a set of basic functions (building blocks) and coefficients (weights) in order to minimize the error ε in case of given y and x. The standard basic functions are constant, addition, subtraction, multiplication, division, sine, cosine tangent, exponential, power, square root, etc. To select the optimal set of basic functions, Koza (1992) suggested employment of genetic programming (GP). GP is a biologically inspired machine learning method that evolves computer programs to perform a task. In order to carry out genetic programming, the individuals (competing functions) should be represented by a binary tree. In standard GP, the leaves of the binary tree are called terminal nodes represented by variables and constants, while the other nodes, the so called non-terminal nodes are represented by functions. Let us see a simple example. Consider Its binary tree representation can be seen in Fig 12.1, β i (x) = x 1 x x 3 Fig The binary tree representation of a basic function βi In this example, there are three variables (x1, x2, x3), two constants (1, 2), and three elementary functions (plus, times, rational). The binary tree of y(x) can be built up from such trees as sub-trees. Mathematically
3 SymbolicRegression_12.nb 3 y(x) = c 0 + c 1 tree 1 + c 2 tree GP randomly generates a population of individuals ( y k (x), k = 1, 2...n ) represented by tree structures to find the best performing trees. There are two important features of the function represented by a binary tree: complexity and fitness. We define complexity as the number of nodes in a binary tree needed to represent the function, Morales (2004). The fitness qualifies how good a model (y = y(x)) is. Basically, there are two types of measures used in SR; the root mean squared error (RMSE) and the Rsquare. The later returns the square of the Pearson product moment correlation coefficient (R) describing the correlation between the predicted values and the target values, (e.g., Ferreira (2006)). Then, the goodness of the model, the fitness function can be defined as, f = 1 1+ RMSE or f = R2 where 0 f 1 GP tries to minimize this error to improve the fitness of the population consisting of individuals (competing functions) from generation to generation by mutation and cross-over procedure. Mutation is an eligible random change in the structure of the binary tree, which is applied to a randomly chosen sub-tree in the individual. This sub-tree is removed from the individual and replaced by a new randomly created sub-tree. This operation leads to a slightly (or even substantially) different basic function. Let us consider the binary tree on Fig. 12.2a, where the sub-tree of y 2 is replaced by y+x 2. Then, the mutated binary tree can be seen in Fig. 12.2b Fig Binary tree representations of the mutation. y 2 in a) is replaced by y+x 2 in b) The operation cross-over representing sexuality can accelerate the improvement of the fitness of a function more effectively than mutation alone can do. It is a random combination of two different basic functions (parents), based on their fitness, in order to create a new generation of functions, more fitter than the original functions. To carry out cross-over, crossing points (nonterminal nodes) in tree of both parents should be randomly selected, as it can be seen in Fig. 3. Then subtrees belonging to these nodes will be exchanged creating offsprings. Let us consider the parents before the cross-over. The first parent (x-y)/3, with its crossing point (x) is shown in Fig. 12.3a, the second parent 3x+y 2 /5 with its crossing point (see the sub-tree of y 2 ) is presented in Fig b.
4 4 SymbolicRegression_12.nb Fig The parents before the cross over The children produced by the cross-over are given on Fig The first child (y 2 -y)/3 is shown in Fig a while the second child (16/5)x is shown in Fig. 12.4b. Fig The children produced by cross-over The generalization of GP was invented by Cramer (1985) and further developed by Koza (1992). GP is a class of evolutionary algorithms working on executable tree structures (parse trees). Koza (1992) showed that GP is capable of doing symbolic regression (or function identification) by generating mathematical expressions approximating a given sample set very closely or in some cases even perfectly. Therefore, GP finds the entire approximation model and its (numerical) parameters simultaneously. An important goal in symbolic regression is to get a solution, which is numerically robust and does not require high levels of complexity to give accurate output values for given input parameters. Small mean errors may lead to wrong assumptions about the real quality of the found expressions. To be on the safer side, a worst case absolute error should be determined. Sometimes, alternating the worst case absolute error and RMSE or R 2 as target to be minimized during GP process is the best strategy. Complexity and fitness are conflicting features leading to a multiobjective problem, (e.g., Smits and Kotanchek (2004)). Useful expression is both predictive and parsimonious. Some expressions may be more accurate but overfit the data, whereas others may be more parsimonious but oversimplify. The prediction error versus complexity or 1-fitness versus complexity Pareto front represents the optimal solutions as they vary over expression complexity and maximum prediction error, (e.g., Paláncz and Awange (2012)). As Fig shows, functions representing the Pareto front have the following features: In case of fixed complexity, there is no such a solution (function), which could provide less error as the Pareto solution, Reversely, in case of fixed error, there is no such a solution (function), which would
5 SymbolicRegression_12.nb 5 Reversely, in case of fixed error, there is no such a solution (function), which would have smaller complexity than the Pareto solution. Fig The Pareto - front In order to illustrate the method, let us start with a simple, didactic example: Kepler s Third Law. This problem was considered and rediscovered by Langley et al. (1987) by a family of heuristic techniques as well as by Koza (1992) by genetic programming Didactic example- Kepler third law Kepler s problem The third law of Kepler states: "The square of the orbital period of a planet is directly proportional to the cube of the semi-major axis of its orbit (average distance from the Sun)." P 2 a 3 where P is the orbital period of the planet and a is the semi-major axis of the orbit. For example, suppose planet A is 4 times as far from the Sun as planet B. Then planet A must traverse 4 times the distance of planet B each orbit, and moreover it turns out that planet A travels at half the speed of planet B, in order to maintain equilibrium with the reduced gravitational centripetal force due to being 4 times further from the Sun. In total it takes 4 2 = 8 times as long for planet A to travel an orbit, in agreement with the law (8 2 = 4 3 ). The third law currently receives additional attention as it can be used to estimate the distance from an exoplanet to its central star, and help to decide if this distance is inside the habitable zone of that star. The exact relation, which is the same for both elliptical and circular orbits, is given by the equation This third law used to be known as the harmonic law, because Kepler enunciated it in a laborious attempt to determine what he viewed as the "music of the spheres" according to precise laws, and express it in terms of musical notation. His result based on the Rudolphine table containing the observations of Tycho Brache 1605, see Table 12.1 Table 12.1 Normalized observation planetary data Planet where a is give in units of Earth s semi-major axis. Period Semimajor axis P [yr] a Mercury Venus Earth Mars Jupiter Saturn Uranus Neptune Pluto
6 6 SymbolicRegression_12.nb Let us assume that Kepler could have employed one of the function approximation techniques like polynomial regression, artificial neural networks, support vector machine, thin plate spline. Could he find this simple relation with these sophisticated methods? Polynomial regression Let us employ a fifth order algebraic polynomial. Then the resulted polynomial is P = a a a a a Neural network Employing a single-layer feedforward network (Kecman 2001) with two nodes where sigmoid activation function is implemented, we obtained P = a a a Support vector machine Employing wavelet - kernel (Palancz et al. 2005) the result is the following, P = (0.39-a)2 Cos[ ( a)] (0.72-a)2 Cos[ ( a)] (1.-a)2 Cos[ (1. - a)] (1.52-a)2 Cos[ ( a)] (5.2-a)2 Cos[ (5.2 - a)] (9.54-a)2 Cos[ ( a)] (19.19-a)2 Cos[ ( a)] (30.06-a)2 Cos[ ( a)] (39.53-a)2 Cos[ ( a)] Thin plate - spline Employing polyharmonic interpolation (Fasshauer 2007) the approximation is, P = φ φ φ φ φ φ φ φ φ 9 ] where φ i = r 2 log (r) and alternatively r = a - a i for a a i r = 0 for a = a i As we can see, in this way one can hardly find a simple relation, although all of these methods give quite satisfactory solution for the problem, see Table Table 12.2 Errors of the different methods of approximation of the observed planetary data Model 1 - R 2 Standard Deviation Polynomial Regression Neural Network Support Vector Machine Thin plate spline interpolation where R stands for the Root Square Mean Error (RMSE). Remarks. 1 fifth order algebraic polynomial. 2 feedforward neural network with two nodes and linear tail
7 SymbolicRegression_12.nb 7. feedforward neural network with two nodes and linear tail. 3 with Gaussian- kernel. 4 basis function: x 2 Log( x ) Now let us see, how can we solve this problem with symbolic regression Symbolic regression Fig shows the Pareto - front of the generated models via DataModeler. The points represent the generated models. The red points stand for the models belong to the Pareto-front. Fig The Pareto-front (red points) and the evaluated models in case of the Kepler s problem In Table 12.3 some of the models of the Pareto front can be seen. Table Model selection riport Model Complexity 1 - R 2 Function x x ( x) x x x a 3/ x 3/2 + 2 x x 3/2 + 2 x x x 3/2 + 2 x * x x 3/ x x + 5 x x 3/ a +a x x x + x 12+2 x+x 2 It goes without saying that our candidate is the 4th model, since it has a small error and at the same time its complexity is low. In practice, it is useful to carry out an additional nonlinear regression with the candidate model in order to refine the value of its parameters. This means, we use symbolic regression to find the best model structure and then its
8 8 SymbolicRegression_12.nb the value of its parameters. This means, we use symbolic regression to find the best model structure and then its parameters will be improved via standard nonlinear regression, in order to avoid a long run of the symbolic regression procedure. Now let us compute the parameters p i of the candidate model, P = p 0 + p 1 a p 2 Table 12.4 indicates that the estimation of the parameter p 0 is unreliable, since its confidence interval is very wide and the statistic P-Value is high, For this model 1 - R 2 = Table The statistics of the estimated parameters via nonlinear regression Parameter Estimated Value Confidencia Interval P- Value p { , } p { , } p { , } Now let us repeat the regression for the model P = p 1 a p2 we get the following values of the two parameters, see Table 12.5 Table The statistics of the estimated parameters via nonlinear regression Parameter Estimated Value Confidence Interval P- Value p { , p { , From practical point of view one considers p 1 = 1 and a = p 2 = 1.5. Although the error of this simplified model is somewhat higher, since 1 - R 2 = < Now let us see, how we compute the symbolic result via Mathematica Application Mathematica Mathematica itself has not a built-in function for symbolic regression, however a third party product DataModeler package developed by Evolved Analytics ( can be employed. Let us load the package << DataModeler` First we should provide the input - output corresponding data, see Table 12.1 samplepts = SetPrecision[{0.39, 0.72, 1., 1.52, 5.2, 9.54, 19.19, 30.06, 39.53}, 10]; observedresponse = SetPrecision[{0.24, 0.61, 1., 1.88, 11.86, 29.46, 84.01, , }, 10]; Then we compute the Pareto front. The computation is carried out parallel in order to decrease the running time.
9 SymbolicRegression_12.nb 9 Timing@ ParetoFrontLogPlot[ quickmodels1 = SymbolicRegression[ samplepts, observedresponse, DataVariables {x}, AlignModel True, RobustModels True, TimeConstraint 50, FitnessPrecision 10 ] ] , Fig The Pareto-front (red points) computed by Mathematica. The models represented by the Pareto front are in the following table. Here we listed out models which have less error than err and their complexity is lower than comp err = 0.01; comp = 150; ModelSelectionReport[quickModels1, QualityBox -> {comp, err}] Model Selection Report Complexity 1-R 2 Function x ( x) x x x x x x 3/ x 3/ x x 3/ x (10 +x) x x x 3/ x x x 3/ x x 3/ x (4.93+x) x x 3/ x x x 3/2 - (7+x) x2 (11.82+x) x x 3/ x 2 (( x) ( x)) x x x 5/ x +x x x x x + x 2 Table The some good models representing the Pareto- front Kepler considered the 5th model, and leaved out the small constant: x 3! In order to use the selected model for
10 10 SymbolicRegression_12.nb further computations we should do the following operations Here we select the model having less error than and complexity is not greater than 30! kp = SelectModels[quickModels1, QualityBox -> {30, }]; This already a Mathematica expression kpm = ModelPhenotype[kp] // First x 3 and, since u = kpm[[2]] x 3 then the Kepler model is KP = Round Coefficient u, x 3 u[[2]] x Teaching Example Let us try to approximate the following function on the basis of some discrete points. The function is z[{x_, y_}] := (x 2 - y 2 ) Sin[0.5 x] p1 = Plot3D[z[{x, y}], {x, -10, 10}, {y, -10, 10}, PlotPoints {30, 30}, BoxRatios {1, 1, 0.5}] Fig The function to be approximated Let the discrete data (x i, y i, f (x i, y i ) data = Flatten[Table[N[{i - 10, j - 10, z[{i - 10, j - 10}]}], {i, 0, 30, 4}, {j, 0, 30, 4}], 1];
11 SymbolicRegression_12.nb 11 Show[{p1, ListPointPlot3D[data, PlotStyle Directive[Blue, PointSize[0.015]], BoxRatios {1, 1, 0.5}]}] Fig The given points We employ symbolic regression, since we should like to have an explicit expression for discribing the surface. Now we separate the data: the inputs (x i, y i ) and the outputs f (x i, y i ) inputs = Transpose[Take[Transpose[data], {1, 2}]] {{-10., -10.}, {-10., -6.}, {-10., -2.}, {-10., 2.}, {-10., 6.}, {-10., 10.}, {-10., 14.}, {-10., 18.}, {-6., -10.}, {-6., -6.}, {-6., -2.}, {-6., 2.}, {-6., 6.}, {-6., 10.}, {-6., 14.}, {-6., 18.}, {-2., -10.}, {-2., -6.}, {-2., -2.}, {-2., 2.}, {-2., 6.}, {-2., 10.}, {-2., 14.}, {-2., 18.}, {2., -10.}, {2., -6.}, {2., -2.}, {2., 2.}, {2., 6.}, {2., 10.}, {2., 14.}, {2., 18.}, {6., -10.}, {6., -6.}, {6., -2.}, {6., 2.}, {6., 6.}, {6., 10.}, {6., 14.}, {6., 18.}, {10., -10.}, {10., -6.}, {10., -2.}, {10., 2.}, {10., 6.}, {10., 10.}, {10., 14.}, {10., 18.}, {14., -10.}, {14., -6.}, {14., -2.}, {14., 2.}, {14., 6.}, {14., 10.}, {14., 14.}, {14., 18.}, {18., -10.}, {18., -6.}, {18., -2.}, {18., 2.}, {18., 6.}, {18., 10.}, {18., 14.}, {18., 18.}} outputs = Transpose[Take[Transpose[data], {3, 3}]] // Flatten {0., , , , , 0., , , , 0., , , 0., , , , , , 0., 0., , , , , , , 0., 0., , , , , , 0., , , 0., , , , 0., , , , , 0., , , , , , , , , 0., , , , , , , , , 0.} Then we compute the Pareto front, which now is not convex, see Fig ,
12 12 SymbolicRegression_12.nb ParetoFrontLogPlot[ model = SymbolicRegression[inputs, outputs, AlignModel True, RobustModels True, FitnessPrecision 10, TimeConstraint 500 ] ] , Fig The Pareto front now is not convex The selected models can be seen in Table 12.7 ModelSelectionReport[model, QualityBox {500, 0.01}] Model Selection Report Complexity 1-R 2 Function x x x x x x x x 1 +x x x 1 x x 1 3 x x 1 5 x x 1 3 x 1 2 x x x x (7.78- x 1) x x x 5 1 x x x 1 x x 3 1 x x 1 5 x x 1 x 1 2 x x x x 1 3 x x x x x x x x x 1 + x x x x 1 x x 1 3 x x 1 5 x x 1 7 x x 1 3 x 1 2 x Model selection Table 12.7 The eminent models representing the Pareto- front Now we select the model with complexity 262 and error kp = SelectModels[model, QualityBox -> {262, 0.005}]; kpm = ModelPhenotype[kp] // Flatten x x x x x x x x1 + x1 2 + x x1 x x1 3 x x1 5 x x1 3 x1 2 x2 2
13 SymbolicRegression_12.nb 13 p2 = Plot3D[kpm, {x1, -10, 10}, {x2, -10, 10}, PlotPoints {30, 30}, BoxRatios {1, 1, 0.5}] Fig The selected model Now let us see, how the data points fit to the surface of the model, see Fig Show[{p2, ListPointPlot3D[data, PlotStyle Directive[Blue, PointSize[0.015]], BoxRatios {1, 1, 0.5}]}] Fig The selected model and the data points The fitting looks quite good. However, comparing the surfaces of the original model and the selected one, we can see the differences, see Fig left side,
14 14 SymbolicRegression_12.nb Show[{p2, p1}] Fig The surfaces of the original and the selected model In the data points, we have good fitting but where were no points, the fitting is less efficient.this fact shows that we should not select the model with the smallest error this would mean in general overlearning. Let us select the best model kp = SelectModels[model, QualityBox -> {302, 0.000}]; kpm = ModelPhenotype[kp] // Flatten x x x x x x x x1 + x1 2 + x x x1 x x1 3 x x1 5 x x1 7 x x1 3 x1 2 x2 2 p2 = Plot3D[kpm, {x1, -10, 10}, {x2, -10, 10}, PlotPoints {30, 30}, BoxRatios {1, 1, 0.5}]
15 SymbolicRegression_12.nb 15 Show[{p2, ListPointPlot3D[data, Fig The best selected model PlotStyle Directive[Blue, PointSize[0.015]], BoxRatios {1, 1, 0.5}]}] Show[{p2, p1}] Fig The best selected model and the data points 12-7 Extention of function space Fig The surfaces of the original and the best selected model In the basic option, the choice of the type basic functions, which can be employed by the method is quite poor. To gettin better results these function space can be extended, see Help DataModeler SymbolicRegression FunctionPatterns The Pareto - front can be seen in Fig Table? shows that we have got back the original model!
16 16 SymbolicRegression_12.nb ParetoFrontLogPlot[ model = SymbolicRegression[inputs, outputs, AlignModel True, RobustModels True, FitnessPrecision 10, FunctionPatterns BuildFunctionPatterns["ExtendedMath", Sin], MemoryLimit 5 "GB", TimeConstraint 500] ] , Fig The Pareto front in case of extended function space ModelSelectionReport[model, QualityBox {500, }] Model Selection Report Complexity 1-R 2 Function Sin x1 Table 12.8 The bets model of the Pareto- front kp = SelectModels[model, QualityBox -> {52, 0.2}]; kpm = ModelPhenotype[kp] // Flatten x1 2 + x2 2 Sin x x12 +x 2 2
17 SymbolicRegression_12.nb 17 p2 = Plot3D[kpm, {x1, -10, 10}, {x2, -10, 10}, PlotPoints {30, 30}, BoxRatios {1, 1, 0.5}] Show[{p2, ListPointPlot3D[data, Fig The selected model PlotStyle Directive[Blue, PointSize[0.015]], BoxRatios {1, 1, 0.5}]}] Now the two surfaces fit perfectly! Fig The selected model with the data points
18 18 SymbolicRegression_12.nb Show[{p2, p1}] 12-8 Application example Geometric transformation Problem definition Fig The surface of the original and the selected models Transformation of coordinates is important in computer vision, photogrammetry as well as in geodesy. In this example we consider some standard 2D transformations as similarity, affine and projective transformations between the coordinates of the fiducial marks on the comparator plate and that of the corresponding points on the reseau plate. The 16 observed (x, y) and calibrated reseau (X, Y) coordinates are given in Table 12.9 Similarity transformation Table 12.9 The coordinates of the corresponding observation points on the comparator and reseau planes Points x [mm] [mm] [mm] [mm] The advantage of this model is that it is linear in the coordinates as well as in the parameters. Consequently an iterative solution is not required and the inverse transformation is easy. This transformation can be parametrized in the following form, y X Y
19 SymbolicRegression_12.nb 19 x y = a b -b a X Y + c d For each observed ith point the following pair of observations equation can be written for the residuals (r xi, r yi ), x i y i 1 0 y i -x i 0 1 a b c d + -x i -y i = r x i r yi A minimum of two fiducial marks or reseau crosses are required for a unique solution. Having more observation points linear least square can be directly applied. In our example the result can be seen in Table Table The estimated parameters of the similarity transformation model Parameter Value a b c d The norm of the residual vectors in the observation points are shown in Table Affine transformation Table The norm of the residual vectors at the observation points in case of similarity transformation model Points r xy [mm] This is also a linear model although it needs six parameters, therefore minimum 3 fiducial marks or reseau crosses are required for a unique solution. Iterative solution is not required and the inverse transformation is easy. This model can be parametrized in the following form x y = a d b e X Y + c f For each observed ith point the following pair of observations equation can be written for the residuals (r xi, r yi ), y i x i x i 1 y i a b c d e f + -x i -y i = r x i r yi Having more than 3 observation points linear least square can be directly applied. In our example the result can be seen in Table 12.12
20 20 SymbolicRegression_12.nb Table The estimated parameters of the affine transformation model Parameters Values a b c d e f The norm of the residual vectors in the observation points are shown in Table Projective transformation Table The norm of the residual vectors at the observation points in case of affine transformation model Points r xy [mm] The projective transformation can be expressed by the following two equations, and x = a 1 X + a 2 Y + a 3 c 1 X + c 2 Y + 1 y = b 1 X + b 2 Y + a 3 c 1 X + c 2 Y + 1 These equations are a special case of the collinearity condition for mapping of 2 D points from one plane onto another. There are 8 unknown parameters, so if only four fiducial marks are available, the solution is not unique. This is perhaps the main reason why it is not frequently used. However, it is relevant for systems that have been retro-fitted with a reseau plate, such as the Hasselblad camera used to acquire the assignment imagery. The least square solution is nonlinear due to the rational nature of the functions. However, an approximate linear solution can be implemented if both sides of the equations are multiplied by the denominator and partial derivates with respect to the observables (as in the combined adjutment model) are ignored. The inverse of the transformation can be also computed by solving the equation system in symbolic form for X and Y, we obtain
21 SymbolicRegression_12.nb 21 and X = (a 2 (-y + a 3 ) + (x - a 3 ) b 2 + (-x + y) a 3 c 2 )/(-x b 2 c 1 + a 2 (-b 1 + y c 1 ) + x b 1 c 2 + a 1 (b 2 - y c 2 )) Y = (a 1 (y - a 3 ) + (-x + a 3 ) b 1 + (x - y) a 3 c 1 ) /(-x b 2 c 1 + a 2 (-b 1 + y c 1 ) + x b 1 c 2 + a 1 (b 2 - y c 2 )) Having more than 4 observation points nonlinear least square can be applied. In our example the result can be seen in Table Table The estimated parameters of the projective transformation model Parameters Values a a a b b b c 1 0 c The norm of the residual vectors in the observation points are shown in Table Table The norm of the residual vectors at the observation points in case of projective transformation model Application of symbolic regression Points r xy [mm] {{-110, -110}, {-40, -110}, {40, -110}, {110, -110}, {110, -40}, {40, -40}, {-40, -40}, {-100, -40}, {-110, 40}, {-40, 40}, {40, 40}, {110, 40}, {110, 110}, {40, 110}, {-40, 110}, {-110, 110}} outputx = { , , , , , , , , , , , , , , , }; outputy = {-107.4, , , , , , , , , , , , , , , };
22 22 SymbolicRegression_12.nb inputxy = {{-110, -110}, {-40, -110}, {40, -110}, {110, -110}, {110, -40}, {40, -40}, {-40, -40}, {-100, -40}, {-110, 40}, {-40, 40}, {40, 40}, {110, 40}, {110, 110}, {40, 110}, {-40, 110}, {-110, 110}}; Per definition neither the model nor the parameters of the model are not known. However in order to have a chance for computing the inverse of the transformation trigonometrical functions are excluded from the functionset. Now, we create linear and nolinear models, but linear models will be optimized, Timing@ ParetoFrontLogPlot[ ] model = SymbolicRegression[inputXY, outputx, OptimizeLinearModel True, ] TimeConstraint , Fig The Pareto fron in case of Optimal Linear model for x = (X, Y)
23 SymbolicRegression_12.nb 23 ModelSelectionReport model, QualityBox 250, 10-8 Page 1 Page 2 Model Selection Report Complexity 1-R 2 Function x x x x x x x 1 x x x x 1 x x x x 1 x x x1 x1 x1 x2 x x x 1 x x x x 1 x x x x1 x2 x x x 1 x x x x 1 x x x x x1 x2 x2 2 x x x 1 x x 2 2 x x x 1 x x x 2 2 x x x 1 x x x 2 2 x x x x x 1 x (3 -x x1 2) x x x x x x 1 x x (7 - x 2 ) x 3 2 x x x x x x2 x1 x1 2 2 x x x (x 1 - x 2) 2 + Table Linear and nonlinear models for x = (X, Y) The best linear model among the linear and nonlinear models is kp = SelectModels model, QualityBox -> 30, ; kpm = ModelPhenotype[kp] // Flatten x x x1 x2 Now, only linear models will be created
24 24 SymbolicRegression_12.nb ParetoFrontLogPlot[model = CreateLinearModel[ inputxy, outputx, PolynomialOrder 4, TimeConstraint 10]] , Fig The Pareto fron in case of direct linear modelling for x = (X, Y) ModelSelectionReport model, QualityBox 230, 10-8 Model Selection Report Complexity 1-R 2 Function x x 1 x x 2 1 x x 3 1 x x x x x x x 1 x x 2 1 x x x 1 x x 4 2 Table The linear model for x = (X, Y) kp = SelectModels model, QualityBox -> 205, 10-9 ; kpm = ModelPhenotype[kp] // Flatten x x x x x x1 x x1 2 x x1 3 x x x1 x x1 2 x x x1 x x2 4 Surprisingly, the best linear model is among the linear and nonlinear models is better than direct linear model! Now we select both xv1 = ` `X `Y `*^-7X Y X Y X Y and
25 SymbolicRegression_12.nb 25 xv2 = `*^ `x `*^-6x `*^-9x `*^-10x `x `*^-6x1 x `*^-10x1 2 x `*^-11x1 3 x `x `*^-9x1 x `*^-13x1 2 x `*^-9x `*^-11x1 x `*^-7x2 4 /. {x1 X, x2 Y} X X X X Y X Y X 2 Y X 3 Y Y X Y X 2 Y Y X Y Y 4 Similarly let us compute the y (X, Y) function, too. Timing@ ParetoFrontLogPlot[ model = SymbolicRegression[inputXY, outputy, OptimizeLinearModel True, TimeConstraint 150 ] ] , Fig The Pareto fron in case of Optimal Linear model for y = (X, Y)
26 26 SymbolicRegression_12.nb ModelSelectionReport model, QualityBox 250, 10-8 Page 1 Page 2 Model Selection Report Complexity 1-R 2 Function x x x x x x x 1 x x x x x x x x x x 1 x x x1 x x x x x x x x2 x x x x x x 1 2 x / ( x 1 + x 2) x x x x x / ( x 1 +x 2) x x x x x 1 2 x / ( x 1 + x 2) x x x x x 2 1 x x x 2 x x x x x 1 x x 1 +x / ( x 1 + x 2 ) x x x x 1 x x1 +x / ( x 1 + x 2) x x x x x x x (4+ x 1 +x 2) x x /( x 1 -x 2) x / ( x 1 +x 2) + x2 The best linear model is Table Linear and nonlinear models for y = (X, Y) kp = SelectModels model, QualityBox -> 42, ; kpm = ModelPhenotype[kp] // Flatten x x x1 x x2 2 Now, only linear models will be created
27 SymbolicRegression_12.nb 27 ParetoFrontLogPlot[model = CreateLinearModel[ inputxy, outputy, PolynomialOrder 4, TimeConstraint 10]] , Fig The Pareto fron in case of direct linear modelling for y = (X, Y) ModelSelectionReport model, QualityBox 230, 10-8 Model Selection Report Complexity 1-R 2 Function x x 1 x x 2 1 x x 3 1 x x x x x x x 1 x x 2 1 x x x 1 x x 4 2 Table The linear model for y = (X, Y) kp = SelectModels model, QualityBox -> 205, ; kpm = ModelPhenotype[kp] // Flatten x x x x x x1 x x1 2 x x1 3 x x x1 x x1 2 x x x1 x x2 4 Now the best linear model is among the linear and nonlinear models has smaller complexity but higher error level than the direct linear model! Therefore we select both yv1 = ` `x `x `*^-7x1 x `*^-7x2 2 /. {x1 X, x2 Y} X Y X Y Y 2
28 28 SymbolicRegression_12.nb yv2 = `*^ `x `x `*^-9x `*^-10x `x `*^-8x1 x `*^-9x1 2 x `*^-12x1 3 x `x `*^-9x1 x `*^-11x1 2 x `*^-9x `*^-11x1 x `*^-8x2 4 /. {x1 X, x2 Y} X X X X Y X Y X 2 Y X 3 Y Y X Y X 2 Y Y X Y Y 4 Now we compute the sum of residuals rxy1 = MapThread[Norm[{(#2 - xv1), (#3 - yv1)}] /. {X #1[[1]], Y -> #1[[2]]} &, {inputxy, outputx, outputy}]; TableForm[rxy1] Apply[Plus, rxy1] RMSE % / rxy2 = MapThread[Norm[{(#2 - xv2), (#3 - yv2)}] /. {X #1[[1]], Y -> #1[[2]]} &, {inputxy, outputx, outputy}]; TableForm[rxy2] Apply[Plus, rxy2]
29 SymbolicRegression_12.nb 29 RMSE % / So the higher order pair give better result. Table The norm of the residual vectors at the observation points in case of the nonlinear transformation model obtained by SR Points r xy [mm] The Fig illustrates how this nonlinear transformation warpes (deformates) the original (X, Y) plane into the (x, y) plane
30 30 SymbolicRegression_12.nb Fig The original plane and below the plane after warping via the nonlinear transformation Comparison of the different transformation models Max[rxy2] StandardDeviation[rxy2] It is not surprising that transformation models having more parameters and nonlinear form can provide less error than others. Let us summarize some statistic values of the different methods in Table Transformation model Table Some statistical values of the different transformation models Max of Absolute Errors Standard Deviation of the Absolute Errors RMSE of the residual Errors [mm] [mm] [mm] Similarity Affine Projective Symbolic Regression Both SR models are better than the other models. Although the nonlinear transformation obtained by the symbolic regression gives the best fitting, to compute the inverse of the transformation is not an easy task. However it can be solved via Groebner basis.
Physics Unit 7: Circular Motion, Universal Gravitation, and Satellite Orbits. Planetary Motion
Physics Unit 7: Circular Motion, Universal Gravitation, and Satellite Orbits Planetary Motion Geocentric Models --Many people prior to the 1500 s viewed the! Earth and the solar system using a! geocentric
More informationKoza s Algorithm. Choose a set of possible functions and terminals for the program.
Step 1 Koza s Algorithm Choose a set of possible functions and terminals for the program. You don t know ahead of time which functions and terminals will be needed. User needs to make intelligent choices
More informationKepler's constant in Earth-Moon System is gradually. decreasing
Kepler's constant in Earth-Moon System is gradually decreasing Hu Wang Email:vortex.wh@hotmail.com Abstract Studies on nautiloids, coral fossils, rotation of the Earth, and Earth-Moon distance variation
More informationInvestigating the Solar System
Investigating the Solar System This Workbook belongs to: Our Local Star: The Sun Location in The Solar System Interesting Facts 1. 2. 3. 4. Name of Star: THE SUN 5. Draw and Color your own Sun in the blank
More informationObservational Astronomy - Lecture 4 Orbits, Motions, Kepler s and Newton s Laws
Observational Astronomy - Lecture 4 Orbits, Motions, Kepler s and Newton s Laws Craig Lage New York University - Department of Physics craig.lage@nyu.edu February 24, 2014 1 / 21 Tycho Brahe s Equatorial
More informationCSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18
CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$
More informationKernel-based Approximation. Methods using MATLAB. Gregory Fasshauer. Interdisciplinary Mathematical Sciences. Michael McCourt.
SINGAPORE SHANGHAI Vol TAIPEI - Interdisciplinary Mathematical Sciences 19 Kernel-based Approximation Methods using MATLAB Gregory Fasshauer Illinois Institute of Technology, USA Michael McCourt University
More informationDiscriminative Direction for Kernel Classifiers
Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering
More informationHoldout and Cross-Validation Methods Overfitting Avoidance
Holdout and Cross-Validation Methods Overfitting Avoidance Decision Trees Reduce error pruning Cost-complexity pruning Neural Networks Early stopping Adjusting Regularizers via Cross-Validation Nearest
More informationGeometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators
Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators Andrea Mambrini 1 University of Birmingham, Birmingham UK 6th June 2013 1 / 33 Andrea Mambrini GSGP: theory-laden
More informationYes, inner planets tend to be and outer planets tend to be.
1. Planet Density Make some general comments about inner and outer planets density Inner Planets Density Outer Planets Density Is there a pattern or a trend in planet density? Yes, inner planets tend to
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationSVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels
SVMs: Non-Separable Data, Convex Surrogate Loss, Multi-Class Classification, Kernels Karl Stratos June 21, 2018 1 / 33 Tangent: Some Loose Ends in Logistic Regression Polynomial feature expansion in logistic
More informationChapter 4 Thrills and Chills +Math +Depth Acceleration of the Moon +Concepts The Moon is 60 times further away from the center of Earth than objects on the surface of Earth, and moves about Earth in an
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationHSC General Maths. Name: Tracey Hughes-Butters 2011 Mathtastic Learning GENERAL MATHEMATICS NOTES HSC ALGEBRAIC MODELLING 3
Allgebraic Modelllling 3 Notes... HSC General Maths Name: Tracey Hughes-Butters 011 Mathtastic Learning 1 HSC CAPACITY MATRIX GENERAL MATHEMATICS TOPIC: Algebraic Modelling 3 & 4a Algebra skills & Linear
More informationModeling Data with Functions
Chapter 11 Modeling Data with Functions 11.1 Data Modeling Concepts 1 11.1.1 Conceptual Explanations: Modeling Data with Functions In school, you generally start with a function and work from there to
More informationUnit: Planetary Science
Orbital Motion Kepler s Laws GETTING AN ACCOUNT: 1) go to www.explorelearning.com 2) click on Enroll in a class (top right hand area of screen). 3) Where it says Enter class Code enter the number: MLTWD2YAZH
More informationLecture 9 Evolutionary Computation: Genetic algorithms
Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Simulation of natural evolution Genetic algorithms Case study: maintenance scheduling with genetic
More informationName. Satellite Motion Lab
Name Satellite Motion Lab Purpose To experiment with satellite motion using an interactive simulation in order to gain an understanding of Kepler s Laws of Planetary Motion and Newton s Law of Universal
More informationImportance of the study of extrasolar planets. Exoplanets Introduction. Importance of the study of extrasolar planets
Importance of the study of extrasolar planets Exoplanets Introduction Planets and Astrobiology (2017-2018) G. Vladilo Technological and scientific spin-offs Exoplanet observations are driving huge technological
More informationSupport Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM
1 Support Vector Machines (SVM) in bioinformatics Day 1: Introduction to SVM Jean-Philippe Vert Bioinformatics Center, Kyoto University, Japan Jean-Philippe.Vert@mines.org Human Genome Center, University
More informationCSC 4510 Machine Learning
10: Gene(c Algorithms CSC 4510 Machine Learning Dr. Mary Angela Papalaskari Department of CompuBng Sciences Villanova University Course website: www.csc.villanova.edu/~map/4510/ Slides of this presenta(on
More information1. The two triangles shown below are similar. This means that all the angles are equal and the sides are proportional.
1. The two triangles shown below are similar. This means that all the angles are equal and the sides are proportional. a. How many times bigger is the big triangle in comparison to the little triangle?
More information1 The Solar System. 1.1 a journey into our galaxy
1 The Solar System Though Pluto, and the far-flung depths of the Solar System, is the focus of this book, it is essential that Pluto is placed in the context of the planetary system that it inhabits our
More informationForecasting & Futurism
Article from: Forecasting & Futurism December 2013 Issue 8 A NEAT Approach to Neural Network Structure By Jeff Heaton Jeff Heaton Neural networks are a mainstay of artificial intelligence. These machine-learning
More informationARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92
ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 BIOLOGICAL INSPIRATIONS Some numbers The human brain contains about 10 billion nerve cells (neurons) Each neuron is connected to the others through 10000
More informationConceptual Explanations: Modeling Data with Functions
Conceptual Explanations: Modeling Data with Functions In school, you generally start with a function and work from there to numbers. Newton s Law tells us that F=ma. So if you push on a 3kg object with
More informationRelated Concepts: Lecture 9 SEM, Statistical Modeling, AI, and Data Mining. I. Terminology of SEM
Lecture 9 SEM, Statistical Modeling, AI, and Data Mining I. Terminology of SEM Related Concepts: Causal Modeling Path Analysis Structural Equation Modeling Latent variables (Factors measurable, but thru
More informationNSTA Web Seminar: Discover the Universe from Galileo to Today
LIVE INTERACTIVE LEARNING @ YOUR DESKTOP NSTA Web Seminar: Discover the Universe from Galileo to Today Presented by: Dr. Natalie Batalha Tuesday, January 20, 2009 International Year of Astronomy: Advances
More informationARTIFICIAL NEURAL NETWORK WITH HYBRID TAGUCHI-GENETIC ALGORITHM FOR NONLINEAR MIMO MODEL OF MACHINING PROCESSES
International Journal of Innovative Computing, Information and Control ICIC International c 2013 ISSN 1349-4198 Volume 9, Number 4, April 2013 pp. 1455 1475 ARTIFICIAL NEURAL NETWORK WITH HYBRID TAGUCHI-GENETIC
More informationPhysics 12. Unit 5 Circular Motion and Gravitation Part 2
Physics 12 Unit 5 Circular Motion and Gravitation Part 2 1. Newton s law of gravitation We have seen in Physics 11 that the force acting on an object due to gravity is given by a well known formula: F
More informationLarge and small planets
Large and small planets Journey through the Solar System C 41 time 50 minutes. Preparation For the activity Planets show the planets under each other on the board, as shown in the table. learning outcomes
More information9/12/2010. The Four Fundamental Forces of Nature. 1. Gravity 2. Electromagnetism 3. The Strong Nuclear Force 4. The Weak Nuclear Force
The Four Fundamental Forces of Nature 1. Gravity 2. Electromagnetism 3. The Strong Nuclear Force 4. The Weak Nuclear Force The Universe is made of matter Gravity the force of attraction between matter
More informationLecture 13. Gravity in the Solar System
Lecture 13 Gravity in the Solar System Guiding Questions 1. How was the heliocentric model established? What are monumental steps in the history of the heliocentric model? 2. How do Kepler s three laws
More informationGENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS
GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS A genetic algorithm is a random search technique for global optimisation in a complex search space. It was originally inspired by an
More informationIntroduction. Chapter 1
Chapter 1 Introduction In this book we will be concerned with supervised learning, which is the problem of learning input-output mappings from empirical data (the training dataset). Depending on the characteristics
More informationSupport Vector Machines and Kernel Methods
2018 CS420 Machine Learning, Lecture 3 Hangout from Prof. Andrew Ng. http://cs229.stanford.edu/notes/cs229-notes3.pdf Support Vector Machines and Kernel Methods Weinan Zhang Shanghai Jiao Tong University
More informationCS145: INTRODUCTION TO DATA MINING
CS145: INTRODUCTION TO DATA MINING 5: Vector Data: Support Vector Machine Instructor: Yizhou Sun yzsun@cs.ucla.edu October 18, 2017 Homework 1 Announcements Due end of the day of this Thursday (11:59pm)
More informationARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD
ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD WHAT IS A NEURAL NETWORK? The simplest definition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided
More informationCOMS 4771 Introduction to Machine Learning. Nakul Verma
COMS 4771 Introduction to Machine Learning Nakul Verma Announcements HW1 due next lecture Project details are available decide on the group and topic by Thursday Last time Generative vs. Discriminative
More information18 An Eclipsing Extrasolar Planet
Name: Date: 18 An Eclipsing Extrasolar Planet 18.1 Introduction One of the more recent new fields in astronomy is the search for (and discovery of) planets orbiting around stars other than our Sun, or
More informationEarly Theories. Early astronomers believed that the sun, planets and stars orbited Earth (geocentric model) Developed by Aristotle
Planetary Motion Early Theories Early astronomers believed that the sun, planets and stars orbited Earth (geocentric model) Developed by Aristotle Stars appear to move around Earth Observations showed
More informationAPPENDIX III - PLANETARY SYMBOLS DECODED CDEFGHIJ VENUS MARS JUPITER SATURN URANUS NEPTUNE PLUTO
PAGE 1 APPENDIX III - PLANETARY SYMBOLS DECODED There have been many attempts to decode the symbols of the planets. Most all of these decodings are either very complicated or somewhat mystical in nature.
More informationState Vector Reference Body Transformations Throughout The Solar System
Introduction One of the more mind-boggling programming challenges encountered during interplanetary trajectory design software development is how to transform a specified inertial position and velocity
More informationPedagogical information
SHOOTING STAR Shooting Star, an interactive computer simulation using calculation power of super computers. Students should investigate and become familiar with Kepler's laws, Newton's theory of gravitation,
More informationCharacterization of Jet Charge at the LHC
Characterization of Jet Charge at the LHC Thomas Dylan Rueter, Krishna Soni Abstract The Large Hadron Collider (LHC) produces a staggering amount of data - about 30 petabytes annually. One of the largest
More informationPHYSICS. Chapter 13 Lecture FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E RANDALL D. KNIGHT Pearson Education, Inc.
PHYSICS FOR SCIENTISTS AND ENGINEERS A STRATEGIC APPROACH 4/E Chapter 13 Lecture RANDALL D. KNIGHT Chapter 13 Newton s Theory of Gravity IN THIS CHAPTER, you will learn to understand the motion of satellites
More informationUnit 1: The Earth in the Universe
Unit 1: The Earth in the Universe 1. The Universe 1.1. First ideas about the Universe 1.2. Components and origin 1.3. Sizes and distances 2. The Solar System 3. The planet Earth 3.1. Movements of the Earth
More informationSupport Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar
Data Mining Support Vector Machines Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar 02/03/2018 Introduction to Data Mining 1 Support Vector Machines Find a linear hyperplane
More informationIntroduction to Computer Graphics (Lecture No 07) Ellipse and Other Curves
Introduction to Computer Graphics (Lecture No 07) Ellipse and Other Curves 7.1 Ellipse An ellipse is a curve that is the locus of all points in the plane the sum of whose distances r1 and r from two fixed
More informationLecture 6. Regression
Lecture 6. Regression Prof. Alan Yuille Summer 2014 Outline 1. Introduction to Regression 2. Binary Regression 3. Linear Regression; Polynomial Regression 4. Non-linear Regression; Multilayer Perceptron
More informationCopyright 2008 Pearson Education, Inc., publishing as Pearson Addison-Wesley.
Chapter 13. Newton s Theory of Gravity The beautiful rings of Saturn consist of countless centimeter-sized ice crystals, all orbiting the planet under the influence of gravity. Chapter Goal: To use Newton
More informationA FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE
A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE Li Sheng Institute of intelligent information engineering Zheiang University Hangzhou, 3007, P. R. China ABSTRACT In this paper, a neural network-driven
More informationVISUAL PHYSICS ONLINE
VISUAL PHYSICS ONLINE PRACTICAL ACTIVITY HOW DO THE PANETS MOVE? One of the most important questions historically in Physics was how the planets move. Many historians consider the field of Physics to date
More informationPortable Assisted Study Sequence ALGEBRA IIB
SCOPE This course is divided into two semesters of study (A & B) comprised of five units each. Each unit teaches concepts and strategies recommended for intermediate algebra students. The second half of
More informationFinding Other Earths. Jason H. Steffen. Asset Earth Waubonsee Community College October 1, 2009
Finding Other Earths Jason H. Steffen Asset Earth Waubonsee Community College October 1, 2009 True Earth Analog Necessities: 1) Main Sequence Star 2) Within the Stellar Habitable Zone 3) Roughly Earth
More informationLecture 5: Linear Genetic Programming
Lecture 5: Linear Genetic Programming CIU036 Artificial Intelligence 2, 2010 Krister Wolff, Ph.D. Department of Applied Mechanics Chalmers University of Technology 41296 Göteborg, Sweden krister.wolff@chalmers.se
More informationGiven a planet s radius, its orbital period in days is given by:
Sidereal Messenger by Daniel Browning, April 22 2013 Introduction A mathematician s job is never done. This brief item will attempt to improve on an equation that forges a relationship between the orbital
More informationThe Law of Ellipses (Kepler s First Law): all planets orbit the sun in a
Team Number Team Members Present Learning Objectives 1. Practice the Engineering Process a series of steps to follow to design a solution to a problem. 2. Practice the Five Dimensions of Being a Good Team
More informationTransit Tracks. Activity G14. What s This Activity About? Tips and Suggestions. What Will Students Do? What Will Students Learn?
G14 Transit Tracks Activity G14 Grade Level: 7 12 Source: This is a first version of an activity intended to be part of the Full Option Science System (FOSS) Planetary Science Course. Transit Tracks is
More informationBioinformatics: Network Analysis
Bioinformatics: Network Analysis Model Fitting COMP 572 (BIOS 572 / BIOE 564) - Fall 2013 Luay Nakhleh, Rice University 1 Outline Parameter estimation Model selection 2 Parameter Estimation 3 Generally
More informationChapter 14 Satellite Motion
1 Academic Physics Mechanics Chapter 14 Satellite Motion The Mechanical Universe Kepler's Three Laws (Episode 21) The Kepler Problem (Episode 22) Energy and Eccentricity (Episode 23) Navigating in Space
More informationASTR 200 : Lecture 6 Introduction to the Solar System Pearson Education Inc., publishing as Addison-Wesley
ASTR 200 : Lecture 6 Introduction to the Solar System 1 2004 Pearson Education Inc., publishing as Addison-Wesley Comparative Planetology Studying the similarities among and differences between the planets
More informationFIFTH MIDTERM -- REVIEW PROBLEMS
Physics 2210 Fall 2005 Paolo Gondolor FIFTH MIDTERM -- REVIEW PROBLEMS A solution set is available on the course web page in pdf format (no solutions for 27-29). 7 1. Assume that the planet Uranus has
More informationGravity and the Orbits of Planets
Gravity and the Orbits of Planets 1. Gravity Galileo Newton Earth s Gravity Mass v. Weight Einstein and General Relativity Round and irregular shaped objects 2. Orbits and Kepler s Laws ESO Galileo, Gravity,
More information7.4 Universal Gravitation
Circular Motion Velocity is a vector quantity, which means that it involves both speed (magnitude) and direction. Therefore an object traveling at a constant speed can still accelerate if the direction
More informationEvolution Strategies for Constants Optimization in Genetic Programming
Evolution Strategies for Constants Optimization in Genetic Programming César L. Alonso Centro de Inteligencia Artificial Universidad de Oviedo Campus de Viesques 33271 Gijón calonso@uniovi.es José Luis
More informationUnderstanding Generalization Error: Bounds and Decompositions
CIS 520: Machine Learning Spring 2018: Lecture 11 Understanding Generalization Error: Bounds and Decompositions Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the
More informationGraphical Analysis and Errors MBL
Graphical Analysis and Errors MBL I Graphical Analysis Graphs are vital tools for analyzing and displaying data Graphs allow us to explore the relationship between two quantities -- an independent variable
More informationEvidence that the Earth does not move: Greek Astronomy. Aristotelian Cosmology: Motions of the Planets. Ptolemy s Geocentric Model 2-1
Greek Astronomy Aristotelian Cosmology: Evidence that the Earth does not move: 1. Stars do not exhibit parallax: 2-1 At the center of the universe is the Earth: Changeable and imperfect. Above the Earth
More informationScaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.).
Local Search Scaling Up So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.). The current best such algorithms (RBFS / SMA*)
More informationDecision Tree Learning
Decision Tree Learning Berlin Chen Department of Computer Science & Information Engineering National Taiwan Normal University References: 1. Machine Learning, Chapter 3 2. Data Mining: Concepts, Models,
More informationAlgebra II Unit Breakdown (Curriculum Map Outline)
QUARTER 1 Unit 1 (Arithmetic and Geometric Sequences) A. Sequences as Functions 1. Identify finite and infinite sequences 2. Identify an arithmetic sequence and its parts 3. Identify a geometric sequence
More informationData Warehousing & Data Mining
13. Meta-Algorithms for Classification Data Warehousing & Data Mining Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 13.
More informationTycho Brahe
Tycho Brahe 1546-1601 At the time of Shakespeare and Elizabeth I and Champlain Lost part of his nose in a duel over who was the best mathematician At 27 he measured the distance of a supernova and a comet
More informationJohannes Kepler ( ) German Mathematician and Astronomer Passionately convinced of the rightness of the Copernican view. Set out to prove it!
Johannes Kepler (1571-1630) German Mathematician and Astronomer Passionately convinced of the rightness of the Copernican view. Set out to prove it! Kepler s Life Work Kepler sought a unifying principle
More informationLearning Outcomes in Focus
Contextual strand: E&S 3 Learning Outcomes in Focus Students should be able to interpret data to compare the Earth with other planets and moons in the solar system, with respect to properties including
More informationASTR 200 Reminder Center section = 'no laptop' zone
ASTR 200 Reminder Center section = 'no laptop' zone Laptops OK 1 NO Laptops Laptops OK ASTR 200 Frontiers of Astrophysics www.phas.ubc.ca/~gladman/a200 2 Homework in ASTR 200 The clarity and logic and
More informationIN THE ALMIGHTY GOD NAME Through the Mother of God mediation I do this research
Jupiter Uranus Distance (Metonic Cycle Reason) By Gerges Francis Twadrous 2 nd course student Physics Department - Physics & Math Faculty Peoples' Friendship University Moscow Russia -2010-2013 TEL +201022532292
More informationJeff Howbert Introduction to Machine Learning Winter
Classification / Regression Support Vector Machines Jeff Howbert Introduction to Machine Learning Winter 2012 1 Topics SVM classifiers for linearly separable classes SVM classifiers for non-linearly separable
More informationDetecting Dark Matter Halos Sam Beder, Brie Bunge, Adriana Diakite December 14, 2012
Detecting Dark Matter Halos Sam Beder, Brie Bunge, Adriana Diakite December 14, 2012 Introduction Dark matter s gravitational pull affects the positions of galaxies in the universe. If galaxies were circular,
More informationThe Doppler Method, or Radial Velocity Detection of Planets: I. Technique
ASTs309L The Doppler Method, or Radial Velocity Detection of Planets: I. Technique 1. Keplerian Orbits 2. Spectrographs/Doppler shifts 3. Precise Radial Velocity measurements ASTs309L The Doppler Effect:
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More informationGeometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators
Geometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators Andrea Mambrini University of Birmingham, UK NICaiA Exchange Programme LaMDA group, Nanjing University, China 7th
More informationAssignment 1. Due Jan. 31, 2017
Assignment 1 Due Jan. 31, 2017 Show all work and turn in answers on separate pages, not on these pages. Circle your final answers for clarity. Be sure to show/explain all of your reasoning and that your
More informationMathematics - High School Algebra II
Mathematics - High School Algebra II All West Virginia teachers are responsible for classroom instruction that integrates content standards and mathematical habits of mind. Students in this course will
More informationWhich of the following planets are all made up of gas? When a planets orbit around the Sun looks like an oval, it s called a(n)
When a planets orbit around the Sun looks like an oval, it s called a(n) - ellipse - circle - axis - rotation Which of the following planets are all made up of gas? - Venus, Mars, Saturn and Pluto - Jupiter,
More informationEvolutionary computation
Evolutionary computation Andrea Roli andrea.roli@unibo.it DEIS Alma Mater Studiorum Università di Bologna Evolutionary computation p. 1 Evolutionary Computation Evolutionary computation p. 2 Evolutionary
More informationChapter 2. Altitude Measurement
Chapter Altitude Measurement Although altitudes and zenith distances are equally suitable for navigational calculations, most formulas are traditionally based upon altitudes which are easily accessible
More informationLecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University
Lecture 18: Kernels Risk and Loss Support Vector Regression Aykut Erdem December 2016 Hacettepe University Administrative We will have a make-up lecture on next Saturday December 24, 2016 Presentations
More informationPatterns in the Solar System (Chapter 18)
GEOLOGY 306 Laboratory Instructor: TERRY J. BOROUGHS NAME: Patterns in the Solar System (Chapter 18) For this assignment you will require: a calculator, colored pencils, a metric ruler, and meter stick.
More informationData Mining Part 5. Prediction
Data Mining Part 5. Prediction 5.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline How the Brain Works Artificial Neural Networks Simple Computing Elements Feed-Forward Networks Perceptrons (Single-layer,
More informationPutting Earth In Its Place
Teacher Instructions Overview: During this activity, students build a model of our Solar System to gain insight into the relative sizes and distances involved. ives: The student will: create a scale model
More informationTAKEN FROM HORIZONS 7TH EDITION CHAPTER 1 TUTORIAL QUIZ
TAKEN FROM HORIZONS 7TH EDITION CHAPTER 1 TUTORIAL QUIZ 1. If the solar system is scaled down so that the Sun is represented by a basketball, a. a ping-pong ball located 500 feet away would properly represent
More informationIndirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina
Indirect Rule Learning: Support Vector Machines Indirect learning: loss optimization It doesn t estimate the prediction rule f (x) directly, since most loss functions do not have explicit optimizers. Indirection
More informationLecture 4: Kepler and Galileo. Astronomy 111 Wednesday September 6, 2017
Lecture 4: Kepler and Galileo Astronomy 111 Wednesday September 6, 2017 Reminders Online homework #2 due Monday at 3pm Johannes Kepler (1571-1630): German Was Tycho s assistant Used Tycho s data to discover
More informationKernel Methods & Support Vector Machines
Kernel Methods & Support Vector Machines Mahdi pakdaman Naeini PhD Candidate, University of Tehran Senior Researcher, TOSAN Intelligent Data Miners Outline Motivation Introduction to pattern recognition
More information