Geo-location estimation from two shadow trajectories

Similar documents
Computer Vision and Image Understanding

Earth & Space Science, Interpreting Data DURATION Preparation: 5 minutes Activity: 40 minutes (total over one day)

Gnomon (a thin, round stick at least a foot long and capable of being put into the ground or stood up vertically)


R.L. Kellogg, Ph.D. North American Sundial Society June 2013

ZW2000 and Your Vertical Sundial Carl Sabanski

Solar Schoolhouse Human Sundial. Building a Sundial with the Online Calculator

ANALEMMAS. on the gnomon and on the dial plate. Declining Shadows (Declining-Shadows.pdf ) for declination curve insights and methods.

Introduction To Modern Astronomy I: Solar System

Compendium* A Mean Time Sundial by Hendrik Hollander

Solar Schoolhouse Human Sundial Building a Sundial with the Online Calculator

A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES

CA1 2.11: Designing an Equatorial Sundial Activity

ZW2000 and Your Sundial

For most observers on Earth, the sun rises in the eastern

Exercise 6. Solar Panel Orientation EXERCISE OBJECTIVE DISCUSSION OUTLINE. Introduction to the importance of solar panel orientation DISCUSSION

ME 476 Solar Energy UNIT THREE SOLAR RADIATION

The Earth is a Rotating Sphere

ILLUSTRATING SHADOWS TABLE OF CONTENTS

Polar Sundial. Cut carefully on this line. Cut on this line.

Horizontal/Vertical/Inclining Sundial

Fundamentals of Satellite technology

Tracking Human Heads Based on Interaction between Hypotheses with Certainty

These notes may contain copyrighted material! They are for your own use only during this course.

Discovering the Night Sky

Discovering the Night Sky

3. The Sun s Position

Knowing the Heavens. Chapter Two. Guiding Questions. Naked-eye (unaided-eye) astronomy had an important place in ancient civilizations

Sunlight and its Properties II. EE 446/646 Y. Baghzouz

announcements 5/1/08 L10 Analysis and Design Tools 5/1/08 Analysis and Design Tools

The Sun -D ial of the I.H.B. Le cadran solaire du B.H.I.

Determining absolute orientation of a phone by imaging celestial bodies

Cartesian Coordinates Need two dimensional system 2 number lines perpendicular to each other X-axis is horizontal Y-axis is vertical Position relative

Single view metrology

Chapter 1 Solar Radiation

Oberth: Energy vs. Momentum

Chapter 3 Models of the Earth. 3.1 Finding Locations on the Earth. 3.1 Objectives

Research On Sun Shadow Positioning Genetic Algorithm Based On BP Neural Network Xu Ying

CUTTING Shadows SOME CUTOUT DIALS FOR YOUR AMUSEMENT. ALL YOU NEED ARE SOME SCISSORS, AND YOUR LATITUDE, AND LONGITUDE.

MiSP Astronomy - Seasons Worksheet #1 L2

The celestial sphere, the coordinates system, seasons, phases of the moon and eclipses. Chapters 2 and S1

Chapter 2 Lecture. The Cosmic Perspective Seventh Edition. Discovering the Universe for Yourself Pearson Education, Inc.

MiSP Astronomy Seasons Worksheet #1 L1

THE EARTH AND ITS REPRESENTATION

Knowing the Heavens. Goals: Constellations in the Sky

Chapter 6. Solar Geometry. Contents

Geostationary Satellites and Astrophotography

2. Modern: A constellation is a region in the sky. Every object in the sky, whether we can see it or not, is part of a constellation.

Geometry of Earth Sun System

Naples: A Disguised Analemmatic Dial

Knowing the Heavens. Goals: Constellations in the Sky

Modern Navigation. Thomas Herring

ClassAction: Coordinates and Motions Module Instructor s Manual

Geographers Tools: Location Systems Prof. Anthony Grande Hunter College Geography

Before you Sit. Please Pick-up: Blue Information Sheet for Evening Observing. 1 Red and 1 Blue ticket for Observing/ Planetarium

Which Earth latitude receives the greatest intensity of insolation when Earth is at the position shown in the diagram? A) 0 B) 23 N C) 55 N D) 90 N

LOCATING CELESTIAL OBJECTS: COORDINATES AND TIME. a. understand the basic concepts needed for any astronomical coordinate system.

Making a Sundial. Build a sundial and discover how time can be measured. Space Awareness, Leiden Observatory. iau.org/astroedu

A METHOD OF FINDING IMAGE SIMILAR PATCHES BASED ON GRADIENT-COVARIANCE SIMILARITY

Earth Motions Packet 14

6/17. Universe from Smallest to Largest:

REFERENCING COORDINATE SYSTEMS MAP PROJECTIONS GEOREFERENCING

Test Bank Chapter 2: Representations of Earth

4th National Conference on Electrical, Electronics and Computer Engineering (NCEECE 2015)

NAVIGATION THEORY QUESTIONS Basics of Navigation

A UNIFORM METHOD TO COMPUTE FLAT SUNDIALS

Degeneracies, Dependencies and their Implications in Multi-body and Multi-Sequence Factorizations

Chapter 1: Discovering the Night Sky. The sky is divided into 88 unequal areas that we call constellations.

RESTORATION OF VIDEO BY REMOVING RAIN

Determining the Translational Speed of a Camera from Time-Varying Optical Flow

Introduction to Geographic Information Science. Updates/News. Last Lecture. Geography 4103 / Map Projections and Coordinate Systems

Meridian Circle through Zenith, North Celestial Pole, Zenith Direction Straight Up from Observer. South Celestial Pole

5 - Seasons. Figure 1 shows two pictures of the Sun taken six months apart with the same camera, at the same time of the day, from the same location.

3. a. In the figure below, indicate the direction of the Sun with an arrow.

b. Assuming that the sundial is set up correctly, explain this observation.

SOLAR ENERGY: THAT S HOT Grades 4-6

Chapter 2 Discovering the Universe for Yourself. What does the universe look like from Earth? Constellations. 2.1 Patterns in the Night Sky

The Celestial Sphere. Chapter 1. Constellations. Models and Science. Constellations. Diurnal vs. Annular Motion 9/16/2010

Chapter 2 Discovering the Universe for Yourself

ASTRONOMY 25 SUMMER 2017 PROJECT 2: THE HEIGHT OF THE SUN

CELESTIAL COORDINATES

The Sundial Primer - "Dialling Guides" Double Gnomon Polar Sundial

Solar Time, Angles, and Irradiance Calculator: User Manual

1.4j interpret simple shadow stick data to determine local noon and observer s longitude

drinking straw, protractor, string, and rock. observer on Earth. Sun across the sky on March 21 as seen by an

Latitude and Longitude. North. Latitude. South

Map Projections (Part 1)

Lecture #03. January 20, 2010, Wednesday

Chapter 2 Discovering the Universe for Yourself. Copyright 2012 Pearson Education, Inc.

Chapter S1 Lecture. The Cosmic Perspective Seventh Edition. Celestial Timekeeping and Navigation Pearson Education, Inc.

arc setup and seasonal adjustments manual Solar timepiece

01) The Sun s rays strike the surface of the Earth at 90 degrees at the on December 22.

Geography involves the study of places: their locations, their characteristics, and how humans use and move around them.

Answer Key for Exam C

Answer Key for Exam B

Science : Introduction to Astronomy. Lecture 2 : Visual Astronomy -- Stars and Planets. Robert Fisher

FRIOT'S EXPLANATION OF THE TIP OF THE EARTH'S CONTRIBUTION TO THE EQUATION OF TIME

COMPUTER PROGRAM FOR THE ANGLES DESCRIBING THE SUN S APPARENT MOVEMENT IN THE SKY

Time and Diurnal Motion. 1a. The Earth Is Flat. 1c. Aristotle ( BC) 1b. The Earth Is Round. Time and Diurnal Motion

Sunlight and its Properties Part I. EE 446/646 Y. Baghzouz

Transcription:

Geo-location estimation from two shadow trajectories Lin Wu School of Computer Science and Technology Tianjin University, China wulin@tju.edu.cn Xiaochun Cao School of Computer Science and Technology Tianjin University, China xcao@tju.edu.cn Abstract The position of a world point s solar shadow depends on its geographical location, the geometrical relationship between the orientation of the sunshine and the ground plane where the shadow casts. This paper investigates the property of solar shadow trajectories on a planar surface and shows that camera parameters, latitude, longitude can be estimated from two observed shadow trajectories. Our contribution is that we use the design of the analemmatic sundial to get the shadow conic and furthermore recover the camera s geographical location. The proposed method does not require the shadow casting objects or a vertical object to be visible in the recovery of camera calibration. This approach is thoroughly validated on both synthetic and real data, and tested against various sources of errors including noise and number of observations.. Introduction Estimating geographic information is a high-level problem in computer vision [9, 8,, 7, 4]. A lot of cameras are mounted over the earth to undertake video surveillance, observe weather patterns, monitor coastline or navigation. Geo-location can provide plenty of background information including local climate, average temperature, rainfall and population density [8]. Currently, there are mainly three types of geolocating methods: feature matching based methods [9, 8, ], photometry based method [7] and geometry based methods [4]. The limitations shared within feature matching methods are the dependence on a great quantity of GPS tagged signals and the high complexity in the matching process between input images and known signals. The photometry based method provides the access to the angular velocity of the sun and the angle of each scene point s projection on the solar plane by analyzing the time-varying color model. The geometry based method avoids the GPS tagged information and relieves the constraints for the scene. The inputs are observed objects, e.g. the solar shadows, which are able to provide necessary geometric relationships among geographic location, scene geometry, and camera orientation. Our approach belongs to this line of research which needs to remove the distortions as well as [4]. Metric rectification is preliminary for geo-location estimation since the achievement of latitude from the geometric relationship residing in the Analemmatic sundial, holds true only in the world coordinate system. Shadows are important because they provide important visual cues, e.g. depth and shape, in our perception of the world [3, 4]. They are also interesting for camera calibration []and useful for computer vision applications such as planar object detection [8], visual surveillance [9, 3] and inverse lighting [5]. There has been some investigation of outdoor camera calibration (e.g. [5]) while most methods attempt to eliminate shadows for better performance rather than exploiting their properties to estimate camera parameters. In this paper, we take advantage of the fact that most of the webcams and surveillance cameras remain static, and propose a new method to recover the camera calibration and geo-location directly from shadow points. In this setup, only four shadow observations are required to calibrate the camera without requiring the objects to be visible. For geolocation, we employ the design of the analemmatic sundial to obtain geographical information. Compared to existing methods [4, 8], our approach does not require shadow casting objects s visibility and any GPS information.. Related work Vanishing points and the horizon line play central roles in perspective geometry [7] for applications such as camera calibration and image-based metrology [6]. The calibration method proposed by Antone and Bosse [] requires accurate geography information, which is not always possible to obtain from images. Inter-image constraints on the Image of Absolute Conic from objects and their cast shadows were introduced by Cao and Foroosh [], where the authors argue that two views of an outdoor scene with two vertical objects under sunshine are sufficient to calibrate a camera. Both these methods [, ] require the entire objects and their 978--444-6985-7//$6. IEEE 585

shadows to be present in the image. Junejo and Foroosh [4] use shadow points from shadow trajectory of objects to determine the geo-location of a camera. For the camera calibration, they argue that when shadow casting objects are not visible, it requires six imaged shadow positions of the same 3D points to fit a conic which can meet the line at infinity of the ground plane at two circular points. To estimate the geo-location, they require the entire shadow casting objects to be visible. Instead, we use the method of computing latitude inspired by the design of the analemmatic sundial which just requires the footprints of shadow casting objects ( [4] requires the entire object to be visible since they rely on the estimation of the Azimuth and Altitude of the sun). Sunkavalli et al. [7] focus on color changes and aim to separate the direct sunlight and ambient skylight components for the recovery of partial scene information. They recover scene geometry as well as the geo-location and geoorientation of the camera by utilizing common geometric constraints. Our method has the similar goals but different in three respects: () we recover the geometric constraints from only shadow trajectories while they assume the three mutually orthogonal vanishing lines are provided, i.e. a calibrated camera. () the inputs to our algorithm are shadow trajectories which are common in the world while they use color changes induced by variations in the spectral composition of daylight, and (3) our approach is simpler. The paper is organized as follows: Section 3 presents the elementary knowledge of geography and analemmatic sundial. Section 4 introduces the computation of camera calibration parameters and proposes a novel method to calculate a stationary camera s longitude and latitude. In Section 5, we verify the proposed method on both simulated and real data. Section 6 concludes this work. 3. Preliminaries 3.. Introduction of the analemmatic sundial An analemmatic sundial, also known as a Human Sundial, is flexible and movable as it uses a vertical gnomon such as a person standing in the proper position for the month. Its hour lines are the vertical projection of the hour lines of a circular equatorial sundial onto a flat plane. Therefore, the analemmatic sundial is an ellipse, where the long axis is aligned East-West and the short axis points North- South. The most important is that the hour points consisting of the sundial reflect the direction of shadows. With the availability of orthogonal projection on the plane casted by vertical objects, the varied shadow trajectories during the day can be obtained. We will show in Section 4. that from the time stamped observations of solar shadows we can recover the longitude, latitude and camera calibration. (a) (b) (c) (d) Figure. Automatic shadow trajectory detection. (a) A frame of real data. (b) Background. (c) Shadow point detection. (d) Shadow trajectory. 3.. About time: local and standard time Sundials indicate the local solar time, not the clock time, unless otherwise corrected. To obtain the solar time, three types of corrections need to be considered when clock time are available. First, due to the facts that the orbit of the Earth is not perfectly circular, and its rotational axis is not perfectly perpendicular to its orbit, there are small variations in the sundial time throughout the year. The required correction may be as large as a quarter-hour early or late which is described by the equation of time (t EOT ). Note that this correction is global and does not depend on the local latitude of sundial. Second, the solar time needs to be corrected for the longitude of the sundial relative to the central meridian at which the official time zone is defined. Within a given time zone, if the sundial is located at the central meridian of that time zone, it will show the clock time with the appropriate correction of the equation of time. However, if the sundial is located to the west (east), it will indicate slow (fast). This correction is often made by rotating the hour-lines by an angle equaling the difference in longitudes (We use t off to describe it). Third, the practice of daylight saving time (t DST ), which is either or, shifts the official time away from solar time by an hour advanced in some countries during the summer. 3.3. Shadow trajectory detection We employ a semi-automatic approach to detect the shadow point for an input video. First, we construct a background image B for the set of input images V= {I,I,...I k }: B(x, y) = max (I i(x, y)) () i [,k] In B, each pixel (x,y) contains the brightest value among V. The brightest values in B are computed on gray scale (Fig. (b)). After applying the background subtraction, the shadow region stands out by thresholding the absolute differences. Then we select a point in the shadow region and enforce the flood fill algorithm to determine the area. Finally, the prominent shadow tip is highlighted by using the PCA technique (Fig. (c)). Shown in Fig. (d) is the shadow trajectory of the left object that is made up of shadow tips at a quarter s interval. 586

Figure. Symmetric relationship in the shadow trajectory. s i is the shadow points with the offset i hours to solar time :. b is the imaged footprint of a shadow casting object. l noon is the symmetric axis of the shadow trajectory and l is the horizon. 4. Our Method 4.. Camera calibration using only four shadow observations It has been demonstrated that any two 3D corresponding lines on the shadow trajectories are parallel to each other, and hence the imaged lines provide the vanishing points []. Therefore, it is evident that the shadows observed over time are sufficient to provide the horizon line. Assuming a unit aspect ratio and zero camera skew, the imaginary point conic ω can be expanded up to scale as ω u v u v f + u + v, () where (u,v ) is the principal point that is known to be approximately at the center of a image [, 7], and the focal length f, the remaining constraint, can be achieved from the orthogonal relationship provided by v T x ωv y =, (3) where v x and v y are vanishing points of lines with perpendicular directions. By exploring the shadow trajectory s property, we find that the shadow points with the same offset to the solar noon point indicated by the solar time are symmetric with respect to the local noon line l noon. As shown in Fig., shadow points s i and s i indicated by solar time have the same offset, i hours, with respect to solar noon point s, and therefore are symmetric with l noon. s i is i hours before s while s i after it. The pair (s i, s i ) is defined as a paired shadow points. The lines connecting the paired shadow points are parallel to each other and intersect at a vanishing point v p on the horizon. Therefore, we can get midpoints which lie on the local noon line by using the cross ratio relationship: s i m i s i v p m i v p s i s i =, (4) Figure 3. The Analemmatic sundial. The hour point A indicates 5 pm in solar time with A=[X A,Y A] T lying on the sundial whereas the real shadow point lies in D. O is the center of the sundial and E is the footprint location (i.e. the scale of date). where (s i, s i ) are paired shadow points, m i is the midpoint of line segment s i s i, and v p is the vanishing point. Herein, if the imaged footprint b is available, l noon can be determined by connecting b and m i. The vanishing point v y =l noon l is perpendicular to v p. Then v y and v p satisfy the Eq. (3) to obtain ω. In the case where b is not available, another pair of shadow points s j and s j provides the second point m j on l noon. When the local longitude is unknown, we need to compute the time correction t off that can be determined by finding paired shadow points which can induce a vanishing point. The symmetric relation in the shadow trajectory yields that the lines connecting paired shadow points are parallel to each other. In Fig., lines s s and s 3 s 3 intersect on a vanishing point v p. The lines s s 3 and s 3 s 4 also intersect at a point which, however, does not lie on the horizon. Assume (s i, s i ) and (s j, s j ) to be shadow points indicated by clock time, we wish to search around two paired shadow points to find the correct t off to make the lines connecting (s i+t off, s i+t off ) and (s j+t off, s j+t off ) converge to a vanishing point p(t off ). Therefore, we estimate the time correction by minimizing the following distances between the converging point p(t off ) and the horizon. ˆt off = arg min t off d(l, p(t off )) (5) where d(l, p(t off )) represents the Euclidean distance between the line l and the point p(t off ) which varies along with t off in the searching process. 4.. Computation of Geographical location In the design of sundial, the horizontal dimension lies on the East-West axis, the vertical dimension lies on the North- South axis with the shadow-cast object s position varies along the sundial s short axis [6], seeing Fig. 3. Therefore, the corresponding shadow-tip, if falls onto a flat surface, will trace out a conic section, seeing the green points 587

in Fig. 3. The conic s symmetric axis is exactly the North- South direction. i.e. the analemmatic sundial s short axis. There exists an offset between the solar time indicated by the sundial and the standard clock time because of equation of time, difference between central meridian and local meridian (i.e. t off ), and daylight saving time. The local longitude can be achieved by correcting the reference longitude with a longitude correction t lc which is computed as t lc = t off + t EOT t DST, (6) Since every movement of the sun is equivalent to 4 minutes, the longitude correction is applied to the sundial as L = L meridian.5 t lc (7) Relative error % Absolute error 8 6 4..4.6.8..4.6.8 4 3.5 3.5.5.5 (a) f (without footprint) Relative error % Absolute error 4 3.5 3.5.5.5.4.35.3.5..5..5..4.6.8..4.6.8 (b) f (with footprint) where L meridian is the central meridian of the time zone which indicates the standard time, L is the local meridian where the shadow casting object locates. In addition, the equation of time and daylight saving time are available from the date when the video is recorded. In terms of the design of the sundial, the horizontal dimension is defined as: X =sin, (8) where =(T 4 ) 5 is the hour angle, T 4 is the solar time in hours that can be calculated from the longitude correction and clock time. The vertical dimension is: Y =sinφ cos (9) The footprint s location E=[,Y E ] T relative to the central point of the analemmatic sundial is: Y E = tan δ cos φ () where δ is the sun s declination which varies from -3 to 3 during a year. Therefore, the latitude φ is estimated as Y Y E X = sin φ cos tan δ cos φ sin = DC EC. () where DC and EC are the shadow length s measurements in North-South and West-East direction, respectively. 5. Experimental results 5.. Computer simulation The simulated camera has the focal length of f =, the aspect ratio of λ=, the skew of γ =, and the principal point at u = and v = 5. The synthetic light source is at infinity with the declination of δ = 3, the equation of time of t EOT = 5. minutes (4 th Dec), the latitude of φ = 39. The two stationary 3D objects locating at the northern hemisphere (39 N, 7. E) have heights of and units from the ground plane, respectively. The distance between the two footprints on the ground plane is units...4.6.8..4..4.6.8..4.6.8 (c) latitude (d) longitude Figure 4. Performance with respect to added random noise ranging from to pixels. 5.. Performance versus noise level In this experiment, the simulated shadow points are collected from 8: am to 4: pm with a frame speed at 3 frames per second. Gaussian noise with zero mean and a standard deviation of σ. pixels was added to the image shadow points. For each noise level, independent trials were performed. For the focal length f, we measured the relative error with respect to its ground truth. The geolocation estimation was measured with absolute error. Due to the important role played by the footprint in camera calibration, we evaluate our method in situations with and without known footprint location. Fig. 4 (a) and (b) show the average relative errors of f as functions of noise variance without and with footprint s influence. It can be seen that the two kinds of errors increase almost linearly as functions of the noise level and better results can be achieved by taking footprint into consideration. Since the footprint is required in the geo-location estimation, we employ it to get preferable results in the following experiments. For σ =, the relative error of focal length f is.%. For higher noise, σ =, the relative error in focal length increases steadily to reach 3.66%. Meanwhile the maximum absolute error of latitude s mean value (Fig. 4 (c)) is 5.7 with the noise level reaching pixels. In Fig. 4 (d), the absolute error of longitude compared with ground truth value is within ±.38 when the noise level reaches pixels. 5.. Performance versus number of observations Herein, the assumption is given that longitude has been determined using the proposed method in Section 4. The selected shadow points should have the same offset with re- 588

3.5 3.8 Relative error %.5.5.5 5 5 5 3 35 Quantity of selected points Absolute error.6.4..8.6.4 5 5 5 3 35 Quantity of selected points (a) f (b) latitude Figure 5. Performance with respect to the number of shadow observations. (a) (b) (c) Figure 7. Camera calibration and geolocation estimation on real images. (a) Two shadow trajectories are labeled with red and blue in a real frame captured by an indoor camera. (b) The red line is the estimated noon line and the cyan squares are midpoints. The blue lines connecting solar points converge at vanishing points labeled with green squares. (c) The latitude estimation illustration on the calibrated image. The green and blue are measurements along the North-South and West-East, respectively. (a) Tianjin, China (b) Zätec (c) Tianjin, China Figure 6. Calibration using the orthogonal vanishing points recovered from real images. Two set of lines labeled with red and blue provide the orthogonal relationship to calibrate the camera. spect to the solar noon point i.e. :, and also in pairs. For example, the 4 points we used in the experiment were selected as: 8:, 9:, 5:, 6: (solar time). (8:, 6:) and (9:, 5:) were two pairs which have the same offset with respect to :. We also added a noise level of σ = pixels to all projected image points. We repeated the process times and computed the average errors. The number of observations varied from 4 to 3 in step of 4, and the results are shown in Fig. 5. It can be seen that the relative errors decrease as the number of shadow points increase, and the same situations hold for the latitude estimation. Both the focal length and latitude become stable when the number of observations reaches 4. 5.. Real data We also applied our method to three sets of real data as shown in Fig. 7 and Fig. 8. In the first test, a surveillance camera was set up indoor to capture two vertical objects and casted shadows between 9: am and 3: pm (Beijing time). This dataset was captured on the 4 th December at longitude 7. E and latitude 39. N, Tianjin, China. The casted shadow trajectories of two objects in the scene were marked in red and blue in Fig. 7 (a), and the footprints (magenta) were selected for the computation of geolocation. The shadow points and footprints were picked manually. Fig. 7 (b) shows the two objects shadow points picked over time and the recovery of horizon. The time correction was determined by searching solar points having the same offsets with respect to the solar noon. The lines (a) (c) (d) Figure 8. Geolocation estimation on images captured by outdoor cameras. (a) and (c) Images captured by a static webcam and camera, respectively. Two shadow trajectories are labeled with red and blue. The cyan and red line are the estimated horizon and solar noon line. The green squares are midpoints and the green plus is the footprint. (b) and (d) The latitude estimation illustration on the calibrated image. connecting the solar points with the correct time correction converge at a vanishing point, which lies on the horizon (the red square in Fig. 7 (b)). Otherwise, the node is just a point with a certain distance from the horizon and all the searching nodes (cyan squares in Fig. 7 (b) ) locate on the same line. Consequently, only the correct time correction can induce the minimal distance between the point i.e. vanishing point and the horizon. The latitude estimating process are illustrated in Fig. 7 (c). Another performance was evaluated on the images (b) 589

captured by a webcam installed in Zätec, Czech Republic. In this set, 5 images were captured live from a square of Zätec, using one of the webcams available online at http://siemens.mesto-zatec.cz/webcam.html. The ground truth for this data set is as follows: longitude 3.5 E, latitude 5.3 N, and the declination angle 5.5. As illustrated in Fig. 8 (a), two shadow trajectories were composed of manual-picked shadow points from images captured from : am to 4: pm (clock time). We estimate the longitude as 3.77 E, the latitude as 48.37 N. Therefore, the absolute error of longitude estimation is equivalents to 9 km in remote geographic location. In the third real set, a static camera was installed to capture two objects in a square of Tianjin, China on th October. As shown in Fig.8 (c), shadow trajectories are estimated automatically. Calibration and geolocation results are given in the third row of Table. Note: The ground truth values of focal length for the three experiments were obtained by using calibrated objects, seeing Fig. 6. The calibration and geolocation results for the three sets of real images compared with the ground truth are shown in Table. Table. Calibration and geolocation on real images City Date f GT f est long GT long est lat GT lat est Tianjin 4 th Dec 663 687 7. 7.45 39. 37.7 Zätec nd May 568 64 3.5 3.77 5.3 48.37 Tianjin th Oct 84 84 7. 7.675 39. 36.8 6. Conclusion This paper addresses the problem of computing the geolocation from only two shadow trajectories from uncalibrated images. Camera calibration is achieved by deriving the horizon line and the metric rectification with only four shadow observations, which demands no measurements of any distance of angle in 3D world. More importantly, it relaxes the constrains found in other geo-location estimating methods using the altitude and azimuth angles of the sun orientations, and hence is especially useful in cases where only limited information is available. These advantages make our method an efficient calibration technique that could be especially suitable for outdoor video surveillance systems. 7. *Acknowledgement This work was supported by NSFC (No. 6959), One Hundred Talents Programme of Tianjin, Scientific Research f GT and f est denotes the ground truth and estimated values for focal length. Such references (lat GT, lat est) (long GT, long est) havethe same meaning for latitude and longitude. Foundation for the Returned Overseas Chinese Scholars, SEM., Open Projects Program of NLPR. References [] M. Antone and M. Bosse. Calibration of outdoor cameras from cast shadows. In Proc. IEEE Int. Conf. Systems, Man and Cybernetics, pages 34 345, 4. [] X. Cao and H. Foroosh. Camera calibration and light source estimation from images with shadows. Computer Vision and Image UnderStanding, 4():7 4, 7. [3] Y. Caspi and M. Werman. Vertical parallax from moving shadows. In Proc. IEEE CVPR, pages 39 35, 6. [4] Y. Chuang, D. Goldman, B. Curless, D. Salesin, and R. Szeliski. Shadow matting and compositing. ACM Trans. Graph., (3):494 5, 3. [5] R. Collins and Y. Tsin. Calibration of an outdoor active camera system. In Proc. CVPR, pages 58 534, 999. [6] A. Criminisi, I. Reid, and A. Zisserman. Single view metrology. Int. J. Comput. Vision, 4():3 48,. [7] R. I. Hartley and A. Zisserman. Multiple View Geometry in Computer Vision. Cambridge University Press, 4. [8] J. Hays and A. Efros. Imgps: estimating geographic information from a single image. In Proc. CVPR, pages 8, 8. [9] O. Javed and M. Shah. Tracking and object classification for automated surveillance. In Proc. ECCV, pages 343 357,. [] M. K. Johnson and H. Farid. Detecting photographic composites of people. Proc. IWDW, 7. [] F. Lu, X. Cao, Y. Shen, and H. Foroosh. Camera calibration from two shadow trajectories. In Proc. ICPR, pages II: 4, 6. [] N. R. R. S. N. Jacobs, S. Satkin and R. Pless. Geolocating static cameras. In Proc. ICCV, pages 6, 7. [3] S. Nadimi and B. Bhanu. Physical models for moving shadow and object detection in video. IEEE Trans. Pattern Anal. Mach. Intell., 6(8):79 87, 4. [4] I. N.Junejo and H. Foroosh. Estimating geo-temporal location of stationary cameras using shadow trajectories. In Proc. ECCV, pages 38 33, 8. [5] T. Okabe, I. Sato, and Y. Sato. Spherical harmonics vs. haarwavelets: Basis for recovering illumination from cast shadows. In Proc. IEEE CVPR, pages 5 57, 4. [6] R. R. Rohr. Sundials: History, Theory, and Practice. Dover Publications, 996. [7] K. Sunkavalli, F. Romeiro, W. Matusik, T. Zickler, and H. Pfister. What do color changes reveal about an outdoor scene? In Proc. IEEE CVPR, pages 594 6, 8. [8] Y. Yacoob and L. Davis. Segmentation of planar objects and their shadows in motion sequences. Int. J. Comput. Vision, 67():53 69, 6. [9] W. Zhang and J.Kosecka. Image based localization in urban environments. In Proc. 3DPVT, pages 33 4, 6. 59