Sky Segmentation in the Wild: An Empirical Study

Similar documents
Detection of Metadata Tampering through Discrepancy between Image Content and Metadata using Multi-task Deep Learning

A RAIN PIXEL RESTORATION ALGORITHM FOR VIDEOS WITH DYNAMIC SCENES

OBJECT DETECTION FROM MMS IMAGERY USING DEEP LEARNING FOR GENERATION OF ROAD ORTHOPHOTOS

Grade 6 Standard 2 Unit Test Astronomy

Towards Fully-automated Driving

Smoothed Prediction of the Onset of Tree Stem Radius Increase Based on Temperature Patterns

Global Scene Representations. Tilke Judd

Towards a Data-driven Approach to Exploring Galaxy Evolution via Generative Adversarial Networks

Anticipating Visual Representations from Unlabeled Data. Carl Vondrick, Hamed Pirsiavash, Antonio Torralba

Classification of Hand-Written Digits Using Scattering Convolutional Network

Learning theory. Ensemble methods. Boosting. Boosting: history

Road Surface Condition Analysis from Web Camera Images and Weather data. Torgeir Vaa (SVV), Terje Moen (SINTEF), Junyong You (CMR), Jeremy Cook (CMR)

Real-time image-based parking occupancy detection using deep learning. Debaditya Acharya, Weilin Yan & Kourosh Khoshelham The University of Melbourne

TILT, DAYLIGHT AND SEASONS WORKSHEET

Understanding Travel Time to Airports in New York City Sierra Gentry Dominik Schunack

Weather Prediction Using Historical Data

5 - Seasons. Figure 1 shows two pictures of the Sun taken six months apart with the same camera, at the same time of the day, from the same location.

Earth Motions Packet 14

Laboratory Exercise #7 - Introduction to Atmospheric Science: The Seasons

Investigating Factors that Influence Climate

Mitosis Detection in Breast Cancer Histology Images with Multi Column Deep Neural Networks

A Support Vector Regression Model for Forecasting Rainfall

AREP GAW. AQ Forecasting

Weather Cloud Detection. Measuring infrared temperature from the environment to characterize the current weather

DETECTING HUMAN ACTIVITIES IN THE ARCTIC OCEAN BY CONSTRUCTING AND ANALYZING SUPER-RESOLUTION IMAGES FROM MODIS DATA INTRODUCTION

Orbital Insight Energy: Oil Storage v5.1 Methodologies & Data Documentation

Eigen analysis and gray alignment for shadow detection applied to urban scene images

Anomaly Detection for the CERN Large Hadron Collider injection magnets

Forecasting the electricity consumption by aggregating specialized experts

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Asaf Bar Zvi Adi Hayat. Semantic Segmentation

Urban land use information retrieval based on scene classification of Google Street View images

YACT (Yet Another Climate Tool)? The SPI Explorer

CITS 4402 Computer Vision

Global Climates. Name Date

How to evaluate daylight. Initiated by the VELUX Group

MSG/SEVIRI AND METOP/AVHRR SNOW EXTENT PRODUCTS IN H-SAF

Towards understanding feedback from supermassive black holes using convolutional neural networks

GAMINGRE 8/1/ of 7

RESTORATION OF VIDEO BY REMOVING RAIN

Deep Learning for Natural Language Processing. Sidharth Mudgal April 4, 2017

Observing Dark Worlds (Final Report)

Climate Change Impact Analysis

Long-Term Time Series of Water Vapour Total Columns from GOME, SCIAMACHY and GOME-2

Basic human requirements

Convolutional Neural Networks

The Changing Landscape of Land Administration

Lecture 5 Neural models for NLP

PROBA-V CLOUD MASK VALIDATION

Research Article Weather Forecasting Using Sliding Window Algorithm

Technical note on seasonal adjustment for M0

13 SHADOW FLICKER Introduction Methodology

Laboratory Exercise #7 - Introduction to Atmospheric Science: The Seasons and Daily Weather

K-12 Environmental Education

A AVHRR NDVI dataset for Svalbard. Stian Solbø, Inge Lauknes, Cecilie Sneberg Grøtteland, Stine Skrunes, Hannah Vickers, Kjell Arild Høgda

Effect of Environmental Factors on Free-Flow Speed

Temperature Changes OBJECTIVES PREPARATION SCHEDULE MATERIALS. The students. For each student. For each team of two. For the class

Deep Convolutional Neural Networks for Pairwise Causality

Average Weather In March For Fukuoka, Japan

Agricultural Science Climatology Semester 2, Anne Green / Richard Thompson

5.6. Barrow, Alaska, USA

Algorithm-Independent Learning Issues

Exercise 6. Solar Panel Orientation EXERCISE OBJECTIVE DISCUSSION OUTLINE. Introduction to the importance of solar panel orientation DISCUSSION

JOURNAL OF INTERNATIONAL ACADEMIC RESEARCH FOR MULTIDISCIPLINARY Impact Factor 1.393, ISSN: , Volume 2, Issue 4, May 2014

Two-Stream Bidirectional Long Short-Term Memory for Mitosis Event Detection and Stage Localization in Phase-Contrast Microscopy Images

Random Forests: Finding Quasars

Learning with multiple models. Boosting.

Temporal and spatial approaches for land cover classification.

Visibility Estimation of Traffic Signals under Rainy Weather Conditions for Smart Driving Support

Sound Recognition in Mixtures

Weather and Travel Time Decision Support

What Is It Like Outside?

Basic cloud Interpretation using Satellite Imagery

A Hierarchical Convolutional Neural Network for Mitosis Detection in Phase-Contrast Microscopy Images

MODIS Snow Cover Mapping Decision Tree Technique: Snow and Cloud Discrimination

Climatography of the United States No

Introduction to Machine Learning

Statistical Filters for Crowd Image Analysis

COMPLEX INPUT CONVOLUTIONAL NEURAL NETWORKS FOR WIDE ANGLE SAR ATR

u.s. Naval Observatory Astronomical Applications Department

Caesar s Taxi Prediction Services

Prediction of Citations for Academic Papers

Parking Place Inspection System Utilizing a Mobile Robot with a Laser Range Finder -Application for occupancy state recognition-

Lecture 2 Motions in the Sky September 10, 2018

Third Grade Math and Science DBQ Weather and Climate/Representing and Interpreting Charts and Data - Teacher s Guide

W vs. QCD Jet Tagging at the Large Hadron Collider

Recurrent Autoregressive Networks for Online Multi-Object Tracking. Presented By: Ishan Gupta

Solar reflected glare affecting visual performance

Technical note on seasonal adjustment for Capital goods imports

Remote Ground based observations Merging Method For Visibility and Cloud Ceiling Assessment During the Night Using Data Mining Algorithms

CLIMATE OVERVIEW. Thunder Bay Climate Overview Page 1 of 5

DM-Group Meeting. Subhodip Biswas 10/16/2014

VALIDATION RESULTS OF THE OPERATIONAL LSA-SAF SNOW COVER MAPPING

Boosting: Algorithms and Applications

Optical Character Recognition of Jutakshars within Devanagari Script

Corners, Blobs & Descriptors. With slides from S. Lazebnik & S. Seitz, D. Lowe, A. Efros

International Journal of Scientific & Engineering Research, Volume 7, Issue 2, February ISSN

The Cryosphere Radiative Effect in CESM. Justin Perket Mark Flanner CESM Polar Climate Working Group Meeting Wednesday June 19, 2013

Earth Science: Second Quarter Grading Rubric Kindergarten

Transcription:

Sky Segmentation in the Wild: An Empirical Study Radu P. Mihail 1 Scott Workman 2 Zach Bessinger 2 Nathan Jacobs 2 rpmihail@valdosta.edu scott@cs.uky.edu zach@cs.uky.edu jacobs@cs.uky.edu 1 Valdosta State University 2 University of Kentucky Abstract Camera Ground Truth Segmentation Automatically determining which pixels in an image view the sky, the problem of sky segmentation, is a critical preprocessing step for a wide variety of outdoor image interpretation problems, including horizon estimation, robot navigation and image geolocalization. Many methods for this problem have been proposed with recent work achieving significant improvements on benchmark datasets. However, such datasets are often constructed to contain images captured in favorable conditions and, therefore, do not reflect the broad range of conditions with which a real-world vision system must cope. This paper presents the results of a large-scale empirical evaluation of the performance of three state-of-the-art approaches on a new dataset, which consists of roughly 1k images captured in the wild. The results show that the performance of these methods can be dramatically degraded by the local lighting and weather conditions. We propose a deep learning based variant of an ensemble solution that outperforms the methods we tested, in some cases achieving above 5% relative reduction in misclassified pixels. While our results show there is room for improvement, our hope is that this dataset will encourage others to improve the real-world performance of their algorithms. Frame Figure 1. One scene, observed over a long period of time, can change drastically in appearance due to time of day, day of year and weather. Above, we show six sample scenes from one static outdoor webcam and the results of three sky labeling methods (blue represents correct classification, red represents incorrect classification). 1. Introduction Image labeling algorithms assign a label (e.g., car, ground, sky, building) to every pixel in an image. Outdoor imagery captured in the wild poses challenges to these algorithms due to the variety of possible lighting and weather conditions. We focus on sky segmentation in single images which, while seemingly simple, is actually a very challenging and unsolved problem. Outdoor scene labeling has received much attention from vision research in the past few years, since it is an important pre-processing step for many high-level vision algorithms. Existing approaches perform well in favorable conditions (e.g., clear blue sky), however the effects of weather, season, and time drastically alter the appearance of an outdoor scene (Figure 1). This paper presents an extensive evaluation of three existing methods [3, 12, 17] on a challenging real-world dataset. We selected these methods for both their contributions to the community and their performance on benchmark datasets. However, benchmark datasets do not completely capture the variational appearance of the sky. Camera optics combined with the dynamics of sun position, cloud formations, and more, results in tremendous variability of sky appearance. The combination of these factors motivates the construction of a dataset representative of these real-world scenarios. To build such a dataset, we take advantage of imagery collected from static outdoor cameras over long periods of

time, thus increasing the probability of exposure to different atmospheric conditions. Since the cameras we use are static, one ground truth mask can be used to generate a large training set with significant sky appearance variability. In order to highlight various conditions in which labeling performance is suboptimal, we augment each image with weather observations collected from nearby weather stations. Often in real-world applications, it may not be possible to obtain weather data, so we also compute highlevel transient scene attributes, related to weather and appearance, for each image using the method of Laffont et al. [1]. Using this dataset, we perform several experiments to examine the impact that weather and time has on each methods performance. The key contributions of this work are: 1) introducing a large labeled image dataset; 2) presenting the results of a large-scale empirical evaluation of three stateof-the-art techniques for sky segmentation; 3) suggesting how the observations we make from our evaluation should guide future work on sky segmentation, and more generally pixel labeling; 4) a deep ensemble method that combines raw image data and existing methods output to make better predictions. 2. Related work Recent interest in outdoor imagery has led to various algorithms for calibration [8, 2], labeling [3, 18], geolocation [7, 22], geometry recovery [2, 8] and others. For example, Jacobs et al. [5] exploit pixel time-series from static webcams to recover a cloudmap and an estimate of scene geometry. Cues from multiple images (e.g., moving clouds, sun position) can be used as input to higher level vision algorithms (e.g., calibration [2], scene shape [4, 21], horizon estimation, geolocation [7]), but automatic detection of sky regions is difficult from single images. Weather detection and understanding has been successfully used in robot vision for navigation [9], driver assistance systems [14, 23] and image searching [16]. Scene labeling methods attempt to assign each pixel in an image to one of several categories of objects (e.g., sky, ground, road, tree, building). These methods [3, 11, 17, 18, 19] rely on the local appearance of the objects learned from a training set of instances. The scene parsing problem is most commonly addressed by a local classifier (using engineered features, or more recently, learning features using deep learning architectures) constrained by a graphical probability model (e.g., CRF or MRF) where global decisions are made to include high-level reasoning about spatial relationships. In this work, we present an empirical assessment of the performance of sky segmentation in the wild. Our work is most similar in conception to that of Stylianou et al. [15] who analyze feature matching performance over long time periods. We evaluate three methods that output (either directly or as a by-product) a sky segmentation: Hoiem et al. [3], [17] and [12]. The choice for these methods was motivated by their impact in the vision community and publicly available code. The contribution relevant to sky segmentation in the work of [3] is their use of geometric context (e.g.: a rough estimate of scene geometry) for three classes (sky, ground and vertical) inferred using statistical appearance learning. We only use the geometric label outputs from their algorithm. Tighe et al. [17] introduce an image parsing method based on combining region-level features with exemplar-svm sliding window detectors. We only use the sky label from the final results of their algorithm (i.e., other labels are considered not sky ). [12] explore single image weather classification into two classes: sunny or cloudy. We use their sky segmentation output in our evaluation. 3. Dataset We introduce a new dataset [1] of labeled outdoor images captured in a wide range of weather and illumination conditions. To the best of our knowledge, this dataset is the largest in existence with annotated sky pixels and associated weather data. Motivated by difficulties of existing methods to handle extreme appearance variations of sky regions, we take advantage of the long-term webcam imagery from the Archive of Many Outdoor Scenes (AMOS) [6]. We selected 53 cameras from AMOS that were static (i.e., no camera movement throughout one or more calendar years) and downloaded all the available images for that period, (an average of one year, with around ten thousand images per camera). To keep the dataset size reasonable, we keep five randomly selected frames for each day. For each camera we manually created a binary mask segmenting sky and ground. The average coverage of sky pixels for all the webcams is 41.19%, with standard deviation 15.71%. To quantify the effect of weather condition on labelers, we retrieved weather data from Wunderground.com for all of our cameras. Weather data was retrieved at every camera location for the entire period of the downloaded imagery, and the closest observation was associated with each image. For each frame we have indicators for several weather conditions: light rain, partly cloudy, clear, haze, scattered clouds, mostly cloudy, overcast, fog and light freezing fog. In addition, we augment each image in the dataset with Laffont et al. s [1] transient attributes, which are highlevel properties describing the appearance of a scene, identified via crowdsourcing on thousands of images from 11 webcams. We take advantage of their learned regression models to extract transient attributes for all of our images.

Sample Camera.1.2.3.4.5.6 Percentage of Images in Dataset with best Figure 3. Distribution of best per image with respect to method for the entire dataset. Figure 2. Sample images from five cameras (left-most column). In the right columns, we show cumulative per-pixel for all images, for all three methods evaluated. Red and blue indicate high and low, respectively, misclassification. 4. Experiments We processed every image in our dataset through the previously mentioned methods using the source code provided by their respective authors. While each method was pre-trained, running approximately 1, images through each method is computationally challenging. Our attempt to compensate for the large-scale processing of each method involved distributing the computation over a 16-node cluster for several weeks. Running each method on tens of thousands of images simply led to software crashes. To minimize this, we optimized various components of the methods, which allowed us to successfully run this large scale evaluation. There were some images that led to failure. These images were either completely dark, saturated, or with reduced visibility due to fog, etc. If a method failed for a given image, we excluded the image from the evaluation set. The tunable parameters used in this evaluation were the same as the original authors in their own evaluations. 4.1. Overall Accuracy To evaluate the methods we use a per-pixel performance metric, the misclassification rate (), computed per classified pixels frame as follows: M CR = # of incorrectly. total # of pixels Each of the 53 webcams has roughly 15 images. In Figure 2 we show five sample cameras (left column) with average for every method (right three columns). Overall, achieved the best performance, with lowest average of 16.41%, σ = 18.98%. Hoiem et al. came in second place, with average of 2.69%, σ = 22.13%. Finally, with average of 27.69%, σ = 23.5%. While is a good overall performance indicator, we also report type I errors (false positive, pixel is labeled as sky when it is not) and type II (false negative, a pixel is labeled as ground when it is not) for each method. As seen in Table 1, and have significantly less false positive errors than who achieved lower overall. To gain further insight into the methods, we counted the number of images in the dataset for each method with the best. While the overall lowest was achieved by, their method was outperformed on slightly less than half of the images in our dataset. The lowest was achieved on roughly 4% of the images by, and 1% by, as seen in Figure 3. We now consider individual pixels. It is often the case pixels are independently labeled correctly by at least one method. We compute the per-pixel, by counting a pixel as correctly classified when either one of the three methods labeled it correctly, and incorrect otherwise. Overall, the per-pixel is 1.9%. This suggests that improvements can be achieved using the methods results as a strong prior, combined with other factors that were found to have an impact on accuracy. 4.2. Impact of Lighting on Accuracy We observe that lighting conditions have a significant impact on the accuracy of scene labeling methods. When the sun is at its highest point in the sky, the scene is most equally illuminated. As the day progresses, and the sun lowers, the possibility of shadows or the appearance of the sun in the view increases. To investigate the effects of lighting on accuracy, we visualize the average with respect to the time of day in Figure 4. When the sun is at its highest point in the sky, around noon, all labelers are at peak performance. A similar result is obtained when visualizing average with respect to Table 1. Average type I & II errors. Method Avg. type I error Avg. type II error.87%.91% 14.68% 26.81% 19.7% 1.73%

1 2 3 4 5 6 7 8 9 1 11 12 13 14 15 16 17 18 19 2 21 22 23.5.45.4 w.r.t. hour of day.35.3 w.r.t. month of year.7.6 "clouds" transient attribute effect on, N=13292.35.3 "gloomy" transient attribute effect on, N=5699.35.25.5.25.3.25.2.15.2.15.1 P( clouds>.6).4.3.2 P( gloomy>.6).2.15.1.1.5.5.1.5 Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec Month of Year Figure 4. Time of day has a significant impact on labeler performance. All methods achieve their best performance when the sun is at its highest point, around noon. Less dramatic, month of year also has an impact on performance, with the best results seen during the spring and summer months. P( cold>.6).45.4.35.3.25.2.15.1.2.3.4.5.6.7.8.9 1 "cold" transient attribute effect on, N=6393 P( cold>.6).18.16.14.12.1.8.6.1.2.3.4.5.6.7.8.9 1 "night" transient attribute effect on, N=28779 in different weather conditions.1.5.4.2 Snow (N=68) Scattered Clouds (N=4157) Partly Cloudy (N=3516) Overcast (N=3426) Mostly Cloudy (N=6392) Light Snow (N=45) Light Rain (N=1161) Heavy Rain (N=48) Haze (N=172) Fog (N=172) Clear (N=9911).5.1.15.2.25.3.35 Figure 5. Average for each method given a subset of weather conditions. To reduce the effect of lighting, we only evaluated images taken between 8: a.m. and 6: p.m. with respect to the local time of each image. month of year. We believe the combined higher rates of failure in the winter months of the year are likely due to shorter days as a result of the Earth s tilt. 4.3. Impact of Weather on Accuracy We now explore the effects of known weather conditions on for the three methods. In Figure 5 we plot the s of each method given a subset of weather conditions. We highlight four weather indicators (fog, heavy rain, light snow and snow) where the most accurate labeler, Tighe et al., is outperformed by We believe this effect is attributed to labeler confidence, i.e., low type I error methods have an advantage for images with general occlusion. Contrary to our expectations, two of the sky labeling approaches ( and ) are mostly robust to cloud coverage. 4.4. Impact of Other Attributes on Accuracy We select three transient attributes that are related to a scene appearance as a function of time of day: bright, night, midday, and three weather-related attributes: dry, winter, summer. We threshold the regressor responses (a real-valued score on the interval [, 1] indicating the presence of the attribute) to >.6 and plot the distribution of the resulting images in Figure 7. We observed that the transient features, when above threshold, are indeed related to time.1.2.3.4.5.6.7.8.9 1.1.2.3.4.5.6.7.8.9 1 Figure 6. Probability distribution of given transient features above a threshold. Top left: and more robust to cloudiness than Top right: high failure rates for all three methods when gloomy is detected. Bottom left: all three methods are not robust to cold images. Bottom right: as expected, high failure rates occur for poorly lit images. The threshold for the transient attribute regressors was.6. Frequency of transient feature responses 5 45 4 35 3 25 2 15 1 5 bright >.6, N=14732 night >.6, N=31481 midday >.6, N=347 1 2 3 4 5 6 7 8 9 1 11 12 13 14 15 16 17 18 19 2 21 22 23 Frequency of transient feature responses 14 12 1 8 6 4 2 dry >.6, N=9326 winter >.6, N=349 summer >.6, N=2643 Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec Month Figure 7. Left: transient attributes (bright, day, midday) are related to ground truth hour of day. Right: transient attributes (dry, winter, summer) are related to ground truth month of year. of day and day of year. This provides strong support for using such features as a surrogate for weather data. We now explore labeler robustness with respect to transient attributes as predictor of high labeler failure rates. In Figure 6 we show that and are more robust to cloudiness conditions than The best predictors of high labeler failure are the gloomy, night and cold transient attributes. The methods we evaluated seem to be robust to cloudiness. In Figure 8 we show sample images with high cold and gloomy attributes. 5. Deep Ensemble Approach Based on our experiments and the insights we gained from the data analysis, we show improvements can be made by combining the outputs of the three methods we evaluated. While, achieved best overall performance, on a per-image count, they are outperformed on nearly half of the images in the dataset (Figure 3). This suggests that an ensemble method which uses the three methods output,

gloomy.35.3.25 w.r.t. hour of day NN Best/Image.2.15.1 cold Figure 8. Sample images with gloomy and cold transient attributes above.6. These images are difficult due to sky color and clouds/haze/fog..5.2.18.16.14.12.1 1 2 3 4 5 6 7 8 9 1 11 12 13 14 15 16 17 18 19 2 21 22 23 w.r.t. month of year NN BestPerImage.8 combined with raw image data, could outperform individual methods on our challenging dataset. We now describe an ensemble method based on a deep recurrent convolutional neural network (rcnn) architecture. 5.1. Recurrent CNN We use a recurrent convolutional neural network, similar to that of Pinheiro et al. [13]. The network consists of three convolutional layers with hyperbolic tangent activation functions. The recurrent architecture involves the composition of three instances of a convolutional neural network, with each instance sharing identical parameters. Input to the rcnn is a 3D matrix with the smallest dimension indexing over the color channels. The full model file, solver definition and learned weights of our networks are available at [1]. We trained two rcnns, one only with raw image data and another with the raw image data augmented by the outputs of the methods we evaluate. For the second network, we augmented the RGB input with the binary output of the three methods we evaluate, which results in a width height 6 input matrix. The output of the networks are probability maps of the same size as the input image. We threshold the output to obtain a binary label. Our data was split into a training set of 4 cameras and 13 test cameras. We trained each network on an NVIDIA Tesla M275 for two days. 5.2. Evaluation The results show rcnns can be successfully used as ensemble methods to learn a nonlinear combination of raw image data and the outputs of other methods to improve accuracy. Overall, our ensemble outperforms the three methods evaluated. On the test cameras, averaged an of 18.19%, averaged 23.8%, and averaged 3.6%, while our rcnn ensemble averaged an of 12.96%, a relative improvement of 28.75% from The rcnn trained on raw image data alone achieved an average of 17.28%. We compare this baseline approach to the ensemble in Figure 1. To gain a better understanding of where improvements are most significant, we aggregate with respect to month of year and hour of day, as seen in Figure 9. We note.6.4.2 Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec Month of Year Figure 9. Comparison of our recurrent neural network ensemble with three compositions. The comparisons are made w.r.t. Tighe et al. and the best per-image. rcnn outperforms the best method per-image metric when aggregated by month. Aggregation by hour of day reveals highest performance gains during daylight hours. significant improvements compared to the best per image 1 during daylight hours. Aggregation with respect to month of year shows improvements for all months, with higher performance gains during summer months, possibly attributable to longer days and easier weather conditions. 6. Conclusion To analyze sky segmentation performance in real outdoor scenes, we created a new challenging dataset from static outdoor webcams over long periods of time with known ground truth location to supplement local weather observations. We extensively evaluated the performance of three sky labeling methods ( [17], Hoiem et al. [3] and [12]) under real-world weather and illumination conditions. This exploratory study was driven by the importance of accurately segmenting sky pixels from outdoor imagery, as it serves as input to many high-level vision algorithms. Our results show that sky labeling algorithm performance varies most significantly with respect to illumination conditions, i.e., sun position as indicated by time of day. In addition, we found that certain weather conditions and time of day are good predictors of current labeler errors. We proposed a deep ensemble method that combines the output of existing methods with raw image data using an rcnn. Our model achieves better overall performance than any of the individual methods. Additionally, we aggregated results with respect to hour of day (the most important factor driving sky labeler performance) and compared our ensemble with a metric that uses an oracle to select the best method, showing improved performance during daylight hours. This work suggests two directions for future 1 The best per-image metric gives us an empirical performance boundary treating all three input methods equally.

.3.25 Comparison of baseline RCNN with ensemble RCNN baseline RCNN ensemble change from best per-image w.r.t. hour of day 6 5 4 NN improvement.2.15.1 Change in (%) 3 2 1-1.5-2 -3 1 2 3 4 5 6 7 8 9 1 11 12 13 14 15 16 17 18 19 2 21 22 23-4 1 2 3 4 5 6 7 8 9 1 11 12 13 14 15 16 17 18 19 2 21 22 23 Figure 1. Left: We compare the baseline recurrent neural network with the deep ensemble. We note improvements for all hours of the day. Right: We show relative changes in the deep ensemble when aggregated by hour of day and compared to the best per image. Positive change represents a decrease in, while negative change reflects an increase in error. work. One is in exploring alternative methods for integrating weather metadata into the sky segmentation algorithms. In particular, we think that adding the metadata earlier in the processing pipeline would be beneficial. The other is in exploring how incorporating temporal context and weather metadata can improve performance on other vision tasks. References [1] http://mypages.valdosta.edu/rpmihail/ skyfinder/. 2, 5 [2] D. Hoiem, A. A. Efros, and M. Hebert. Automatic photo pop-up. ACM Transactions on Graphics (SIGGRAPH), 24(3):577 584, 25. 2 [3] D. Hoiem, A. A. Efros, and M. Hebert. Geometric context from a single image. In IEEE International Conference on Computer Vision, 25. 1, 2, 5 [4] N. Jacobs, B. Bies, and R. Pless. Using Cloud Shadows to Infer Scene Structure and Camera Calibration. In Computer Vision and Pattern Recognition, 21. 2 [5] N. Jacobs, J. King, D. Bowers, and R. Souvenir. Estimating Cloud Maps from Outdoor Image Sequences. In IEEE Winter Conference on Applications of Computer Vision, 214. 2 [6] N. Jacobs, N. Roman, and R. Pless. Consistent Temporal Variations in Many Outdoor Scenes. In Computer Vision and Pattern Recognition, 27. 2 [7] N. Jacobs, S. Satkin, N. Roman, R. Speyer, and R. Pless. Geolocating static cameras. In IEEE International Conference on Computer Vision, 27. 2 [8] N. Jacobs, S. Workman, and R. Souvenir. Scene Geometry from Several Partly Cloudy Days. In International Conference on Distributed Smart Cameras, 213. 2 [9] H. Katsura, J. Miura, M. Hild, and Y. Shirai. A viewbased outdoor navigation using object recognition robust to changes of weather and seasons. In IEEE International Conference on Intelligent Robots and Systems, 23. 2 [1] P.-Y. Laffont, Z. Ren, X. Tao, C. Qian, and J. Hays. Transient attributes for high-level understanding and editing of outdoor scenes. ACM Transactions on Graphics (SIGGRAPH), 33(4), 214. 2 [11] C. Liu, J. Yuen, and A. Torralba. Nonparametric scene parsing: Label transfer via dense scene alignment. In Computer Vision and Pattern Recognition, 29. 2 [12] C. Lu, D. Lin, J. Jia, and C.-K. Tang. Two-class weather classification. In Computer Vision and Pattern Recognition, 214. 1, 2, 5 [13] P. Pinheiro and R. Collobert. Recurrent convolutional neural networks for scene labeling. In International Conference on Machine Learning, 214. 5 [14] M. Roser and F. Moosmann. Classification of weather situations on single color images. In IEEE Intelligent Vehicles Symposium, 28. 2 [15] A. Stylianou, A. Abrams, and R. Pless. Characterizing feature matching performance over long time periods. In IEEE Winter Conference on Applications of Computer Vision, 215. 2 [16] L. Tao, L. Yuan, and J. Sun. Skyfinder: attribute-based sky image search. ACM Transactions on Graphics (SIGGRAPH), 28(3):68, 29. 2 [17] J. Tighe and S. Lazebnik. Superparsing: scalable nonparametric image parsing with superpixels. In European Conference on Computer Vision, 21. 1, 2, 5 [18] J. Tighe and S. Lazebnik. Finding things: Image parsing with regions and per-exemplar detectors. In Computer Vision and Pattern Recognition, 213. 2 [19] J. Tighe, M. Niethammer, and S. Lazebnik. Scene parsing with object instances and occlusion ordering. In Computer Vision and Pattern Recognition, 214. 2 [2] S. Workman, R. P. Mihail, and N. Jacobs. A pot of gold: Rainbows as a calibration cue. In European Conference on Computer Vision, 214. 2 [21] S. Workman, R. Souvenir, and N. Jacobs. Scene Shape Estimation from Multiple Partly Cloud Days. Computer Vision and Image Understanding, 134:116 129, 215. 2 [22] S. Workman, R. Souvenir, and N. Jacobs. Wide-area image geolocalization with aerial reference imagery. In IEEE International Conference on Computer Vision, 215. 2 [23] X. Yan, Y. Luo, and X. Zheng. Weather recognition based on images captured by vision system in vehicle. In International Symposium on Neural Networks, 29. 2