Automatic detection and of dipoles in large area SQUID magnetometry

Similar documents
Review for Exam 1. Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA

Final Exam, Machine Learning, Spring 2009

Holdout and Cross-Validation Methods Overfitting Avoidance

Jeff Howbert Introduction to Machine Learning Winter

CS 188: Artificial Intelligence Spring Announcements

SUPPLEMENTARY INFORMATION

CS4495/6495 Introduction to Computer Vision. 8C-L3 Support Vector Machines

CS 188: Artificial Intelligence Spring Announcements

Algorithm User Guide:

Machine Learning. Lecture 9: Learning Theory. Feng Li.

Announcements. CS 188: Artificial Intelligence Spring Classification. Today. Classification overview. Case-Based Reasoning

Machine Learning, Midterm Exam: Spring 2009 SOLUTION

Lecture 7: Edge Detection

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

Predicting New Search-Query Cluster Volume

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Discriminative part-based models. Many slides based on P. Felzenszwalb

Support Vector Machines

Jeffrey D. Ullman Stanford University

Support Vector Machines

Machine Learning, Midterm Exam

W vs. QCD Jet Tagging at the Large Hadron Collider

FINAL EXAM: FALL 2013 CS 6375 INSTRUCTOR: VIBHAV GOGATE

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

Support Vector Machines: Maximum Margin Classifiers

Microarray Data Analysis: Discovery

Machine Learning Lecture 6 Note

Introduction to Deep Learning

Linear Classifiers as Pattern Detectors

6.036 midterm review. Wednesday, March 18, 15

Introduction to Machine Learning Midterm Exam

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

Cheng Soon Ong & Christian Walder. Canberra February June 2018

HYPERGRAPH BASED SEMI-SUPERVISED LEARNING ALGORITHMS APPLIED TO SPEECH RECOGNITION PROBLEM: A NOVEL APPROACH

Can a Magnetic Field Produce a Current?

Learning Methods for Linear Detectors

Stephen Scott.

MIRA, SVM, k-nn. Lirong Xia

Data Mining Prof. Pabitra Mitra Department of Computer Science & Engineering Indian Institute of Technology, Kharagpur

Training the linear classifier

Midterm: CS 6375 Spring 2015 Solutions

Advanced Topics in Machine Learning, Summer Semester 2012

SVMs, Duality and the Kernel Trick

Neural networks and optimization

Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012

Final Exam, Fall 2002

Deep Learning. Convolutional Neural Network (CNNs) Ali Ghodsi. October 30, Slides are partially based on Book in preparation, Deep Learning

Laplacian Filters. Sobel Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters. Laplacian Filters

Support vector machines Lecture 4

Style-aware Mid-level Representation for Discovering Visual Connections in Space and Time

Machine Learning. Nonparametric Methods. Space of ML Problems. Todo. Histograms. Instance-Based Learning (aka non-parametric methods)

Mining Classification Knowledge

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas

Unsupervised Learning: K- Means & PCA

CSE 546 Final Exam, Autumn 2013

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

SATELLITE REMOTE SENSING

MRC: The Maximum Rejection Classifier for Pattern Detection. With Michael Elad, Renato Keshet

CMU-Q Lecture 24:

Click Prediction and Preference Ranking of RSS Feeds

Electric Fields and Equipotentials

Probabilistic modeling. The slides are closely adapted from Subhransu Maji s slides

CS 231A Section 1: Linear Algebra & Probability Review. Kevin Tang

CS 231A Section 1: Linear Algebra & Probability Review

1 Machine Learning Concepts (16 points)

Support Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar

Metric Embedding of Task-Specific Similarity. joint work with Trevor Darrell (MIT)

Chapter 3 Engineering Solutions. 3.4 and 3.5 Problem Presentation

The exam is closed book, closed notes except your one-page (two sides) or two-page (one side) crib sheet.

Mining of Massive Datasets Jure Leskovec, AnandRajaraman, Jeff Ullman Stanford University

6.867 Machine learning

Learning from Examples

Lecture 04 Image Filtering

Problem Session #5. EE368/CS232 Digital Image Processing

Neural networks and optimization

Linear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction

Support Vector Machine (SVM) and Kernel Methods

When Dictionary Learning Meets Classification

Day 6: Classification and Machine Learning

Chapter 18 Sampling Distribution Models

Linear Classifiers. Michael Collins. January 18, 2012

CSE546: SVMs, Dual Formula5on, and Kernels Winter 2012

Instance-based Learning CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2016

Pattern Recognition and Machine Learning. Perceptrons and Support Vector machines

Linear Classifiers as Pattern Detectors

Support Vector and Kernel Methods

Probabilistic Machine Learning. Industrial AI Lab.

Support Vector Machines

CS 179: LECTURE 16 MODEL COMPLEXITY, REGULARIZATION, AND CONVOLUTIONAL NETS

CS798: Selected topics in Machine Learning

CS 534: Computer Vision Segmentation III Statistical Nonparametric Methods for Segmentation

CSCI 315: Artificial Intelligence through Deep Learning

Introduction to Support Vector Machines

CS6375: Machine Learning Gautam Kunapuli. Support Vector Machines

Templates, Image Pyramids, and Filter Banks

Algorithms for Classification: The Basic Methods

Midterm Exam Solutions, Spring 2007

Anomaly Detection for the CERN Large Hadron Collider injection magnets

CS534 Machine Learning - Spring Final Exam

Introduction to Machine Learning

Transcription:

Automatic detection and of dipoles in large area SQUID magnetometry Lisa Qian December 4, Introduction. Scanning SQUID magnetometry Scanning SQUID magnetometry is a powerful tool for metrology of individual nanomagnets because of its incredible flux sensitivity (as low as electron spins) and non-invasive nature. With micron-sized pick-up loops, this technique has been used to study individual magnetic dipoles in a variety of systems such as magnetotactic bacteria, nanofabricated bar magnets, and naturally occurring nanomagnetic patches in complex oxide heterostructures. Figure a shows a cartoon of a scanning SQUID magnetometer over individual point dipoles. A DC magnetometry image of a silicon substrate containing individual magnetotactic bacteria is shown in Figure b, where the individual dipole orientations are clearly visible. Such an image is a convolution of the local magnetic field component perpendicular to the pick up loop with the point spread function defined by the pick up loop geometry. pickup loop 5μm 8. 6 6 μm mφ 8.5 Figure : Cartoon of a scanning SQUID imaging magnetic dipoles. A typical magnetometry image, which is a convolution of the sample s local magnetic field with the SQUID pick up loop shown in.

Using both fine and course motion scanners, we can cover a sample area as large as 5mm 5mm. Each scan can be as large as 3 3, and often contain many individual dipoles that may be clustered together. Once a scan is taken, we often need to analyze properties of each individual dipole such as their dipole moments. Current practice is to identify each dipole by eye, manually crop the image, and send the cropped image to an analysis program. In crowded images, selection of dipoles is a slow and monotonous process. In addition to finding each dipole, the crop window, or bounding box, for each dipole is also of importance. Since magnetic fields are long range, we want to make the bounding box as large as possible without picking up magnetic signals from neighboring dipoles or uneven background. The goal of this project is to automatically detect magnetic dipoles given a large area magnetometry scan and to determine the appropriately sized bounding box. Supervised learning The idea is to go through the entire large area scan with a sliding box of fixed dimensions. Each image is an example and is classified as either a part of a dipole () or not (). Because stronger dipoles extend across more space than weaker dipoles, many boxes classified as () might be needed to cover an entire dipole. At the end, neighboring boxes are merged to form the bounding box for a dipole.. Challenges Classification and feature design for this project was challenging for several reasons. First, many magnetometry scans have slowly varying background signals that can be on the order of the dipolar signals themselves. But because the dipolar fields themselves are long range, it is often challenging to properly remove background fields while preserving the structure of the dipoles. Second, the strengths of the dipoles we measure vary across roughly four orders of magnitude (from 5 µ B to 9 µ B ). Thus the spatial extent of the dipoles also vary greatly, ranging from approximately 5 5 for the weakest dipoles to 35 35 for the strongest ones. Initially I used a very large feature vector to cover all the entire range of pixel intensity, along with sliding boxes of varying sizes. Both of these ideas proved to be not computationally feasible.. Feature selection and SVM The background problem was solved by using the gradient of the large area scan instead of the magnetometry image itself. Figure shows a magnetometry image with particularly strong background fields and how simply taking the gradient nearly completely removes the background. To obtain training/test examples, I take the gradient of the large area magnetometry scan and go through this gradient image with a sliding window as described previously. The feature vector for each image obtained through with the sliding window is a histogram of pixel intensity of that image. I discretize the feature vector linearly from

-mφ / to +mφ /mum in increments of.mφ /mum. For a given training/test example, I count up the number of pixels in that example corresponding to each discretization level. This results in a length sparse feature vector. I used an SVM to train and classify. Figure : Large area magnetometry scan, without background removal. The gradient of, without additional background filtering..3 Training and test examples I collected training samples by looking at a total of seven large area scans. These scans contain a total of 93 dipoles. To get a good variety of dipoles, I chose large area scans taken at different heights and with levels of crowdedness. A large scan image is usually around 3 pixels 4 pixels and covers about 3 3. The size of the sliding window is set to 6 pixels, which is the size of the smallest dipole I found amongst the seven large area scans I looked at. For training examples, I skipped by 5 pixels between each image. For test data, I skipped by pixels between images. It is important to note that since dipolar fields are long range, there is no precise way of determining when a dipole ends and draw the appropriate bounding box. In my work here, I labeled an example as when any stray dipolar fields are on the order of the noise level of the background. 3 3. Results Leave one out cross-validation Because I had a limited sample size (seven large area scans), I used LOOCV to train, test and compare my models. Using an SVM with a polynomial kernel gives 8% precision and 83% recall. An SVM with a Gaussian kernel gives 85% precision and 87% recall. Only four false positives were found among all 7 large are scans. These results are very satisfactory - the dipoles that were missed by the classifier were very weak and right at the noise level. Figure 3 a and c show the results of the model on two large area scans. Each of these scans were classified using an SVM with Gaussian kernel trained on the other six scans. Further improvements could be made by including more positive training data. 3

3. Merging to form bounding box The final step is to merge neighboring test images that have been classified as (part of a dipole) to form the final bounding box for the dipoles. Figure 3 b and d show the final results. While this process works well on scans with sparse dipoles, it does not work well for crowded scans (d). This is because the sliding window is currently quite large and can cover more than one dipole at a time in crowded images. While this problem will be alleviated using a smaller window size (breaking up the scan into more and smaller test images), it will always be a problem because of the long range nature of the dipolar fields. When the dipoles are crowded, it becomes impossible to separate out individual dipoles from a cluster. In these cases, I would eventually like to identify the cluster as a separate classification. 5.5.5 5.5.5 5 5 5.5.5 5.5.5 3 mφ 3 mφ.5.5 5 5.5.5.5.5 5 5.5.5 5 5 mφ 5 5 mφ (c) (d) Figure 3: Results of SVM classification on a large area magnetometry scan. After merging to form one bounding box for each dipole. (c) - (d)the same for another, more crowded scan. 4

3.3 Conclusions In this project, I successfully trained an SVM classifier to automatically detect dipoles in a large area SQUID magnetometry image. Using a Gaussian kernel yields slightly better results over a polynomial kernel, with a precision of 85% and recall of 87%. These results are satisfactory and can be improved by including more positive training examples and reducing the sliding window size. Significant improvements still need to be made to draw appropriate bounding boxes in crowded images. 5