How to forecast solar flares?

Slides:



Advertisements
Similar presentations
ECG Signal processing (2)
Advertisements

An Introduction of Support Vector Machine
Support Vector Machines
11/26/081 AUTOMATIC SOLAR ACTIVITY DETECTION BASED ON IMAGES FROM HSOS NAOC, HSOS YANG Xiao, LIN GangHua
Estimating the magnetic energy in solar magnetic configurations Stéphane Régnier Reconnection seminar on Thursday 15 December 2005.
Nonlinearity of the force-free parameter over active regions. M.Hagino and T.Sakurai National Astronomical Observatory of Japan, Solar Observatory.
H.N. Wang 1 , H. He 1, X. Huang 1, Z. L. Du 1 L. Y. Zhang 1 and Y. M. Cui 2 L. Y. Zhang 1 and Y. M. Cui 2 1 National Astronomical Observatories 2 National.
Study of Magnetic Helicity Injection in the Active Region NOAA Associated with the X-class Flare of 2011 February 15 Sung-Hong Park 1, K. Cho 1,
The Disputed Federalist Papers : SVM Feature Selection via Concave Minimization Glenn Fung and Olvi L. Mangasarian CSNA 2002 June 13-16, 2002 Madison,
Software Quality Ranking: Bringing Order to Software Modules in Testing Fei Xing Michael R. Lyu Ping Guo.
SUPPORT VECTOR MACHINES PRESENTED BY MUTHAPPA. Introduction Support Vector Machines(SVMs) are supervised learning models with associated learning algorithms.
Empirical Forecasting of CMEs from a Free-Energy Proxy: Performance and Extension to HMI David Falconer, Ron Moore, Abdulnasser F. Barghouty, & Igor Khazanov.
Optimizing F-Measure with Support Vector Machines David R. Musicant Vipin Kumar Aysel Ozgur FLAIRS 2003 Tuesday, May 13, 2003 Carleton College.
Free Magnetic Energy in Solar Active Regions above the Minimum-Energy Relaxed State (Regnier, S., Priest, E.R ApJ) Use magnetic field extrapolations.
Study of magnetic helicity in solar active regions: For a better understanding of solar flares Sung-Hong Park Center for Solar-Terrestrial Research New.
Sung-Hong Park Space Weather Research Laboratory New Jersey Institute of Technology Study of Magnetic Helicity and Its Relationship with Solar Activities:
Identifying Computer Graphics Using HSV Model And Statistical Moments Of Characteristic Functions Xiao Cai, Yuewen Wang.
Efficient Model Selection for Support Vector Machines
Comparison on Calculated Helicity Parameters at Different Observing Sites Haiqing Xu (NAOC) Collaborators: Hongqi, Zhang, NAOC Kirill Kuzanyan, IZMIRAN,
Prediction model building and feature selection with SVM in breast cancer diagnosis Cheng-Lung Huang, Hung-Chang Liao, Mu- Chen Chen Expert Systems with.
GA-Based Feature Selection and Parameter Optimization for Support Vector Machine Cheng-Lung Huang, Chieh-Jen Wang Expert Systems with Applications, Volume.
The Disputed Federalist Papers: Resolution via Support Vector Machine Feature Selection Olvi Mangasarian UW Madison & UCSD La Jolla Glenn Fung Amazon Inc.,
Processing of large document collections Part 3 (Evaluation of text classifiers, term selection) Helena Ahonen-Myka Spring 2006.
Practical Calculation of Magnetic Energy and Relative Magnetic Helicity Budgets in Solar Active Regions Manolis K. Georgoulis Research Center for Astronomy.
Beyond Sliding Windows: Object Localization by Efficient Subwindow Search The best paper prize at CVPR 2008.
Efficient Subwindow Search: A Branch and Bound Framework for Object Localization ‘PAMI09 Beyond Sliding Windows: Object Localization by Efficient Subwindow.
Azimuth disambiguation of solar vector magnetograms M. K. Georgoulis JHU/APL Johns Hopkins Rd., Laurel, MD 20723, USA Ambiguity Workshop Boulder,
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
Enhanced Regulatory Sequence Prediction Using Gapped k-mer Features 王荣 14S
1 Yongliang Song & Mei Zhang (National Astronomical Observatory of China) The effect of non-radial magnetic field on measuring helicity transfer rate.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
A Kernel Approach for Learning From Almost Orthogonal Pattern * CIS 525 Class Presentation Professor: Slobodan Vucetic Presenter: Yilian Qin * B. Scholkopf.
The Helioseismic and Magnetic Imager (HMI) on NASA’s Solar Dynamics Observatory (SDO) has continuously measured the vector magnetic field, intensity, and.
Non-separable SVM's, and non-linear classification using kernels Jakob Verbeek December 16, 2011 Course website:
Constructing a Predictor to Identify Drug and Adverse Event Pairs
Principal Component Analysis (PCA)
Support Vector Machines
Data Mining Introduction to Classification using Linear Classifiers
CS 9633 Machine Learning Support Vector Machines
D. Shaun Bloomfield 1,2, K. Domijan 3, P. A. Higgins 2, P. T
Name: Sushmita Laila Khan Affiliation: Georgia Southern University
Support Feature Machine for DNA microarray data
Guillaume-Alexandre Bilodeau
Introduction to Machine Learning
Rule Induction for Classification Using
D. Shaun Bloomfield 1,2, K. Domijan 3, P. A. Higgins 2, P. T
Mixture of SVMs for Face Class Modeling
Effects of Dipole Tilt Angle on Geomagnetic Activities
Basic machine learning background with Python scikit-learn
An Introduction to Support Vector Machines
Students: Meiling He Advisor: Prof. Brain Armstrong
Hidden Markov Models Part 2: Algorithms
Pattern Recognition and Image Analysis
COSC 4335: Other Classification Techniques
iSRD Spam Review Detection with Imbalanced Data Distributions
Implementing AdaBoost
Model Evaluation and Selection
Shih-Wei Lin, Kuo-Ching Ying, Shih-Chieh Chen, Zne-Jung Lee
Support Vector Machines
Support vector machines
Magnetic Configuration and Non-potentiality of NOAA AR10486
Mathematical Foundations of BME
Emerging Active Regions: turbulent state in the photosphere
A Novel Smoke Detection Method Using Support Vector Machine
Linear Discrimination
University of Wisconsin - Madison
Multiple DAGs Learning with Non-negative Matrix Factorization
Support Vector Machines 2
Outlines Introduction & Objectives Methodology & Workflow
Introduction to Machine Learning
Presentation transcript:

How to forecast solar flares? Review of the paper “Solar Flare Prediction Using SDO/HMI Vector Magnetic Field Data with a Machine-learning Algorithm” by Bobra and Couvidat, 2015

Outline Introduction Step 0. Support Vector Machine Classifier Step 1. Data Processing and Feature Extraction Step 2. Definition of the Flaring and non-Flaring Classes Step 3. Feature Selection Step 4. Classifier Tuning and Performance Metrics Results Conclusion

Introduction The prediction of strong solar flares is one of the key questions of Solar Physics Besides many attempts, the operational predictions are still mainly done based on experts’ opinion and experience

Introduction Currently we are receiving tremendous amounts of information about the Sun and its active regions. The prediction problem is the Big Data problem. The magnetic field represents very valuable data for the forecasts. The nonpotential magnetic field is the only reservoir to store the large energies released during the flares The SDO/HMI provides a routine coverage of the vector magnetic field in active regions since 2010

Step 0. Machine-Learning Algorithms. Suppose we are calculating two AR features: total magnetic flux and PIL length The calculations are done each hour for each active region If the M1.0 class solar flare occurs within 24h after the calculation, we define the region as flaring (1), and non- flaring otherwise (0) {Active Region} -> {(Total flux, PIL length) , flaring/non- flaring} Flare No flares Total magnetic flux PIL length

Step 0. Machine-Learning Algorithms. Suppose we are calculating two AR features: total magnetic flux and PIL length The calculations are done each hour for each active region If the M1.0 class solar flare occurs within 24h after the calculation, we define the region as flaring (1), and non- flaring otherwise (0) {Active Region} -> {(Total flux, PIL length) , flaring/non- flaring} One more case {(Total flux, PIL length) , ???} If we correctly set up the division, we can easily classify the new sample. Flare No flares Total magnetic flux PIL length New case

Step 0. Support Vector Machines (SVMs) The “widest street” approach: the margin should be maximized 𝐿= 1 2 𝑤 2 − 𝛼 𝑖 𝑦 𝑖 𝑤 ∗ 𝑥 𝑖 +𝑏 −1 →𝑚𝑖𝑛 After the minimization, purely quadratic optimization problem depending on the sums of inner products 𝐿= 𝛼 𝑖 − 1 2 𝛼 𝑖 𝛼 𝑗 𝑦 𝑖 𝑦 𝑗 ( 𝑥 𝑖 ∗ 𝑥 𝑗 ) 𝑦 𝑖 ( 𝑤 ∗ 𝑥 𝑖 +𝑏)≥1 for any positive/negative sample

Step 0. SVMs. Kernel Trick. One can remap the data to the higher dimensional space For example, we have introduced 𝜑 𝑅 𝑛 → 𝑅 𝑛+1 :𝜑 𝑥 =( 𝑥 , 𝑥 2 ) Because the SVMs depend on ( 𝑥 , 𝑦 ), we now consider 𝜑 𝑥 ,𝜑 𝑦 =𝐾 𝑥 , 𝑦 The Kernel Function 𝐾 𝑥 , 𝑦 is all we need!

Step 0. SVMs Different kernels are suitable for different datasets/problems Each kernel should be tuned for the problem (has its own parameters) Typical kernel functions:

Step 1. Data processing and Feature Extraction Flare selection: GOES class of M1.0 or higher Location within 68O from central meridian Features of the flare <-> features of its parental active region Features (from vector magnetograms): Space-weather HMI Active Region Patches (SHARP) parameters The active region is already traced with 12min time cadence

Step 1. Data processing and Feature Extraction Parameters calculated by (B&C) for each Active Region with 12 min time cadence Parameters (descriptors) form a vector representing the Active Region at the moment of calculation

Step 2. Definition of Classes Structure of each case (active region at each time moment): Vector of parameters/descriptors + its class (1 for potentially flaring active regions, 0 otherwise) In (B&C), the following definitions of classes were tested: “Operational”: Positive if the flare of M1.0 class or higher occurs exactly after 24h in the region after the considered time moment Negative if there is no M1.0 class or higher within the next 24 hours “Segmented”: Negative if there are no flares in the region within 48 hours before and after the considered time moments Total: 303 positive examples + 5000 randomly selected negative samples

Step 2. Definition of Classes

Step 3. Feature selection Generally, the SVM algorithm complexity can be estimated as O(N2*M), where N – number of cases, M – number of characteristics (N >> M) It is reasonable to remove non-discriminative characteristics. One of the ways is to use Fisher ranking score (F-score, B&C): 𝐹(𝑖)= ( 𝑥 𝑖 + − 𝑥 𝑖 ) 2 + ( 𝑥 𝑖 − − 𝑥 𝑖 ) 2 1 𝑛 + −1 𝑘=1 𝑛 + ( 𝑥 𝑘,𝑖 + − 𝑥 𝑖 + ) 2 + 1 𝑛 − −1 𝑘=1 𝑛 − ( 𝑥 𝑘,𝑖 − − 𝑥 𝑖 − ) 2 The F-score was used to select the most “useful” characteristics in (B&C). Most of the top characteristics have “vector” nature

Step 4. Algorithms and performance metrics First of all, we need to simulate the real-time data. For such purpose, the total set of cases is divided into the training test and test set (B&C) with the approximate ratio of 70% to 30%. The question of dataset shuffling is important. In (B&C), the totally random shuffling approach was used. The SVM classifier with RBF kernel was used for the classification (2 parameters, weights of the classes)

Step 4. Performance Metrics How to calculate the performance of the SVM classifier? Basically, each classifier produces the following numbers: TP – True Positives (number of positive cases predicted as positive) TN – True Negatives (number of negative cases predicted as negative) FP – False Positives (number of negative cases predicted as positive) FN – False Negatives (number of positive cases predicted as negative) Our aim is simultaneously minimize FP and FN values. The question is, how?

Step 4. Performance Metrics Various metrics may be constructed from the derived numbers: precision= 𝑇𝑃 𝑇𝑃+𝐹𝑃 recall= 𝑇𝑃 𝑇𝑃+𝐹𝑁 𝐻𝑆𝑆 1 = 𝑇𝑃+𝑇𝑁−𝑁 𝑃 , 𝑃=𝑇𝑃+𝐹𝑁, 𝑁=𝑇𝑁+𝐹𝑃 Scales from (-inf, 1). 0 corresponds to purely negative forecast. 𝐻𝑆𝑆 2 = 𝑇𝑃+𝑇𝑁−𝐸 𝑃+𝑁−𝐸 , 𝐸= 𝑇𝑃+𝐹𝑃 ∗ 𝑇𝑃+𝐹𝑁 + 𝐹𝑃+𝑇𝑁 ∗(𝐹𝑁+𝑇𝑁) 𝑃+𝑁 Measures improvement of the forecast over the random forecast. Used by SWPC. And many others…

Step 4. Performance Metrics All the previously mentioned metrics depend on the class-imbalance ratio. There is the True Skill Statistics (TSS, Hansen-Kuipers skill score, Peirce skill score) which does not suffer from this ratio: TSS= 𝑇𝑃 𝑇𝑃+𝐹𝑁 − 𝐹𝑃 𝐹𝑃+𝑇𝑁 This score was tested in (B&C) among others.

Results (B&C)

Results (B&C) The use of HMI vector magnetograms provides a much larger database and a more uniform quality allowed by space-based observations (in comparison with previous works using ground-based observations) The flare prediction is a strongly imbalanced problem. Calculation of TSS allows to compare different studies. The obtained TSS are larger than in the papers studied. However, results are metrics-dependent: maximizing some scores may result in lower values for other metrics. only the 4 parameters with the highest F-score—the total unsigned current helicity, total magnitude of the Lorentz force, total photospheric magnetic free energy density, and total unsigned vertical current—gives roughly the same TSS score as the top 13 combined.

In Conclusion… Current NICT (National Institute of Information and Communications Technology) space weather forecasting center TSS scores (2000-2015, Nishizuka et al., 2017): TSS = 0.21 for X-class flares TSS = 0.50 for M- and X-class flares Solar Influences Data Center of the Royal Observatory of Belgium TSS scores (2004-2012): TSS = 0.34 for M- and X-class flares The TSS scores obtained via Machine-Learning algorithms are significantly higher than ones based on the opinion of experts. However, the algorithms are needed to be applied accurately and tested on the real-time data.

Thank you for your attention!