metrics for supernova classification · may need a unique metric worth running an anomaly detection...

16
Metrics for Supernova Classification Michelle Lochner PLASTICC Workshop 2017

Upload: others

Post on 04-Oct-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Metrics for Supernova Classification

Michelle LochnerPLASTICC Workshop 2017

Page 2: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

LSST Supernova Survey

LSST

Over 10 years, LSST will detect ~100 000 well-characterised supernovae over the entire southern sky

Page 3: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Supernova types

Conley et al. 2011

Type Ia supernovae Core collapse supernovae

CosmologySupernova astrophysics

Page 4: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Pipeline

Data

https://github.com/LSSTDESC/snmachine

SALT2 Parametric models Wavelets

NB BDTANNSVMKNN

Classified SNe

Page 5: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Evaluating algorithms

Page 6: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Machine Learning Metrics

● Precision → TP/(TP + FP)

● Recall → TP/(TP + FN)

● Accuracy → (TP + TF) / N

● ...

Page 7: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Supernova Photometric Challenge

Kessler et al. (2010)

Rewards completeness Penalises contamination

Page 8: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Receiver Operator Characteristic (ROC) curves

AUC = Area Under Curve

Page 9: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Where does your science live?

Cosmology?

Rare object discovery? Cosmic census?

Page 10: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Advantages of ROC Curves

● Gives you the full picture without arbitrary thresholding

● Allows the fair comparison of different algorithms for different possible science cases

● Rewards completeness while penalising false positives

● Robust to class imbalances

Page 11: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Connecting Metrics to Science

● ROC curves great for comparing algorithms but does a higher AUC mean better science?

Page 12: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Connecting Metrics to Science

● ROC curves great for comparing algorithms but does a higher AUC mean better science?

● Different science cases may need different metrics

Page 13: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Connecting Metrics to Science

● ROC curves great for comparing algorithms but does a higher AUC mean better science?

● Different science cases may need different metrics● Simple metric: for a given purity, how many

supernovae do I find?

Page 14: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Connecting Metrics to Science

● ROC curves great for comparing algorithms but does a higher AUC mean better science?

● Different science cases may need different metrics● Simple metric: for a given purity, how many

supernovae do I find?● Still some subtleties connecting to science

Page 15: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Anomaly detection

● The exciting prospect of discovering new objects with LSST!

● Most algorithms are supervised, anomaly detection is technically an unsupervised problem

● May need a unique metric

● Worth running an anomaly detection component of the challenge?

Page 16: Metrics for Supernova Classification · May need a unique metric Worth running an anomaly detection component of the challenge? Discussion points How do we connect metrics to science?

Discussion points

● How do we connect metrics to science?● Is it even possible to find “one metric to rule them

all”?● Should there be multiple metrics for different science

cases/ different challenges?● How do we (easily) test that the metrics we come

up with are reasonable?