recent achievements of the collaboratory for the study of...

1
Recent Achievements of the Collaboratory for the Study of Earthquake Predictability Establish rigorous procedures for registering and evaluating earthquake forecasting and prediction experiments. Construct community standards and protocols for comparative testing of earthquake predictions. Develop an infrastructure that allows groups of researchers to participate in prediction experiments. Provide access to authorized data sets and monitoring products for calibrating and testing earthquake prediction algorithms. Accommodate experiments involving fault systems in different geographic and tectonic environments. CSEP Software is released under GPL licenses and is distributed to other earthquake forecast testing facilities outside of California. We use Yellowdog Updater, Modified (YUM) repository distribution and Kernel-based Virtual Machine (KVM) images to keep all testing centers automatically up-to-date to the current version of testing center codes. We host the XML-definitions, the software core, and the scripts for testing regions. Objectives Promote rigorous research on earthquake predictability through the SCEC program and its global partnerships. Help the responsible government agencies assess the feasibility of earthquake prediction and the performance of proposed prediction algorithms. Reduce the controversy surrounding earthquake prediction through a collaboratory infrastructure to support a wide range of scientific prediction experiments. Goals D. D. Jackson 1 , M. Werner 2 , A. Christophersen 5 , P. Maechling 3 , D. Rhoades 5 , D. Schorlemmer 3,4 , A. Strader 4 , J. Yu 3 , W. Marzocchi 6 , T.H. Jordan 3 , and the CSEP Working Group 1 Department of Earth, Planetary, and Space Science, UCLA, Los Angeles, USA. 2 School of Earth Sciences and Cabot Institute, University of Bristol, UK. 3 Southern California Earthquake Center, University of Southern California, Los Angeles, USA. 4 Department 2: Physics of the Earth, German Research Centre for Geosciences, Potsdam, Germany. 5 GNS Science, Lower Hutt, New Zealand. 6 INGV, Rome, Italy. Development organization CSEP Software Automatic: Catalog retrieval from authorized data source Forecast generation and evaluation Publishing of results CSEP software capabilities: Reproduce any CSEP experiment Reprocess forecasts and evaluations in batch mode Dataset archive: Raw and post-processed catalogs Input parameters for forecasts models, forecasts Evaluation tests results Mini CSEP Distribution For use by earthquake prediction researchers: To retrieve catalogs from authorized data sources To run evaluation tests for forecasts To create spatial forecast distribution maps http://northridge.usc.edu/trac/csep/wiki/MiniCSEP http://cseptesting.org CSEP Testing Centers Models Web Presentation Figure 1. As of April 2016, there is the total of 438 models under test in CSEP Global 0.1 Degree Testing Region SCEC Testing Center Results Please visit our web page http://cseptesting.org/results. GEAR1: The first version of the Global Earthquake Activity Rate Model is an optimized combination of a global smoothed seismicity model developed by Kagan and Jackson (2011) using shallow m w ≥ 5.767 historical seismicity from the CMT earthquake catalog, and a tectonic model (Bird and Kreemer, 2015) based on global strain rates from version 2.1 of the Global Strain Rate Map (SHIFT_GSRM2F). Using a loglinear combination of the two parent forecasts, the GEAR1 model was optimized using seismicity from 2005-2012. SHIFT_GSRM: The tectonic earthquake forecast was developed from the first version of the Global Strain Rate Model (GSRM). Strain rates, inferred from GPS velocities and plate tectonics, were converted to long-term seismic moment rates (Kreemer et al., 2003) based on the shear modulus and tectonic regions' seismogenic depths. Earthquake rates were derived using the corner magnitudes a n d s p e c t r a l s l o p e s o f t h e frequency/moment relation corresponding to each tectonic region. SHIFT_GSRM2F: GSRM2.1 is an update of the SHIFT_GSRM model, constrained by four times the number of geodetic velocities. Additionally, the grid resolution was increased, with six times the number of cells as SHIFT_GSRM. The number of (assumed) plates and blocks was doubled, to account for regions where rigidity can be assumed but GPS coverage is insufficient to adequately constrain block boundaries and deformation zones. Figure 1: Forecasted numbers of 6.0 m w ≤ 9.0 earthquakes in 0.1˚x0.1˚x70km bins from 1. October 2015 until 1. March 2017 for the GEAR1 (top), SHIFT_GSRM (center) and SHIFT_GSRM2F (bottom) forecasts. Red circles indicate earthquake locations during the evaluation period, and are scaled to indicate relative magnitudes. Within the Pacific subduction zone, forecasted seismicity rates are considerably greater for the GEAR1 model than for the GSRM models, consistent with the abundance of earthquakes occurring in this region. CSEP Comparative-Test Results The T- and W-tests indicate that both SHIFT_GSRM and SHIFT_GSRM2F can be rejected in favor of the GEAR1 model at the 0.05 significance level. This result is consistent with the information scores used to optimize the GEAR1 forecast from 2005-2012 (Bird et al., 2015), which indicate an optimal loglinear combination of 60% from the smoothed seismicity forecast, and 40% from the SHIFT_GSRM2F model. There were no observed anomalous differences in log-likelihood scores between the GEAR1 and GSRM models caused by specific events. Figure 2: T- and W-test results during the one-year evaluation period. The plots display the average information gain per earthquake of the GSRM models over the GEAR1 model. Average information gain is indicated by the circles, while the horizontal lines show the ranges of average information gain within the 0.05 significance level. The numbers above the circles show the number of earthquakes occurring during the testing period. Asterisks indicate that a forecast can be rejected in favor of GEAR1 according to the W-test. Both GSRM models can be rejected in favor of GEAR1. CSEP Consistency-Test Results N-Test (Number of Earthquakes) L-Test (Data Consistency) All forecasts pass the L-test, indicating overall consistency between observed and forecasted earthquake distributions. Although the L-test provides information about the consistency of the number of earthquakes, magnitude and spatial distributions, the total log-likelihood score (L) is most influenced by the number of forecasted earthquakes. Therefore, little information beyond the N-test results is apparent from the L-test. CL-Test (Data Consistency with Normalized Earthquake Count) The number of earthquakes forecasted by all three models is consistent with observed seismicity for the one-year prospective evaluation period. As the duration of the evaluation period increases, forecasted earthquake numbers approach the observed earthquake count. M-Test (Magnitude Distribution) S-Test (Spatial Distribution) All of the forecasts fail the S-test, which evaluates the consistency of the forecasts' spatial distributions with that of observed seismicity. For both GSRM forecasts, the log-likelihood score from observed earthquakes falls below the range considered consistent with the forecasts. In the case of the GEAR1 forecast, the observed log-likelihood score exceeds the forecasted range. Because in both cases the forecasted and observed log-likelihood scores are inconsistent, the S-test (as well as all likelihood consistency tests) were conducted as two-tailed tests. Figure 3: CSEP consistency test results for the one-year prospective evaluation period. The squares indicate observed numbers of earthquakes (N-test) or log-likelihood scores (all other tests). Green squares indicate the observed earthquake number or log-likelihood score is consistent with what would be simulated from the forecasts. The horizontal lines indicate the range of simulated earthquake numbers or log-likelihood scores within the 0.05 significance level. All three forecasts pass the M-test during the one-year evaluation period, indicating that the magnitude distribution of observed earthquakes during the one-year evaluation period is consistent with the forecasted magnitude distribution. The forecasted magnitude distributions were obtained from the union of various tapered Gutenberg-Richter distributions for different tectonic regions. The conditional likelihood (CL) test compares the spatial-magnitude distribution of observed versus synthetic earthquake catalogs by normalizing the number of forecasted events to be equivalent with the observed number. During the evaluation period, only the SHIFT_GSRM forecast passes the CL- test. The other two forecasts fail the CL- test due to inconsistencies in the forecasted versus observed spatial earthquake distribution (see S-test results). Testing Fault-based Earthquake Models Earthquakes on faults play an important role in hazard studies, yet models of finite ruptures on faults are difficult to test for several reasons. Fault ruptures are subjective and lack systematic reporting in authoritative catalogs. Surface observations may not reveal the true extent of subsurface rupture, and Surface ruptures are relatively rare, and earthquakes observed seismically and by surface rupture are too rare to provide adequate statistics. Earthquakes may occur off of but near previously mapped faults, making forecasts difficult to state in a testable fashion. Here we explore ways of testing Uniform California Earthquake Rupture Forecasts (UCERF) both retrospectively and prospectively. Key features include Specifying polygons about fault sections, such that earthquakes not precisely on the fault but within the polygons affect the stress and recurrence properties on the fault, and Determining rupture extent by seismically observed aftershock locations, and Defining goodness of fit by whether rupture stops or continues through test polygons along faults with prescribed conditional probability that rupture will stop within a polygon if it enters. We test probabilities based on hypotheses (e.g. segmentation) assumed in the recurrence model against those based on a null hypothesis that earthquakes occur on an infinitely long fault with a length distribution based on Gutenberg-Richter size distribution and simple magnitude-length scaling. . Figure 4: Map of seismio-tectonic zones used in the 1995 WGCEP forecast, and M3+ aftershocks within 30 days of the four subsequent M6+ earthquakes. Probabilities were assigned to each of the zones in 1995. The figure shows how rupture extent can be defined using aftershock epicenters. Figure 6: Conditional stopping probabilities as a function of distance southward from Mendocino for the San Andreas Fault (adapted from Figure 29 of Field et al., 2014). Red curve is based on modeled earthquake ruptures in the UCERF3 report, while the blue curve illustrates the null hypothesis. Both are evaluated on about 5 km sub-segments. Figure 5: Conditional stopping probabilities for 20 km fault segment boundary zones on the southern San Andreas fault. Blue bars are calculated from segment rupture forecasts in UCERF2 (Field et al, 2009); horizontal red line is for null hypothesis of randomly placed self-similar GR ruptures on infinitely long fault. Present efforts include mapping rupture zones as illustrated in Figure 4 onto the stopping probabilities in Figure 3. The UCERF3 model implies relevant earthquakes along the San Andreas fault at a rate of about 0.15 per year. A definitive prospective test might take several decades, but likelihood scores from just one large earthquake might falsify the UCERF3 fault model if it ruptured across more than one of the “spikes” on the red curve.

Upload: others

Post on 25-Aug-2021

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Recent Achievements of the Collaboratory for the Study of ...hypocenter.usc.edu/research/SSA/CSEPTestingCenterPoster...CMT earthquake catalog, and a tectonic model (Bird and Kreemer,

Recent Achievements of the Collaboratory for the Study of Earthquake Predictability

● Establish rigorous procedures for registering and evaluatingearthquake forecasting and prediction experiments.

● Construct community standards and protocols for comparativetesting of earthquake predictions.

● Develop an infrastructure that allows groups of researchers toparticipate in prediction experiments.

● Provide access to authorized data sets and monitoringproducts for calibrating and testing earthquake predictionalgorithms.

● Accommodate experiments involving fault systems in differentgeographic and tectonic environments.

CSEP Software is released under GPL licenses and isdistributed to other earthquake forecast testing facilities outside ofCalifornia. We use Yellowdog Updater, Modified (YUM) repositorydistribution and Kernel-based Virtual Machine (KVM) images tokeep all testing centers automatically up-to-date to the currentversion of testing center codes. We host the XML-definitions, thesoftware core, and the scripts for testing regions.

Objectives● Promote rigorous research on earthquake predictability

through the SCEC program and its global partnerships.● Help the responsible government agencies assess the

feasibility of earthquake prediction and the performance ofproposed prediction algorithms.

● Reduce the controversy surrounding earthquake predictionthrough a collaboratory infrastructure to support a wide rangeof scientific prediction experiments.

Goals

D. D. Jackson1, M. Werner2, A. Christophersen5, P. Maechling3, D. Rhoades5, D. Schorlemmer3,4, A. Strader4, J. Yu3, W. Marzocchi6, T.H. Jordan3, and the CSEP Working Group1 Department of Earth, Planetary, and Space Science, UCLA, Los Angeles, USA.2 School of Earth Sciences and Cabot Institute, University of Bristol, UK.3 Southern California Earthquake Center, University of Southern California, Los Angeles, USA.4 Department 2: Physics of the Earth, German Research Centre for Geosciences, Potsdam, Germany.5 GNS Science, Lower Hutt, New Zealand.6 INGV, Rome, Italy.

Development organization

CSEP Software Automatic:● Catalog retrieval from authorized data source● Forecast generation and evaluation● Publishing of resultsCSEP software capabilities:● Reproduce any CSEP experiment● Reprocess forecasts and evaluations in batch mode Dataset archive:● Raw and post-processed catalogs● Input parameters for forecasts models, forecasts● Evaluation tests results

Mini CSEP Distribution For use by earthquake prediction researchers:● To retrieve catalogs from authorized data sources● To run evaluation tests for forecasts● To create spatial forecast distribution maps http://northridge.usc.edu/trac/csep/wiki/MiniCSEP

http://cseptesting.org

CSEP Testing Centers Models

Web Presentation

Figure 1. As of April 2016, there is the total of 438 models under test in CSEP

Global 0.1 Degree Testing Region

SCEC Testing Center Results

Please visit our web page http://cseptesting.org/results.

GEAR1: The first version of the GlobalEarthquake Activity Rate Model is anoptimized combination of a globalsmoothed seismicity model developed byKagan and Jackson (2011) using shallowmw ≥ 5.767 historical seismicity from theCMT earthquake catalog, and a tectonicmodel (Bird and Kreemer, 2015) based onglobal strain rates from version 2.1 of theG l o b a l S t r a i n R a t e M a p(SHIFT_GSRM2F). Using a loglinearcombination of the two parent forecasts,the GEAR1 model was optimized usingseismicity from 2005-2012.

SHIFT_GSRM: The tectonic earthquakeforecast was developed from the firstversion of the Global Strain Rate Model(GSRM). Strain rates, inferred from GPSvelocities and plate tectonics, wereconverted to long-term seismic momentrates (Kreemer et al., 2003) based on theshear modulus and tectonic regions'seismogenic depths. Earthquake rateswere derived using the corner magnitudes a n d s p e c t r a l s l o p e s o f t h efrequency/moment relation correspondingto each tectonic region.

SHIFT_GSRM2F: GSRM2.1 is anupdate of the SHIFT_GSRM model,constrained by four times the number ofgeodetic velocities. Additionally, the gridresolution was increased, with six timesthe number of cells as SHIFT_GSRM. Thenumber of (assumed) plates and blockswas doubled, to account for regions whererigidity can be assumed but GPScoverage is insufficient to adequatelyc o n s t r a i n b l o c k b o u n d a r i e s a n ddeformation zones.

Figure 1: Forecasted numbers of 6.0 ≤ mw ≤ 9.0 earthquakes in0.1˚x0.1˚x70km bins from 1. October 2015 until 1. March 2017 for theGEAR1 (top), SHIFT_GSRM (center) and SHIFT_GSRM2F (bottom)forecasts. Red circles indicate earthquake locations during the evaluation period, and are scaled to indicate relative magnitudes. Within the Pacificsubduction zone, forecasted seismicity rates are considerably greater forthe GEAR1 model than for the GSRM models, consistent with theabundance of earthquakes occurring in this region.

CSEP Comparative-Test Results

The T- and W-tests indicate that both SHIFT_GSRM and SHIFT_GSRM2F can be rejected in favor of the GEAR1 model at the 0.05 significance level. This result is consistent with the informationscores used to optimize the GEAR1 forecast from 2005-2012 (Bird et al., 2015), which indicate anoptimal loglinear combination of 60% from the smoothed seismicity forecast, and 40% from theSHIFT_GSRM2F model. There were no observed anomalous differences in log-likelihood scoresbetween the GEAR1 and GSRM models caused by specific events.

Figure 2: T- and W-test results during the one-year evaluation period. The plotsdisplay the average information gain per earthquake of the GSRM models over theGEAR1 model. Average information gain is indicated by the circles, while thehorizontal lines show the ranges of average information gain within the 0.05significance level. The numbers above the circles show the number of earthquakesoccurring during the testing period. Asterisks indicate that a forecast can be rejectedin favor of GEAR1 according to the W-test. Both GSRM models can be rejected infavor of GEAR1.

CSEP Consistency-Test ResultsN-Test (Number of Earthquakes)

L-Test (Data Consistency) All forecasts pass the L-test, indicatingoverall consistency between observedand forecasted earthquake distributions.Although the L-test provides informationabout the consistency of the number ofearthquakes, magnitude and spatialdistributions, the total log-likelihood score(L) is most influenced by the number offorecasted earthquakes. Therefore, littleinformation beyond the N-test results isapparent from the L-test.

CL-Test (Data Consistency withNormalized Earthquake Count)

The number of earthquakes forecasted by all three models is consistent withobserved seismicity for the one-yearprospective evaluation period. As theduration of the evaluation periodincreases, forecasted earthquakenumbers approach the observedearthquake count.

M-Test (Magnitude Distribution)

S-Test (Spatial Distribution) All of the forecasts fail the S-test, whichevaluates the consistency of theforecasts' spatial distributions with thatof observed seismicity. For both GSRMforecasts, the log-likelihood score fromobserved earthquakes falls below therange considered consistent with theforecasts. In the case of the GEAR1forecast, the observed log-likelihoodscore exceeds the forecasted range.Because in both cases the forecastedand observed log-likelihood scores areinconsistent, the S-test (as well as alllikelihood consistency tests) wereconducted as two-tailed tests.Figure 3: CSEP consistency test results for the

one-year prospective evaluation period. Thesquares indicate observed numbers of earthquakes (N-test) or log-likelihood scores (all other tests).Green squares indicate the observed earthquakenumber or log-likelihood score is consistent withwhat would be simulated from the forecasts. Thehorizontal lines indicate the range of simulatedearthquake numbers or log-likelihood scores withinthe 0.05 significance level.

All three forecasts pass the M-testduring the one-year evaluation period,i n d i c a t i n g t h a t t h e m a g n i t u d edistribution of observed earthquakesduring the one-year evaluation period is cons is ten t w i th the fo recas tedmagnitude distribution. The forecastedmagnitude distributions were obtainedfrom the union of various taperedGutenberg-Richter distributions fordifferent tectonic regions.

The conditional likelihood (CL) testcompares the spat ia l -magni tudedistribution of observed versus synthetic earthquake catalogs by normalizing thenumber of forecasted events to beequivalent with the observed number.During the evaluation period, only theSHIFT_GSRM forecast passes the CL-test. The other two forecasts fail the CL-test due to inconsistencies in theforecasted versus observed spatialearthquake distribution (see S-testresults).

Testing Fault-based Earthquake Models

Earthquakes on faults play an important role in hazard studies, yet models of finiteruptures on faults are difficult to test for several reasons.

● Fault ruptures are subjective and lack systematic reporting in authoritative catalogs.● Surface observations may not reveal the true extent of subsurface rupture, and● Surface ruptures are relatively rare, and earthquakes observed seismically and bysurface rupture are too rare to provide adequate statistics. ● Earthquakes may occur off of but near previously mapped faults, making forecastsdifficult to state in a testable fashion.

Here we explore ways of testing Uniform California Earthquake Rupture Forecasts(UCERF) both retrospectively and prospectively.

Key features include● Specifying polygons about fault sections, such that earthquakes not precisely on the faultbut within the polygons affect the stress and recurrence properties on the fault, and● Determining rupture extent by seismically observed aftershock locations, and ● Defining goodness of fit by whether rupture stops or continues through test polygonsalong faults with prescribed conditional probability that rupture will stop within a polygon ifit enters.

We test probabilities based on hypotheses (e.g. segmentation) assumed in the recurrencemodel against those based on a null hypothesis that earthquakes occur on an infinitelylong fault with a length distribution based on Gutenberg-Richter size distribution andsimple magnitude-length scaling..

Figure 4: Map of seismio-tectoniczones used in the 1995 WGCEPforecast, and M3+ aftershocks within 30days of the four subsequent M6+earthquakes. Probabilities were assignedto each of the zones in 1995. The figureshows how rupture extent can be definedusing aftershock epicenters.

Figure 6: Conditional stoppingprobabilities as a function of distancesouthward from Mendocino for the SanAndreas Fault (adapted from Figure 29of Field et al., 2014). Red curve is basedon modeled earthquake ruptures in theUCERF3 report, while the blue curveillustrates the null hypothesis. Both areevaluated on about 5 km sub-segments.

Figure 5: Conditional stoppingprobabilities for 20 km fault segmentboundary zones on the southern SanAndreas fault. Blue bars are calculatedfrom segment rupture forecasts inUCERF2 (Field et al, 2009); horizontalred line is for null hypothesis of randomlyplaced self-similar GR ruptures oninfinitely long fault.

Present efforts include mapping rupture zones as illustrated in Figure 4 onto thestopping probabilities in Figure 3. The UCERF3 model implies relevant earthquakes alongthe San Andreas fault at a rate of about 0.15 per year. A definitive prospective test mighttake several decades, but likelihood scores from just one large earthquake might falsify the UCERF3 fault model if it ruptured across more than one of the “spikes” on the red curve.