Professional Documents
Culture Documents
Science Aau0323
Science Aau0323
Machine learning for data-driven The two primary classes of ML algorithms are
supervised and unsupervised techniques. In sup-
ervised learning, the ML algorithm “learns” to
discovery in solid Earth geoscience recognize a pattern or make general predictions
using known examples. Supervised learning algo-
rithms create a map, or model, f that relates a
Karianne J. Bergen1,2, Paul A. Johnson3, Maarten V. de Hoop4, Gregory C. Beroza5*
data (or feature) vector x to a corresponding
label or target vector y: y = f(x), using labeled
Understanding the behavior of Earth through the diverse fields of the solid Earth geosciences
training data [data for which both the input and
is an increasingly important task. It is made challenging by the complex, interacting, and
corresponding label (x(i), y(i)) are known and
multiscale processes needed to understand Earth’s behavior and by the inaccessibility of nearly
available to the algorithm] to optimize the mod-
all of Earth’s subsurface to direct observation. Substantial increases in data availability and
el. For example, a supervised ML classifier might
in the increasingly realistic character of computer simulations hold promise for accelerating
learn to detect cancer in medical images using
progress, but developing a deeper understanding based on these capabilities is itself
a set of physician-annotated examples (9). A well-
challenging. Machine learning will play a key role in this effort. We review the state of the field
trained model should be able to generalize and
and make recommendations for how progress might be broadened and accelerated.
make accurate predictions for previously un-
T
seen inputs (e.g., label medical images from new
point belongs to one of two classes. Reynen and of alpine rockslides (40), volcanic signals (41, 42), whereas nonlinear kernel functions allow for
Audet (37) apply a logistic regression classifier to regional earthquakes (43), and induced earth- nonlinear decision boundaries between classes
distinguish automatically between earthquake quakes (44). A detailed explanation of HMMs [see Cracknell and Reading (50) and Shahnas
signals and explosions, using polarization and and their application to seismic waveform data et al. (51) for explanations of SVMs and kernel
frequency features extracted from seismic wave- can be found in Hammer et al. (42). Dynamic methods, respectively].
form data. They extend their approach to detect Bayesian networks (DBNs), another type of graph- Shahnas et al. (51) use an SVM to study mantle
earthquakes in continuous data by classifying each ical model that generalizes HMMs, have also convection processes by solving the inverse prob-
time segment as earthquake or noise. They used been used for earthquake detection (45, 46). In lem of estimating mantle density anomalies from
class probabilities at each seismic station to com- exploration geophysics, hierarchical graphical the temperature field. Temperature fields com-
bine the detection results from multiple stations models have been applied to determine the con- puted by numerical simulations of mantle con-
in the seismic network. Pawley et al. (38) use nectivity of subsurface reservoirs from time-series vection are used as training data. The authors
logistic regression to separate aseismic from seis- measurements using priors derived from convection- also train an SVM to predict the degree of mantle
mogenic injection wells for induced seismicity, diffusion equations (47). The authors report that flow stagnation. Both support vector machines
using features in the model to identify geologic use of a physics-based prior is key to obtaining (18) and support vector regression (19) have been
factors, including proximity of the well to base- a reliable model. Graph-based ML emulators used for rapid magnitude estimation of seismic
ment, associated with a higher risk of induced were used by Srinivasan et al. (48) to mimic high- events for earthquake early warning. Support vec-
seismicity. performance physics-based computations of flow tor machines have also been used for discrimi-
through fracture networks, making robust un- nation of earthquakes and explosions (52) and
Graphical models certainty quantification of fractured systems pos- for earthquake detection in continuous seismic
Many datasets in the geosciences have a tempo- sible (48). data (53).
ral component, such as the ground motion time-
series data recorded by seismometers. Although Support vector machine Random forests and ensemble learning
most ML algorithms can be adapted for use on Support vector machine (SVM) is a binary clas- Decision trees are a supervised method for clas-
temporal data, some methods, like graphical mod- sification algorithm that identifies the optimal sification and regression that learn a piecewise-
els, can directly model temporal dependencies. boundary between the training data from two constant function, equivalent to a series of if-then
For example, hidden Markov models (HMMs) classes (49). SVMs use kernel functions, similar- rules that can be visualized by a binary tree
are a technique for modeling sequential data ity functions that generalize the inner product, to structure. A random forest (RF) is an ensemble
and have been widely used in speech recognition enable an implicit mapping of the data into a learning algorithm that can learn complex rela-
(39). HMMs have been applied to continuous higher-dimensional feature space. SVMs with tionships by voting among a collection (“forest”)
seismic data for the detection and classification linear kernels separate classes with a hyperplane, of randomized decision trees (54) [see Cracknell
Fig. 4. ML methods and their applications. Most ML applications in include semi-supervised learning, in which both labeled and unlabeled
sEg fall within two classes: unsupervised learning and supervised learning. data are available to the learning algorithm, and reinforcement learning.
In supervised learning tasks, such as prediction (21, 28) and classification Deep neural networks represent a class of ML algorithms that include
(16, 20), the goal is to learn a general model based on known (labeled) both supervised and unsupervised tasks. Deep learning algorithms
examples of the target pattern. In unsupervised learning tasks, the have been used to learn feature representations (32, 89), surrogate
goal is instead to learn structure in the data, such as sparse or models for performing fast simulations (23, 75), and joint probability
low-dimensional feature representations (27). Other classes of ML tasks distributions (98, 100).
and Reading (50) for a detailed description of resentation of discrete fracture networks allowed node takes a weighted linear combination of
RFs]. Random forests are relatively easy to use RF and SVMs to identify subnetworks that char- values from the previous layer and applies a
and interpret. These are important advantages acterize the network flow and transport of the nonlinear function to produce a single value that
over methods that are opaque or require tuning full network. The reduced network representa- is passed to the next layer. “Shallow” networks
many hyperparameters (e.g., neural networks, tions greatly decreased the computational effort contain an input layer (data), a single hidden
described below), and have contributed to the required to estimate system behavior. layer, and an output layer (predicted response).
broad application of RFs within sEg. Rouet-Leduc et al. (28, 29) trained a RF on Valentine and Woodhouse (58) present a de-
Kuhn et al. (35) produced lithological maps in continuous acoustic emission in a laboratory tailed explanation of ANNs and the process of
Western Australia using geophysical and remote- shear experiment to determine instantaneous learning weights from training data. ANNs can
sensing data that were trained on a small subset friction and to predict time-to-failure. From be used for both regression and classification,
of the ground area. Cracknell and Reading (55) the continuous acoustic data using the same depending on the choice of output layer.
found that random forests provided the best laboratory apparatus, Hulbert et al. (30) apply a ANNs have a long history of use in the geo-
performance for geological mapping by com- decision tree approach to determine the instan- sciences [see (59, 60) for reviews of early work],
paring multiple supervised ML algorithms. taneous fault friction and displacement on the and they remain popular for modeling nonlinear
Random forest predictions also improved three- laboratory fault. Rouet-Leduc et al. (31) scaled relationships in a range of geoscience appli-
dimensional (3D) geological models using re- the approach to Cascadia by applying instanta- cations. De Wit et al. (61) estimate both the 1D
motely sensed geophysical data to constrain neous seismic data to predict the instantaneous P-wave velocity structure and model uncertain-
geophysical inversions (56). displacement rate on the subducting plate inter- ties from P-wave travel-time data by solving the
Trugman and Shearer (21) discern a predictive face using GPS data as the label. In the lab- Bayesian inverse problem with an ANN. This
relationship between stress drop and peak ground oratory and the field study in Cascadia, ML neural network–based approach is an alternative
acceleration using RFs to learn nonlinear, non- revealed unknown signals. Of interest is that to using the standard Monte Carlo sampling ap-
parametric ground motion prediction equations the same features apply both at laboratory and proach for Bayesian inference. Käufl et al. (25)
(GMPEs) from a dataset of moderate-magnitude field scale to infer fault physics, suggesting a built an ANN model that estimates source pa-
events in northern California. This departed from universality across systems and scales. rameters from strong motion data. The ANN
the typical use of linear regression to model the model performs rapid inversion for source pa-
relationship between expected peak ground velo- Neural networks rameters in real time by precomputing compu-
city or acceleration and earthquake site and Artificial neural networks (ANNs) are an algo- tationally intensive simulations that are then
source parameters that define GMPEs. rithm loosely modeled on the interconnected used to train the neural network model.
Valera et al. (24) used ML to characterize the networks of biological neurons in the brain (57). ANNs have been used to estimate short-period
topology of fracture patterns in the subsurface ANN models are represented as a set of nodes response spectra (62), to model ground motion
for modeling flow and transport. A graph rep- (neurons) connected by a set of weights. Each prediction equations (22), to assess data quality
for focal mechanism and hypocenter location seismic events (71). Wiszniowski et al. (72) in- tion followed by SOM for clustering. SOMs are
(58), and to perform noise tomography (63). troduced a real-time earthquake detection algo- often used to identify seismic facies, but standard
Logistic regression and ANN models allowed rithm using a recurrent neural network (RNN), SOMs do not account for spatial relationships
Mousavi et al. (64) to characterize the source an architecture designed for sequential data. among the data points. Zhao et al. (82) propose
depth of microseimic events induced by under- Magaña- Zook and Ruppert (73) use a long short- imposing a stratigraphy constraint on the SOM
ground collapse and sinkhole formation. Kong et al. term memory (LSTM) network (74), a sophisticated algorithm to obtain more detailed facies maps.
(17) use an ANN with a small number of easy-to- RNN architecture for sequential data, to discrimi- SOMs have also been applied to seismic wave-
compute features for use on a smartphone-based nate natural seismicity from explosions. An ad- form data for feature selection (83) and to cluster
seismic network to distinguish between earth- vantage of DNNs for earthquake detection is that signals to identify multiple event types (84, 85).
quake motion and motion due to user activity. feature extraction is performed by the network, Supervised and unsupervised techniques
so minimal preprocessing is required. By con- are commonly used together in ML workflows.
Deep neural networks trast, shallow ANNs and other classical learning Cracknell et al. (86) train a RF classifier to iden-
Deep neural networks (DNNs), or deep learning, algorithms require the user to select a set of key tify lithology from geophysical and geochem-
are an extension of the classical ANN that in- discriminative features, and poor feature selection ical survey data. They then apply a SOM to the
corporate multiple hidden layers (65). Deep learn- will hurt model performance. Because it may be volcanic units from the RF-generated geologic
ing does not represent a single algorithm, but a difficult to define the distinguishing characteristics map to identify subunits that reveal composi-
broad class of methods with diverse network of earthquake waveforms, the automatic feature tional differences. In the geosciences it is com-
architectures, including both supervised and extraction of DNNs can improve detection per- mon to have large datasets in which only a
unsupervised methods. Deep architectures in- formance, provided large training sets are available. small subset of the data are labeled. Such cases
clude multiple processing layers and nonlinear Araya-Polo et al. (75) use a DNN to learn an call for semi-supervised learning methods de-
transformations, with the outputs from each inverse for a basic type of tomography. Rather signed to learn from both labeled and unlabeled
layer passed as inputs to the next. Supervised than using ML to automate or improve individ- data. In a semi-supervised approach, Köhler et al.
DNNs simultaneously learn a feature represen- ual elements of a standard workflow, they aim to (87) detect rockfalls and volcano-tectonic events
iterative thresholding and signed K-means dic- use a GAN to map seismic source and receiver matching for earthquake detection (106). Each
tionary learning to enhance sparsity of the re- parameters to synthetic multicomponent seismo- of these three applications requires a database
presentation of the slowness estimated from grams. Mosser et al. (99) use a domain transfer of known or precomputed earthquake features
travel-time perturbations. approach, similar to artistic style transfer, with and uses an efficient search algorithm to reduce
a deep convolutional GAN (DCGAN) to learn the computational runtime. By contrast, Yoon et al.
Deep generative models mappings from seismic amplitudes to geological (107) take an unsupervised pattern-mining ap-
Generative models are a class of ML methods structure and vice versa. The authors’ approach proach to earthquake detection; the authors use
that learn joint probability distributions over enables both forward modeling and fast inver- a fast similarity search algorithm to search the
the dataset. Generative models can be applied sion. GANs have also been applied to geological continuous waveform data for similar or re-
to both unsupervised and supervised learning modeling by Dupont et al. (100), who infer local peating, allowing the method to discover new
tasks. Recent work has explored applications geological patterns in fluvial environments from events with previously unknown sources. This
of deep generative models, in particular genera- a limited number of rock type observations using approach has been extended to multiple stations
tive adversarial networks (GANs) (93). A GAN is a GAN similar to those used for image inpaint- (108) and can process up to 10 years of continuous
a system of two neural networks with opposing ing. Chan and Elsheikh (101) generate realistic, data (109).
objectives: a generator network that uses train- complex geological structures and subsurface Network analysis techniques—methods for
ing data to learn a model to generate realistic flow patterns with a GAN. Veillard et al. (102) use analyzing data that can be represented using a
synthetic data and a discriminator network that both a GAN and a VAE (96) to interpret geolo- graph structure of nodes connected by edges—
learns to distinguish the synthetic data from real gical structures in 3D seismic data. have also been used for data-driven discovery in
training data [see (94) for a clear explanation]. the sEg. Riahi and Gerstoft (110) detect weak
Deep generative models, such as the Deep Other techniques sources in a dense array of seismic sensors
Rendering Model (95), Variational Autoencoders Reinforcement learning is a ML framework in using a graph clustering technique. The authors
(VAEs) (96), and GANs (93), are hierarchical prob- which the algorithm learns to make decisions to identify sources by computing components of
abilistic models that explain data at multiple maximize a reward by trial and error. Draelos et al. a graph where each sensor is a node and the
Domain Challenge
adaptation Problems
lagged some other scientific domains with sim- to go unrecognized. The ImageNet image rec- with benchmark datasets, greater sharing of
ilar data quality issues. ognition dataset (118) has been cited in over the code that implements new ML-based sol-
Our recommendations are informed by the 6000 papers. utions will help accelerate the development and
characteristics of geoscience datasets that pre- Recently, the Institute of Geophysics at the validation of these new approaches. Active re-
sent challenges for standard ML algorithms. Data- Chinese Earthquake Administration (CEA) and search areas like earthquake detection benefit as
sets in the sEg represent complex, nonlinear, Alibaba Cloud hosted a data science competition available open source codes enable direct com-
physical systems that act across a vast range of with more than 1000 teams centered around parisons of multiple algorithms on the same data-
length and time scales. Many phenomena, such automatic detection and phase picking of after- sets to assess the relative performance. To the
as fluid injection and earthquakes, are strongly shocks following the 2008 Ms 8.0 Wenchuan extent possible, this should also extend to the
nonstationary. The resulting data are complex, earthquake (119, 120). The ground-truth phase- sharing of the original datasets and pretrained
with multiresolution, spatial, and temporal struc- arrival data, against which entries were assessed, ML models.
tures requiring innovative approaches. Further, were determined by CEA analysts. Such chal- The use of electronic preprints [e.g., (129–131)]
much existing data are unlabeled. When availa- lenges are useful for researchers seeking to test may also help to accelerate the pace of research
ble, labels are often highly subjective or biased and improve their detection algorithms. Future (132) at the intersection of Earth science and AI.
toward frequent or well-characterized phenome- competitions should have greater impact if they Preprints allow authors to share preliminary re-
na, limiting the effectiveness of algorithms that are accompanied with some form of broader sults with a wider community and receive feed-
rely on training datasets. The quality and com- follow-up, such as publications associated with back in advance of the formal review process.
pleteness of datasets create another challenge as top-performing entries or a summary of effec- The practice of making research available on
uneven data collection, incomplete datasets, and tive methods and lessons learned from compe- preprint servers is common in many science,
noisy data are common. tition organizers. technology, engineering, and mathematics (STEM)
Creating benchmark datasets and determin- fields, including computer vision and natural lan-
Benchmark datasets ing evaluation metrics are challenging when the guage processing—two fields that are driving
The lack of clear ground-truth and standard underlying data are incompletely understood. development in deep learning (133); however,
ML algorithms, including scientist-in-the-loop scientists. A greater component of data science 20. D. Shoji, R. Noguchi, S. Otsuki, H. Hino, Classification of
systems. and ML in geoscience curricula could help, as volcanic ash particles using a convolutional neural network
and probability. Sci. Rep. 8, 8111 (2018). doi: 10.1038/
A challenge that comes with entirely data- would recruiting students trained in data science s41598-018-26200-2; pmid: 29802305
driven approaches is the need for large quan- to work on geoscience research. Interdisciplinary 21. D. T. Trugman, P. M. Shearer, Strong Correlation between
tities of training data, especially for modeling research conferences could and are being used to Stress Drop and Peak Ground Acceleration for Recent M 1–4
through deep learning. Moreover, ML models promote collaborations through identifying com- Earthquakes in the San Francisco Bay Area. Bull. Seismol.
Soc. Am. 108, 929–945 (2018). doi: 10.1785/0120170245
may end up replicating the biases in training mon interests and complementary capabilities. 22. B. Derras, P. Y. Bard, F. Cotton, Towards fully data driven ground-
data, which can arise during data collection or motion prediction models for Europe. Bull. Earthquake Eng. 12,
even through the use of specific training data- 495–516 (2014). doi: 10.1007/s10518-013-9481-0
sets. Thus, extracting maximum value from RE FERENCES AND NOTES 23. P. M. R. DeVries, T. B. Thompson, B. J. Meade, Enabling
large-scale viscoelastic calculations via neural network
geoscientific datasets will require methods cap- 1. P. Baldi, P. Sadowski, D. Whiteson, Searching for exotic
acceleration. Geophys. Res. Lett. 44, 2662–2669 (2017).
able of learning with limited, weak, or biased particles in high-energy physics with deep learning.
doi: 10.1002/2017GL072716
Nat. Commun. 5, 4308 (2014). doi: 10.1038/ncomms5308;
labels. Furthermore, because the phenomena pmid: 24986233
24. M. Valera, Z. Guo, P. Kelly, S. Matz, V. A. Cantu, A. G. Percus,
of interest are governed by complex and dy- J. D. Hyman, G. Srinivasan, H. S. Viswanathan, Machine
2. A. A. Melnikov, H. P. Nautrup, M. Krenn, V. Dunjko, M. Tiersch,
learning for graph-based representations of three-
namic physical processes, there is a need for A. Zeilinger, H. J. Briegel, Active learning machine learns to
dimensional discrete fracture networks. Computat. Geosci.
new approaches to analyzing scientific data- create new quantum experiments. Proc. Natl. Acad. Sci. USA.
22, 695–710 (2018). doi: 10.1007/s10596-018-9720-1
115, 1221–1226 (2018). doi: 10.1073/pnas.1714936115
sets that combine data-driven and physical mod- 3. C. J. Shallue, A. Vanderburg, Identifying Exoplanets with 25. P. Käufl, A. P. Valentine, J. Trampert, Probabilistic point
eling (144). Deep Learning: A Five-planet Resonant Chain around
source inversion of strong-motion data in 3-D media using
Much of the representational power of mod- Kepler-80 and an Eighth Planet around Kepler-90. Astron. J. pattern recognition: A case study for the 2008 M w 5.4 Chino
Hills earthquake. Geophys. Res. Lett. 43, 8492–8498 (2016).
ern ML techniques, such as DNNs, comes from 155, 94 (2018). doi: 10.3847/1538-3881/aa9e09
4. F. Ren, L. Ward, T. Williams, K. J. Laws, C. Wolverton, doi: 10.1002/2016GL069887
the ability to recognize paths to data inversion J. Hattrick-Simpers, A. Mehta, Accelerated discovery of 26. M. T. McCann, K. H. Jin, M. Unser, Convolutional Neural
outside of the established physical and mathe- metallic glasses through iteration of machine learning Networks for Inverse Problems in Imaging: A Review. IEEE
matical frameworks, through nonlinearities that and high-throughput experiments. Sci. Adv. 4, eaaq1566 Signal Process. Mag. 34, 85–95 (2017). doi: 10.1109/
MSP.2017.2739299
data using hidden Markov models. J. Geophys. Res. Earth Surf. traveltimes using neural networks. Geophys. J. Int. 195, 84. A. Esposito, F. Giudicepietro, S. Scarpetta ,L. D’Auria,
121, 351–371 (2016). doi: 10.1002/2015JF003647 408–422 (2013). doi: 10.1093/gji/ggt220 M. Marinaro, M. Martini, Automatic Discrimination among
41. M. Beyreuther, R. Carniel, J. Wassermann, Continuous Hidden 62. R. Paolucci, F. Gatti, M. Infantino, C. Smerzini, Landslide, Explosion-Quake, and Microtremor Seismic
Markov Models: Application to automatic earthquake A. Guney Ozcebe, M. Stupazzini, Broadband ground Signals at Stromboli Volcano Using Neural Networks.
detection and classification at Las Canãdas caldera, Tenerife. motions from 3D physics-based numerical simulations using Bull. Seismol. Soc. Am. 96 (4A), 1230–1240 (2006).
J. Volcanol. Geotherm. Res. 176, 513–518 (2008). artificial neural networks. Bull. Seismol. Soc. Am. 108, doi: 10.1785/0120050097
doi: 10.1016/j.jvolgeores.2008.04.021 1272–1286 (2018). 85. A. Esposito, F. Giudicepietro, L. D’Auria, S. Scarpetta,
42. C. Hammer, M. Beyreuther, M. Ohrnberger, A Seismic- 63. P. Paitz, A. Gokhberg, A. Fichtner, A neural network for noise M. Martini, M. Coltelli, M. Marinaro, Unsupervised Neural
Event Spotting System for Volcano Fast-Response Systems. correlation classification. Geophys. J. Int. 212, 1468–1474 Analysis of Very-Long-Period Events at Stromboli Volcano
Bull. Seismol. Soc. Am. 102, 948–960 (2012). doi: 10.1785/ (2018). doi: 10.1093/gji/ggx495 Using the Self-Organizing Maps. Bull. Seismol. Soc. Am. 98,
0120110167 64. S. M. Mousavi, S. P. Horton, C. A. Langston, B. Samei, 2449–2459 (2008).
43. M. Beyreuther, J. Wassermann, Continuous earthquake Seismic features and automatic discrimination of deep and doi: 10.1785/0120070110
detection and classification using discrete Hidden Markov shallow induced-microearthquakes using neural network and 86. M. J. Cracknell, A. M. Reading, A. W. McNeill, Mapping
Models. Geophys. J. Int. 175, 1055–1066 (2008). doi: 10.1111/ logistic regression. Geophys. J. Int. 207, 29–46 (2016). geology and volcanic-hosted massive sulfide alteration in the
j.1365-246X.2008.03921.x doi: 10.1093/gji/ggw258 Hellyer–Mt Charter region, Tasmania, using Random
44. M. Beyreuther, C. Hammer, J. Wassermann, M. Ohrnberger, 65. Y. LeCun, Y. Bengio, G. Hinton, Deep learning. Forests™ and Self-Organising Maps. Aust. J. Earth Sci. 61,
T. Megies, Constructing a Hidden Markov Model based Nature 521, 436–444 (2015). doi: 10.1038/nature14539; 287–304 (2014). doi: 10.1080/08120099.2014.858081
earthquake detector: Application to induced seismicity. pmid: 26017442 87. A. Köhler, M. Ohrnberger, F. Scherbaum, Unsupervised
Geophys. J. Int. 189, 602–610 (2012). doi: 10.1111/ 66. Z. E. Ross, M.-A. Meier, E. Hauksson, P Wave Arrival pattern recognition in continuous seismic wavefield records
j.1365-246X.2012.05361.x Picking and First-Motion Polarity Determination With Deep using Self-Organizing Maps. Geophys. J. Int. 182, 1619–1630
45. C. Riggelsen, M. Ohrnberger, F. Scherbaum, Dynamic Learning. J. Geophys. Res. Solid Earth 123, 5120–5129 (2010). doi: 10.1111/j.1365-246X.2010.04709.x
Bayesian networks for real-time classification of seismic (2018). doi: 10.1029/2017JB015251 88. B. Sick, M. Guggenmos, M. Joswig, Chances and limits of
signals, in Knowledge Discovery in Databases. PKDD 2007. 67. B. Moseley, A. Markham, T. Nissen-Meyer, Fast approximate single-station seismic event clustering by unsupervised
Lecture Notes in Computer Science, vol. 4702, simulation of seismic waves with deep learning. pattern recognition. Geophys. J. Int. 201, 1801–1813 (2015).
J. N. Kok et al., Eds. (Springer, 2007), pp. 565–572. arXiv:1807.06873 [physics.geo-ph] (2018). doi: 10.1093/gji/ggv126
doi: 10.1007/978-3-540-74976-9_59 68. Y. Wu, Y. Lin, Z. Zhou, D. C. Bolton, J. Liu, P. Johnson, 89. F. Qian et al., Unsupervised seismic facies analysis via deep
46. C. Riggelsen, M. Ohrnberger, A Machine Learning Approach Cascaded region-based densely connected network convolutional autoencoders. Geophysics 83, A39–A43
for Improving the Detection Capabilities at 3C Seismic for event detection: A seismic application. arXiv:1709.07943 (2018). doi: 10.1190/geo2017-0524.1
Stations. Pure Appl. Geophys. 171, 395–411 (2014). [cs.LG] (2017). 90. S. Beckouche, J. Ma, Simultaneous dictionary learning and
Comput. Geosci. 114, 22–29 (2018). doi.10.1016/ 120. L. Fang, Z. Wu, K. Song, SeismOlympics. Seismol. Res. Lett. Wide Web (ACM, 2010), pp. 851–860. doi: 10.1145/
j.cageo.2018.01.001 88, 1429–1430 (2017). doi: 10.1785/0220170134 1772690.1772777
106. R. Tibi, C. Young, A. Gonzales, S. Ballard, A. Encarnacao, 121. D. Sculley, J. Snoek, A. Wiltschko, A. Rahimi, Winner’s curse? 139. P. S. Earle, D. C. Bowden, M. Guy, Twitter earthquake
Rapid and robust cross-correlation-based seismic phase on pace, progress, and empirical rigor, in International detection: Earthquake monitoring in a social world. Ann.
identification using an approximate nearest neighbor Conference on Learning Representations (ICLR) 2018 Geophys. 54, 708–715 (2012). doi: 10.4401/ag-5364
method. Bull. Seismol. Soc. Am. 107, 1954–1968 (2017). Workshop. (2018); https://openreview.net/forum? 140. R. Bossu, G. Mazet-Roux, V. Douet, S. Rives, S. Marin,
doi: 10.1785/0120170011 id=rJWF0Fywf. M. Aupetit, Internet Users as Seismic Sensors for Improved
107. C. E. Yoon, O. O’Reilly, K. J. Bergen, G. C. Beroza, Earthquake 122. Y. Gil et al., Toward the Geoscience Paper of the Future: Earthquake Response. Eos 89, 225–226 (2008).
detection through computationally efficient similarity search. Best practices for documenting and sharing research doi: 10.1029/2008EO250001
Sci. Adv. 1, e1501057 (2015). doi: 10.1126/sciadv.1501057; from data to software to provenance. Earth Space Sci. 3, 141. E. S. Cochran, J. F. Lawrence, C. Christensen, R. S. Jakka,
pmid: 26665176 388–415 (2016). doi: 10.1002/2015EA000136 The Quake-Catcher Network: Citizen Science Expanding
108. K. J. Bergen, G. C. Beroza, Detecting earthquakes over 123. GitHub, https://github.com. Seismic Horizons. Seismol. Res. Lett. 80, 26–30 (2009).
a seismic network using single-station similarity 124. GitLab, https://about.gitlab.com. doi: 10.1785/gssrl.80.1.26
measures. Geophys. J. Int. 213, 1984–1998 (2018). 125. M. Beyreuther et al., ObsPy: A Python Toolbox for 142. DigitalGlobe, http://www.digitalglobe.com.
doi: 10.1093/gji/ggy100 Seismology. Seismol. Res. Lett. 81, 530–533 (2010). 143. A. Karpatne, I. Ebert-Uphoff, S. Ravela, H. A. Babaie,
109. K. Rong, C. E. Yoon, K. J. Bergen, H. Elezabi, P. Bailis, doi: 10.1785/gssrl.81.3.530 V. Kumar, Machine learning for the geosciences:
P. Levis, G. C. Beroza, Locality-sensitive hashing for 126. L. Krischer et al., ObsPy: A bridge for seismology into the Challenges and opportunities. arXiv:1711.04708
earthquake detection: A case study scaling data-driven scientific Python ecosystem. Comput. Sci. Discov. 8, 014003 [cs.LG] (2017).
science. Proceedings of the International Conference (2015). doi: 10.1088/1749-4699/8/1/014003 144. A. Karpatne et al., Theory-Guided Data Science: A New
on Very Large Data Bases (PVLDB) 11, 1674 (2018). 127. S. Heimann, et al., Pyrocko - An open-source seismology Paradigm for Scientific Discovery from Data. IEEE Trans.
doi: 10.14778/3236187.3236214 toolbox and library. V. 0.3. GFZ Data Services. (2017). Knowl. Data Eng. 29, 2318–2331 (2017). doi: 10.1109/
110. N. Riahi, P. Gerstoft, Using graph clustering to locate sources doi: 10.5880/GFZ.2.1.2017.001 TKDE.2017.2720168
within a dense sensor array. Signal Processing 132, 110–120 128. F. Pedregosa et al., Scikit-learn: Machine Learning in Python. 145. J.-Y. Zhu, T. Park, P. Isola, A. A. Efros, Unpaired
(2017). doi: 10.1016/j.sigpro.2016.10.001 J. Mach. Learn. Res. 12, 2825–2830 (2011). image-to-image translation using cycle-consistent
111. A. C. Aguiar, G. C. Beroza, PageRank for Earthquakes. Seismol. 129. arXiv e-Print archive, https://arxiv.org. adversarial networks. arXiv:1703.10593 [cs.CV]
Res. Lett. 85, 344–350 (2014). doi: 10.1785/0220130162 130. EarthArXiv Preprints, https://eartharxiv.org. (2017).
112. M. Campbell, A. J. Hoane Jr., F.-H. Hsu, Deep Blue. Artif. Intell. 131. Earth and Space Science Open Archive, https://essoar.org. 146. I. Ebert-Uphoff, Y. Deng, Three steps to successful
134, 57–83 (2002). doi: 10.1016/S0004-3702(01)00129-1 132. J. Kaiser, The preprint dilemma. Science 357, collaboration with data scientists. Eos 98, (2017).
1344–1349 (2017). doi: 10.1126/science.357.6358.1344;
Science (ISSN 1095-9203) is published by the American Association for the Advancement of Science. 1200 New York Avenue NW,
Washington, DC 20005. The title Science is a registered trademark of AAAS.
Copyright © 2019 The Authors, some rights reserved; exclusive licensee American Association for the Advancement of Science. No claim
to original U.S. Government Works