Automatic classification of volcano seismic signatures
Marielle Malfante, Mauro Dalla Mura, Jerome Mars, Jean-Philippe Métaxian,
Orlando Macedo, Adolfo Inza
To cite this version:
Marielle Malfante, Mauro Dalla Mura, Jerome Mars, Jean-Philippe Métaxian, Orlando Macedo, et
al.. Automatic classification of volcano seismic signatures. Journal of Geophysical Research, American
Geophysical Union, 2018, 123 (12), pp.10,645-10,658. 10.1029/2018jb015470. hal-01949302
HAL Id: hal-01949302
https://hal.archives-ouvertes.fr/hal-01949302
Submitted on 21 Dec 2018
HAL is a multi-disciplinary open access
archive for the deposit and dissemination of scientific research documents, whether they are published or not. The documents may come from
teaching and research institutions in France or
abroad, or from public or private research centers.
L’archive ouverte pluridisciplinaire HAL, est
destinée au dépôt et à la diffusion de documents
scientifiques de niveau recherche, publiés ou non,
émanant des établissements d’enseignement et de
recherche français ou étrangers, des laboratoires
publics ou privés.
Journal of Geophysical Research: Solid Earth
RESEARCH ARTICLE
Automatic Classification of Volcano Seismic Signatures
10.1029/2018JB015470
Key Points:
• We propose a supervised process for
the automatic classification of volcano
seismic signals
• We test and validate the proposed
architecture on 6 years of continuous
seismic recordings from a
short-period seismic station
• The automatic classification scheme
performs better than human analysts
in a crisis situation
Correspondence to:
J.-P. Métaxian,
[email protected]
Citation:
Malfante, M., Dalla Mura, M., Mars, J. I.,
Métaxian, J.-P., Macedo, O., & Inza, A.
(2018). Automatic classification of
volcano seismic signatures. Journal of
Geophysical Research: Solid Earth, 123.
https://doi.org/10.1029/2018JB015470
Received 30 JAN 2018
Accepted 21 SEP 2018
Accepted article online 26 SEP 2018
Marielle Malfante1
Orlando Macedo4,5
, Mauro Dalla Mura1
, and Adolfo Inza4
, Jerome I. Mars1
, Jean-Philippe Métaxian2,3
,
1 Univ. Grenoble Alpes, CNRS, Grenoble INP, GIPSA-lab, Grenoble, France, 2 Univ. Grenoble Alpes, Univ. Savoie Mont Blanc,
CNRS, IRD, ISTerre, Grenoble, France, 3 Institut de Physique du Globe de Paris, Université Sorbonne-Paris-Cité, CNRS, Paris,
France, 4 Instituto Geofísico del Perú, Lima, Peru, 5 Faculdad de Geologia, Geofisica y Minas, Universidad Nacional de San
Agustin de Arequipa, Arequipa, Peru
Abstract
The prediction of volcanic eruptions and the evaluation of associated risks remain a timely and
unresolved issue. This paper presents a method to automatically classify seismic events linked to volcanic
activity. As increased seismic activity is an indicator of volcanic unrest, automatic classification of volcano
seismic events is of major interest for volcano monitoring. The proposed architecture is based on supervised
classification, whereby a prediction model is built from an extensive data set of labeled observations.
Relevant events should then be detected. Three steps are involved in the building of the prediction model:
(i) signals preprocessing, (ii) representation of the signals in the feature space, and (iii) use of an automatic
classifier to train the model. Our main contribution lies in the feature space where the seismic observations
are represented by 102 features gathered from both acoustic and seismic fields. Ideally, observations are
separable in the feature space, depending on their class. The architecture is tested on 109,609 seismic
events that were recorded between June 2006 and September 2011 at Ubinas Volcano, Peru. Six main
classes of signals are considered: long-period events, volcanic tremors, volcano tectonic events, explosions,
hybrid events, and tornillos. Our model reaches 93.5% ± 0.50% accuracy, thereby validating the presented
architecture and the features used. Furthermore, we illustrate the limited influence of the learning algorithm
used (i.e., random forest and support vector machines) by showing that the results remain accurate
regardless of the algorithm selected for the training stage. The model is then used to analyze 6 years of data.
1. Introduction
1.1. Volcano Monitoring
Monitoring volcanoes and forecasting their related hazards has been a concern and a challenge for the scientific community for many years. With the potential for dramatic consequences on populations and the
economy, volcanic hazards are among the most threatening, and thus, the interest in reliable forecasting and
prevention tools is obvious. A common approach for volcano monitoring relies on tracking the evolution of
several parameters, including seismicity, deformation, and gas flux and gas composition (McNutt, 1996). Usually, the tracking of these parameters and of their evolution is still essentially carried out manually. Over the
last decade, the technical evolution of geophysical instruments and their cost reduction have led volcano
observatories to install an increasing number of sensors. The ever increasing amounts of recorded data that
have to be processed makes manual approaches inappropriate for carrying out comprehensive analyses. This
represents another example of the big data phenomenon. As in various domains, machine learning methods
are today considered to automate analyses and to help to improve decision making, which can lead to health
and safety recommendations.
In this study, we propose a novel approach to automatically classify volcano seismic events that is based on
machine learning. The success of these methods relies on signal characterization with a large set (102) of
descriptors.
©2018. American Geophysical Union.
All Rights Reserved.
MALFANTE ET AL.
1.2. Machine Learning
We give here a short introduction to supervised machine learning that is aimed at defining the terminology
and the framework of the proposed technique. Machine learning is a field of artificial intelligence that is aimed
at automatically assigning objects (from a potentially large set) to categories that are associated to semantic
information. This automatic decision process relies on a prediction model that is learned on a set of examples
where the categories (or classes) are known beforehand. More specifically, the analysis is carried out on a data
1
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
set of observations. In this study, we refer to observations as the recorded seismic signal that corresponds to
a volcano seismic event (e.g., an explosion or a tremor). However, in the literature and in many studies related
to machine learning, observations can also be referred to as signatures, samples, examples, data, or simply as
signals. The data set is said to be labeled if each observation has previously been categorized and associated to
a semantic class, that is, to one of the classes of interest (in the present study, classes correspond to the type of
volcanic event that is associated to the observation). Supervised machine learning algorithms provide models that can learn to predict the class of a new observation, where the label is not known. Many supervised
machine learning algorithms are being used today. We will introduce here two of the most effective ones (random forest and support vector machines-SVMs), which were used to automatically classify volcano seismic
observations in this study. Machine learning algorithms are today very effective and are widely used in image
processing, speech processing, medical imaging, finance, and robotics, among others (Friedman et al., 2001).
As in many signal processing applications, a key point in machine learning processes is the representations
used for the observations. Traditionally, these are represented by features that relate to measurements made
on the observations that summarize and characterize their properties. In particular, properties that discriminate observations into their respective classes should be highlighted in the feature space. In machine learning,
the use of a feature space with respect to the original recording space is also motivated by the curse of dimensionality (Bellman, 1956). More specifically, the number of observations, N, needed to train a model increases
with the dimension, d, of the input space, thereby representing the data in a space of low dimension.
Various methods can be used to pass from the original space of representation to a feature space, which
include learning features or designed features, which we will briefly explain (Friedman et al., 2001). However,
using one or the other method of defining features gives no guarantee of finding the optimum space of representation for the data. Finding an appropriate feature space is always the key point when using machine
learning algorithms, for it greatly impacts on the output performance and strongly modifies the results. Learning representations are often referred to as dictionary learning, and they include algorithms such as principal
component analysis, independent component analysis, singular values decomposition, nonnegative matrix
factorization, and convolutional neural networks. Their main idea is to learn a dictionary of elements that can
be used to reconstruct the data. The main advantage of these algorithms is the final representation that is
adapted to best discriminate the observations into their classes. However, such algorithms often require a
large and labeled data set to learn the representation, and they can be very costly. Furthermore, a learning
representation cannot be related to a physical meaning and cannot be generalized to other data.
In the present study, as in many studies that have dealt with natural or environmental data, we use designed
features. In particular, we introduce D = 102 descriptors to extract features that represent and quantify the
physical values of the observations. The main challenge of this method is the difficulty of finding features
that effectively separate the observations into their classes. This step requires deep knowledge of the data
and, especially, why a given observation belongs to a specific class. One advantage of these feature sets is
that they can be generalized: They can be designed for one application, but they can be relevant to other
fields. Moreover, this technique is less constraining on computation time and allows the observation to be
represented by physical quantities that can be associated to a physical meaning of the phenomenon. In signal
processing, classical features would be mean values or standard deviation, for example. Yet another option to
extract features would be to apply dictionary learning over designed features, which would lead to an even
more compressed representation of the signals, although once again, this loses the physical meaning of the
features.
We will now briefly introduce the two learning algorithms that will be used throughout the present study. The
random forest algorithm (Breiman, 2001) is built upon an ensemble of binary decision trees (Quinlan, 1986),
which are combined by majority voting. A binary decision tree is a discriminative algorithm that carries out
iterative tests on features in a hierarchical manner, until a classification rule is defined. Potentially, the main
advantage of the algorithm lies in its simplicity and its explanatory power. However, binary decision trees
tend to overfit (i.e., adapt too closely) the examples from the learning phase, and thereby the learned decision
rules might not be general enough to correctly discriminate newly recorded observations. Random forest
was proposed to overcome this limitation by using bagging: An ensemble of many binary trees is considered,
with each trained on a random subset of the original data set (e.g., typically, a random 10% of the data set
is left for each tree). Furthermore, each node of a tree takes decisions on a random subset of the features
MALFANTE ET AL.
2
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
√
used to represent the observations (e.g., typically, d features are considered, where d is the feature vector
dimension). This second approach aims to further decrease the correlation among the binary trees.
The second machine learning approach we use, SVM algorithms were presented by Cortes and Vapnik (1995).
The main idea is to build a linear hyperplane in the feature space that maximizes the margin (i.e., the distance
between the hyperplane and the closest data). This simple linear classifier works well on data that are linearly
separable in the feature space, which is rarely the case in real-case scenarios. Two variants have been proposed to improve the algorithm performances. The first one uses a kernel function to project the data from
the original to another feature space of higher dimension, in which the data would be linearly separable. This
is known as the kernel trick (Friedman et al., 2001). Several kernel functions have been shown to be effective.
In the case where a radial basis function is used as the kernel, the final feature space has an infinite dimension, and the data are therefore necessarily linearly separable. The second variant to the original algorithm is
to use a soft margin, that is, to introduce a cost parameter, CSVM , that allows some training observations to be
on the wrong side of the margin. Using a soft margin prevents overfitting, thus provides a model that has better generalization potential. In this study, we use a SVM implementing both improvements over the original
algorithm.
1.3. Related Studies
In this section, we review the studies that are related to the present study, in terms of the automatic classification of volcano seismic signals and also the choice of signal representations for machine learning purposes. We
especially review several fields, including acoustic data classification. Indeed, automatic analysis of acoustic
data is today much more advanced than tools currently used for analysis of seismic data. The interest in such
adjusting tools that were developed for other purposes is strong, especially given the similarities between the
data involved (i.e., the transient nature of the signals). In the following, we focus our attention on the features
that vary significantly from one application to another. The learning algorithm has less impact on the results,
and for most studies this was chosen as random forest (Hibert et al., 2017; Maggi et al., 2017), SVMs (Apolloni
et al., 2009), neural networks (Apolloni et al., 2009; Langer et al., 2006), or hidden Markov models (Benítez et al.,
2007; Cortés et al., 2009; Gutiérrez et al., 2009; Hammer et al., 2012; Ibáñez et al., 2009; Ohrnberger, 2003).
Automatic classification systems for volcano seismic waves are not yet widespread, although several studies
on this subject have started to appear. In particular, Hibert et al. (2017), Langet (2014), and Maggi et al. (2017)
classified seismic observations represented by features that included duration, skewness, kurtosis, statistics
ratios, and centroids. The study of Maggi et al. (2017) had a similar approach to our architecture, where
the main difference is that they obtained their best results when using several seismic stations simultaneously, while in the present study, we obtain the same results from a single station. The approach and testing
methods are, however, very similar. The approach presented in Hibert et al. (2017) is also similar but on a
binary classification problem. Other studies propose alternative features, such as used by Langer et al. (2006),
where features based on energetic criteria were considered, such as the signal energy in various frequency
bands. Cannata et al. (2011) used three features: peak frequency, quality factor, and amplitude. Apolloni et al.
(2009) used linear predictive coding as a normalized difference between maximum and minimum amplitudes.
Esposito et al. (2008) propose to use unsupervised method but on the observation of waveforms directly.
Hibert et al. (2014) propose to use five features, including the signal duration, duration of increasing and
decreasing phases, kurtosis of the envelope, and signal maximum over mean ratio. Those features are used
to discriminate rockfall from all other signals. Benítez et al. (2007), Cortés et al. (2009), Gutiérrez et al. (2009),
and Ibáñez et al. (2009) use modified mel frequency cepstral coefficients, which were originally designed for
speech processing and measure the signals energetic repartition in frequency bands. Physical features, such
as polarization and spectral attributes, can also be used (Hammer et al., 2012; Ohrnberger, 2003).
Acoustic signals of various origin can be targeted for automatic classification, including environmental
sounds, music, or speech. For automatic classification of environmental sounds (i.e., from nature and animals,
or human induced), we can mention Tucker and Brown (2005), who classified transient sonar sounds. More
than 20 features were gathered to describe the signals, including duration, peak power, average power, time
of peak power, mean skewness, mean kurtosis, power standard deviation, rate of attack, and rate of decay.
Similar features were used by Zaugg et al. (2010) to distinguish boat sounds from whale sounds. Fagerlund
(2007) also gathered signal descriptors to identify bird species from their calls, which included spectral centroid, bandwidth, and spectral flatness and duration. Spectral centroid, bandwidth, and threshold crossing
rates were also considered by Huang et al. (2009) for a frog sound identification system. For the same purpose,
MALFANTE ET AL.
3
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Han et al. (2011) also used Shannon and Rényi entropies. Similar features have been used for music processing; see, for example, Fujinaga and
MacMillan (2000), Eronen and Klapuri (2000), and Esmaili et al. (2004).
For the present study, our main contribution lies in the definition of the
feature space in which the signals are defined. Namely, we propose the use
of 102 features to obtain a comprehensive description of the observations.
This paper is organized as follows: Section 2 presents the volcano seismic
data set used in this study, which was recorded at Ubinas Volcano (Peru).
We present the automatic classification architecture in section 3, and in
particular, the feature set is described in section 3. The results are then presented and discussed in section 4. Finally, the prospects and conclusions
surrounding this study are summarized in section 5.
2. Experimental Material
2.1. Seismic Monitoring of Ubinas Volcano
Ubinas Volcano is an andesitic stratovolcano in southern Peru, (16∘ 22′ S,
70∘ 54′ W; altitude, 5,672 m). It is considered to be the most active volcano in Peru, and it is closely monitored by the Instituto Geofísico del Perú
(IGP). After nearly 40 years of quiescence, Ubinas Volcano erupted in 2006.
Three eruptions have occurred since 2006, from 2006 to 2011, from 2013 to
Figure 1. Map of Ubinas Volcano with the locations of the permanent IGP
2014, and in 2016. Ubinas Volcano has been monitored seismically by the
seismic network indicated (white triangles). The data used in this study were
IGP since 2006 (Macedo et al., 2009), with the cooperation of the VOLUME
recorded at UBIW station. Inset, top left: location of Ubinas Volcano (black
project (funded by the European Commission 6th Framework Program)
triangle) within Peru.
and the Institut de Recherche pour le Developpement (France). The first permanent telemetered station (i.e., UBIW) was equipped with a short-period vertical 1-Hz sensor that was
installed in May 2006 on the northwest flank of Ubinas Volcano (Macedo et al., 2009). Three additional stations were added in 2007 (i.e., UBIN, UBIE, and UBIS). UBIN was equipped with a broadband vertical sensor,
and the other stations had short-period sensors. In addition, UBIN and UBIS were equipped with biaxial tiltmeters with 0.1-μrad resolution (Inza et al., 2014). These four stations have been working permanently since
2007 (Figure 1). The data are recorded continuously with a sampling rate of 100 Hz, and they are then transmitted in real time to Cayma Volcanological Observatory in Arequipa (Peru). In this paper, our analysis is based
on seismic data from the vertical component of UBIW station.
2.2. Seismic Signatures Overview
The data used for the present study came from a catalog of N = 109, 609 seismic observations of volcanic
events that were recorded between May 2006 and October 2011. Each observation was manually identified
and extracted by the IGP in Arequipa Observatory. Six very heterogeneous classes of signals were defined by
Figure 2. Waveform and spectrogram (Gaussian window of 512 samples width) of volcano seismic signals recorded at
Ubinas Volcano. Six observations are shown, as examples of long-period (LP) events, tremors (TRs), explosions (EXPs),
volcano-tectonic (VT) events, hybrid (HYB) events, and Tornillos (TORs). The amplitude is linear and has been
normalized.
MALFANTE ET AL.
4
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Figure 3. Number of events per day for the six different types of volcano seismic signals recorded at Ubinas Volcano.
The date is indicated in cumulated days starting from 1 January 2006.
the IGP after 10 years of observations of Ubinas Volcano. Each class is associated to a physical state or activity
of the volcano. These types of signals are observed for other volcanoes, and have been described for many
years in the literature, in particular, by (Chouet, 1996; Lahr et al., 1994; Neuberg et al., 2000). They are listed in
the following:
1. Long-period (LP) events (95,243 observations): These originate from fluid processes. They are interpreted
as a time-localized pressure excitation mechanism, followed by the response of a fluid-filled resonator
(Chouet & Matoza, 2013). Different models have been developed to explain the resonance observed for
LP events, including in particular the fluid-filled crack model (Chouet, 1986) and the fluid-filled conduit
model (Neuberg et al., 2000). A wide variety of volcanic processes can produce the excitation mechanism that triggers crack or conduit resonance, as particularly for lava dome growth for andesitic volcanoes
(Chouet & Matoza, 2013; Morgan et al., 2008; Neuberg et al., 2000).
2. Tremors (TRs; 12,309 observations): These are defined by a sustained amplitude that can last from tens
of seconds to days, and they occur over a frequency range from 1 to 9 Hz (McNutt, 1992). This author
reported that many characteristics of LP events, and in some cases also of explosion quakes (see below),
are commonly associated with TRs. Virtually all eruptions are accompanied by TRs (McNutt, 1992). Visual
observations at Ubinas Volcano suggest that TRs are associated with magma extrusion and sporadic or
continuous gas and ash emissions (Macedo et al., 2009).
3. Explosions (EXPs; 159 observations): These are associated to sudden magma extrusion, and ash and gas
emission. Physically, they are related to fragmentation processes in the conduit, as has been observed for
many andesitic volcanoes (Druitt et al., 2002; Iguchi et al., 2008; Ohminato, 2006). Inza et al. (2011) noted
that for Ubinas Volcano, EXPs are related to destruction of the magmatic plug. Traversa et al. (2011) showed
that there is an overall acceleration of the LP events rate above the background level over the 2 to 3 hr
before vulcanian EXPs for Ubinas Volcano. This occurrence of a large number of LP events before EXPs is
consistent with the dome growth process.
MALFANTE ET AL.
5
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Figure 4. Illustration of the architecture used to build a model (light green), and the cross-validation process used to
estimate its performance (dark green). Steps 1, 2, and 3 are used to train a model from a labeled data set of
observations, [O(t)]. The observation O(t)ij is the j-est observation of class ci , with 1 ≤ i ≤ , 1 ≤ j ≤ Ni , and Ni as the
number of observations considered for class ci . Fji is its corresponding feature vector. A fraction 𝛼 with 0 < 𝛼 ≤ 1 of the
data set is used for learning, and the remaining data are used for testing. The process is repeated 50 times, with random
trials of learning and testing observations. Step 1: preprocessing, Step 2: features extraction, Step 3: learning, and
Step 4: testing (or using) the model. SVM = support vector machine; RF = random forest.
4. Volcano-tectonic (VT) events (1,315 observations): These are brittle-failure events. They are associated to
stress changes that are induced by magma movement (Chouet & Matoza, 2013). There are relatively few
VT events at Ubinas Volcano, compared to LP events, but their number increased from 2006 to 2011. VT
events are more numerous at the end of an eruption period.
5. Hybrid (HYB) events (474 observations): These have characteristics of both VT and LP events, with high frequency of onset followed by low frequencies. HYB events were introduced by Lahr et al. (1994) to describe
events observed at Redoubt Volcano. They have also been observed for Soufrière Hills Volcano, Montserrat
Volcano (White et al., 1998), and Mount St. Helens Volcano (Harrington & Brodsky, 2007), where they are
related to dome growth.
6. Tornillos (TORs; also known as screw events; 109 observations): These have a very limited distribution of
frequencies and a very slowly decaying coda. They are related to resonating fluid-filled conduits or cavities. TORs can be considered as a specific type of LP event with a long-duration coda that is composed
of harmonic oscillations. They were observed at Galeras Volcano before several eruptions in 1992 and
1993 (Narváez et al., 1997). TORs are rare for Ubinas Volcano, but they appear to be more common at the
beginning and end of eruptive periods.
See Figure 2 for an example of the waveforms and spectrograms associated to each of these classes. The
number of events of different types per day throughout the time period of the study is showed in Figure 3.
There is no class of rockfalls in our list. Generally speaking, rockfalls do not constitute a significant activity at
Ubinas, perhaps because this part of the Andes is very dry. During the period 2006–2011, Ubinas was in eruption with episodes of dome growth, explosions, and very regular ash emission generating seismic events of
higher amplitude compare to any kind of rockfalls. It is probable that if rockfalls have occurred, the generated
signals were masked by the magmatic activity or not selected due to low amplitudes compared to eruptive
signals.
3. Methods
We here detail the classification architecture proposed. From a large data set of observations of volcano seismic events, a machine learning algorithm is used to train a model to classify new events into one of the
six classes considered (Figure 2). The system architecture is described in Figure 4. The data considered for
the input were a labeled data set, as described in section 2.2. = 6 different classes of observations are
considered. For each class ci (with 1 ≤ i ≤ ), Ni observations were labeled and are taken into account. All of
the observations do not necessarily have the same length. We detail Steps 1 and 2 below that transform the
labeled data set of observations into a labeled data set of feature vectors. Step 3 describes the learning process where the prediction model is actually built. Finally, a fourth step of testing can be included to estimate
the model performances, as is explained in section 4.
MALFANTE ET AL.
6
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Step 1: Preprocessing. A first step of signal preprocessing is needed to standardize the observations. Typically,
observations that were recorded from the same station but with different sensors need to be harmonized
(e.g., due to change of sensor following technical failure). In particular, a filtering above 1 Hz is performed.
The instrumental response was not removed from the original signals. Preprocessing also includes the automatic or manual detection of relevant seismic observations of volcanic events that will constitute the data
set. Observations considered in this catalog were detected either manually by the observatory or automatically using the short-term average/long-term average method. Each observation is also normalized in terms
of its energy. The signal energy is computed and normalized to 1. This is done to build a model that is available for all observations, regardless of their amplitude level. It is worth noting that most state-of-the-art
methods are based on the energy levels (Langer et al., 2006) and therefore cannot be used to detect observations of low energy. Finally, a data set of N = 109,609 observations was considered. We also fixed the
maximum of 5 min per observation, in order to build a model that could permit near real-time analysis.
Step 2: Feature extraction. As specified in sections 1.2 and 1.3, a model sees the data exclusively in the feature space; that is, observations are only considered through their associated feature vectors. The choice of
features used to represent an observation is therefore one of the major and crucial issues when designing
a classification architecture. In the present study, we reviewed, gathered, and selected D = 102 features
that have been used in many classification tasks from various domains (see section 1.3). The use of a large
number of features helps to represent the observations in as thorough a way as possible, to be able to
describe their separating properties. We organized these features into three categories, namely, statistical,
entropy-based, and shape descriptors. The definitions of these features are given in Table 1.
1. Statistical features are valued for their immediate interpretation in terms of the signal shapes. We
used means and standard deviations and also higher moments, such as skewness and kurtosis,
which measure the asymmetry and flatness of the signal, respectively, compared to the Gaussian
distribution.
2. Entropy features come from information theory, and these are used to measure the information
content of a signal. We used Shannon entropy and Rényi measurements (Esmaili et al., 2004;
Huang et al., 2009).
3. Shape descriptors can be used, for instance, maximum over average values. These help to describe the
shape of a signal and therefore its physical properties. Rates of initiation and decay are particularly
useful for this purpose.
Furthermore, we propose to extract these features from three different representations of the observations.
Specifically, all of the features were extracted from three representations:
1. Time domain. Features extracted from the observation x[t] are used to describe the waveform shape
and its specificities (Time).
2. Frequency domain. A Fourier transform ( {⋅}) of the discrete temporal signal x[t] leads to X[f ] =
{x[t]}, which represents the spectral content of the observation (Frequency).
3. "Cepstral" domain. This domain is commonly used in speech processing to describe harmonic properties of the signal, that is, the periodic properties of the spectrum. The features are then extracted
{
}
from ⋅[q] = ⋅ |X[f ]| (Cepstral).
Eventually, each observation is represented by 34 features (i.e., 9 statistical, 9 entropy, and 16 shape descriptor
features) that are extracted in three different domains, which leads to a feature vector of dimension D =
34 × 3 = 102. The original data dimension is thereby reduced while maintaining the information content of
the signals (as a comparison, an original 30-s-long observation has a dimension of 3,000). Using general signal
shape descriptors leads to a precise description of the signal properties. Extraction of these descriptors from
the three different domains (i.e., temporal, spectral, and cepstral) underlines different properties of the original
signal and eventually leads to a thorough description of each observation. Of note, the feature extraction
process used for these data can also be applied to other transient signals.
The labeled data set of feature vectors is then considered for the next step.
Step 3: Learning. This final step can be referred to as the model learning or training. Here a prediction model is
built from a learning algorithm and a data set of labeled feature vectors. Several learning algorithms can be
used for this task (leading, in many cases, to similar results). The choice of the optimal learning algorithm
for a given application is empirical. However, if the feature space has been correctly chosen, the learning
MALFANTE ET AL.
7
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Table 1
List of Features
Statistic features
Feature
Definition
Length
n = length(s)
∑
𝜇s = n1 i s[i]
√
1 ∑
2
𝜎s = (n−1)
i (s[i] − 𝜇s )
)3
(
∑
s[i]−𝜇s
1
Mean
Standard deviation
Skewness
n
1
n
Kurtosis
Used in
i of central energy
Langet (2014) and Hibert et al. (2014)
5
(Tucker & Brown, 2005)
6
Tucker and Brown (2005)
7
̄3
i (i−i) Ei
E⋅B3i
Tucker and Brown (2005)
8
(i−̄i)4 Ei
E⋅B4i
Tucker and Brown (2005)
9
𝜎s
√∑
i
Entropy features
2
4
i
Mean kurtosis
Tucker and Brown (2005)
Langet (2014) and Hibert et al. (2014)
i
𝜎s
∑ ( s[i]−𝜇s )4
Mean skewness
1
3
̄i = 1 ⋅ ∑ Ei ⋅ i
√E∑ i
Bi = 1E i i2 ⋅ Ei − ̄i2
√∑
RMS bandwidth
Ref.
Tucker and Brown (2005)
(with p(sj ) the probability of amplitude level sj )
Feature
Definition
)
(
p(sj ) log2 p(sj )
j
)
(
∑
1
𝛼
p(sj )
⋅ log2
1−𝛼
Shannon entropya
−
Rényi entropyb
Ref.
∑
Esmaili et al. (2004) and Han et al. (2011)
10 to 12
Han et al. (2011)
13 to 18
j
Shape descriptor features
Feature
Rate of attack
Rate of decay
Ratios
Energy descriptors
Specific values
Definition
(
)
Ref.
s[i]−s[i−1]
n
(
)
mini s[i]−s[i+1]
n
maxi
min/mean and max/mean
Signal energy, maximum, average, standard deviation, skewness, and kurtosis
Tucker and Brown (2005)
19
Tucker and Brown (2005)
20
Langet (2014) and Hibert et al. (2014)
21 to 22
Tucker and Brown (2005)
23 to 28
min, max, i of min, i of max, threshold crossing rate, and silence ratio
Tucker and Brown (2005)
29 to 34
∑
n
Note. Features computed for a signal s[i]ni=1 (in which i might correspond to a temporal, frequency, or cepstral sample). E = i=1 s[i]2 and Ei = s[i]2 describe
the signal energy and the energy at sample i, respectively. Some features have a dimension greater than others; e.g., entropy measurements are made on three
different estimations of the amplitude probability (i.e., different histogram bin numbers).
a Bin numbers for probability estimation: 5, 30, and 500. b Bin numbers for probability estimation: 5, 30, 500, 𝛼 = 2, and inf .
algorithm choice should not have a major influence on the results. In this study, we chose to use random
forest and SVMs, as both of these are state-of-the-art techniques in machine learning.
Step 4: Testing. An extra step of testing can also be considered to evaluate the performance of a model and
thus to validate the classification architecture and, in particular, the feature choice. In this study, we used
cross validation, the concept behind which is illustrated in Figure 4 and is explained, in section 4.1.
4. Results and Their Analysis
The code used for this study is based on the Automatic Analysis Architecture that we developed and is available at Malfante (2018). It uses Python 3, and all tests were conducted on a MacBook Pro laptop. All results
are measured in term of accuracy and precision. Accuracy describes the good classification rate within a class
(see equation (1)), and precision describes the rate of observations belonging the the predicted class among
all the prediction of the class (see equation (2)).
MALFANTE ET AL.
Accuracy ci =
#(Observations of ci predicted as ci )
#(Observations of ci )
(1)
Precision ci =
#(Observations of ci predicted as ci )
.
#(Observations predicted as ci )
(2)
8
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Table 2
Classification of the General Results
Features
All
Most valuable features
Valuable features
Time
Frequency
Dimension
102
3
13
34
34
34
SVM
92.1 ± 0.54%
84.4 ± 0.50%
86.9 ± 0.60%
83.9 ± 0.89%
93.5 ± 0.50%
78.4 ± 1.0%
RF
92.5 ± 0.45%
84.2 ± 0.75%
90.3 ± 0.52%
87.2 ± 0.57%
91.3 ± 0.45%
79.3 ± 0.76%
Cepstral
Note. Comparisons of the accuracies of the results using random forest versus support vector machine, for the different feature sets.
4.1. First Test: Method Validation
To validate the proposed classification scheme, we perform cross-validation analysis on the first year of the
data set. To consider relatively balanced classes, we gather 1 year of observations for classes that are highly
frequent (i.e., LP events, TRs, and VT events), and all of the observations available for the less frequent classes
(i.e., HYB events, EXPs, and TORs). For each class, a fraction 𝛼 of the observations is used to train the model.
Because all the learning data need to be loaded in the computer memory at the same time, the number of
training observation per class is also limited to Nmax = 800 for computational reasons. The remaining data are
used for the test stage. Formally, if 0 < 𝛼 ≤ 1 stands for the learning rate, Ni for the number of observations
considered for the class ci with 1 ≤ i ≤ and the number of classes, then min(Ni ⋅ 𝛼, Nmax ) observations are
)
(
therefore considered for each class for the training step. The remaining max Ni ⋅ (1 − 𝛼), Ni − Nmax observations per class are used to test the model that is produced. The random selection of learning and testing data
among all the available observations is repeated 50 times to consider statistically valid results. In the following, the results are expressed as means and standard deviation over the 50 trials. The cross-validation results
for the comparison of the two learning algorithms (i.e., random forest and SVMs) and the four feature sets (i.e.,
Time, Frequency, Cepstral, and All together) are given in Table 2.
Using the All features, the accuracy of the results reaches 92.5% ± 0.54% (i.e., a good classification rate),
thereby showing the effectiveness of the proposed architecture and the feature choice. The results also
validate the limited influence of the learning algorithm (i.e., random forest vs. SVMs), which was expected
(92.5% ± 0.45% vs. 92.1% ± 0.54%, respectively, when considering All features). The influence of the features
used for the automatic classification, however, is much more important. These results vary from 78.4% ± 1.0%
when using cepstral features and to 93.5% ± 0.50% when using frequency features. It is particularly
interesting to note that for these data and this application, the frequency features appear to be particularly
discriminative and, in particular, that manual classification is often based on the frequency content of the
observations.
The confusion matrix is given in Table 3, and it provides more details about the class-by-class results. For
instance, the best classified class is for the LP events, with 58,363 correctly classified observations compared
to 62,030 considered (94.1%), followed closely by the VT events, with an accuracy of 92.2%. The confusion
matrix can also be used to analyze the model limitations. Most of the prediction errors are divided up between:
(i) LP events mixed with TRs, (ii) HYB events mixed with VT and LP events, and (iii) EXPs mixed with HYB
Table 3
Confusion Matrix
True Class (ground truth)
Predicted
Class
LP
TR
VT
EXP
HYB
TOR
Precision
LP
58,363
627
8
0
5
1
98.9%
TR
3,000
4,584
0
1
2
0
60.4%
VT
478
11
475
5
11
3
48.3%
EXP
15
16
2
29
0
0
47.8%
HYB
131
3
28
13
125
0
41.7%
35.9%
TOR
Accuracy
43
94.1%
4
3
0
0
28
87.4%
92.2%
59.8%
87.1%
84.6%
Note. Learning rate 𝛼 = 70%; the model was trained using support vector machine (RBF kernel, CRBF = 10, 𝛾 =
0.01) and feature set Frequency, with cross validation with 50 trials. Overall accuracy: 93.5%. Bold values mean
number of good classifications.
MALFANTE ET AL.
9
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Figure 5. Importance and selection of features. Bottom subplot: Feature weights (mean weights on 1-year
cross-validation models, trained with random forest on All features; 100 trees; 𝛼 = 0.7). Middle subplot: mean
accuracies (models trained with random forest; 100 trees; on the d-st most important features with 1 ≤ d ≤ D, D = 102).
Top subplot: mean accuracies for each class ci , 1 ≤ i ≤ C . The three most valuable features (MVF; filled squares)
and the 13 valuable features (VF; filled and empty squares) are indicated. Features are referenced with T, F, or C
depending on their computation domain (time, frequency, and ceptral, respectively), and their reference
numbers are as given in Table 1.
events. Those mistakes are seen through the precision rates, which are related the the false detection rates,
and are low for some classes. Physical interpretation of these results is valuable, as all of the errors made
by the model translate into physical similarities between the signals. For example, to parallel the main three
prediction errors above: (i) LP events and TRs are in the same frequency range and can overlap, and typically,
a LP event can be found within a TR. This will confuse the model, which predicts a single class at a given
time. Macedo et al. (2009) also showed that on some occasions, LP events occur in a repeated way and can
be separated at the beginning by some tens of seconds, before merging into a TR (e.g., before an EXP). (ii)
HYB events have characteristics of both LP and VT events, and they can belong to one class or the other.
Finally, (iii) EXPs and HYB events have similar frequency contents, and EXPs can produce both low and high
frequencies. The analysis of this error is thus particularly valuable, as these results can help volcanologists to
better analyze the seismicity, the relations between classes of events, and their evolution with time. The ability
of these methods to process very large data sets of recordings is essential for volcanic observatories.
4.2. Second Test: Features Selection
As previously shown and explained, the feature choice is decisive to obtain good results. We here investigate
the possibility to reduce the feature dimension through the selection of the relevant features from among the
whole feature set proposed here. Considering that all of the proposed features might not be optimal, the
feature vectors dimension should be kept as low as possible to avoid accuracy losses due to the curse of
dimensionality (see section 1.2). In particular, features that are strongly correlated can be left out. We intentionally did not use compression algorithms (e.g., principal component analysis), so as to maintain the physical
meanings of the features. Various approaches can be used to select d important features among a feature
set of dimension D, which include forward selection or backward elimination (Langley et al., 1994). However,
the exhaustive search for the d-st optimal features in terms of gain in accuracy is not feasible practically. As a
first approach here, we use a forward selection method, which ranks the features according to their weights
given in a random forest trained model. More specifically, random forest can rank features depending on their
impurity scores, whereby a feature that leads to a high decrease in impurity in the trees will have a high ranking. Nevertheless, the features that are not selected using this method are not necessarily meaningless. If two
features are strongly correlated, then only one will have a high score. Thus, while the random forest ranking
tool is a good approach for features selection, the nonselected features should not be considered as unnecessary when interpreting the data. For more details on this subject, the reader is referred to Quinlan (1986)
and Breiman (2001). Figure 5 shows the features ranked by importance using the impurity measurement and
MALFANTE ET AL.
10
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Figure 6. Monthly accuracy evolution for the LP events, with the model trained on the first 800 observations of each
class (all recorded in May 2006 for the LP events). LP = long-period.
shows the accuracy of the results when the observations are represented by feature vectors of dimension d
with 1 ≤ d ≤ D, composed of the d most important features.
From Figure 5, we can extract the two subsets of the most important features, as the three most valuable features (Figure 5, MVF, filled squares) and the 13 valuable features (Figure 5, VF, filled and
empty squares, which include the most valuable features). These features were selected as subsets,
which led to a notable gain in accuracy with a reasonably small dimension. With just the most valuable
features, the accuracy reaches 84.4% ±0.50%. This result is interesting, as it shows that good classification
can be obtained with a very restrictive number of features. In particular, this is of importance for real-time
or embedded systems and applications. Indeed, the most valuable features contain the three most
important features (3% of the total features), which are (i) F4: the spectrum skewness, which measures the
spectrum asymmetry; (ii) F5: the spectrum kurtosis, which measures the spectrum peakness; and (iii) T7: the
RMS bandwidth in time, which measures the mean length of the signal around which the energy is accumulated. The physical analysis of these features is, as such, a good indicator of the signals for the experts. In
particular, while the features rank based on random forest can be used to select such relevant features, these
features are themselves interesting to analyze, as they gather and embody the information-discriminating
signals across the classes. Similarly, the 13 most important features were extracted (13% of the features) for
the feature set of the valuable features (which includes the most valuable features). This feature set is particularly interesting as it provides a performance that is close to that of the whole feature set,
at 90.3% when using random forest. It is also interesting to note that the valuable features are features from the three computation domains, as time, frequency, and cepstral, thereby confirming the interest
in considering these three different domains for signal representation.
4.3. Third Test: Analysis Over 6 Years of Recordings
In this section, we propose to build a model using the Nmax = 800 first observations of each class recorded
from May 2006 and to test it over the 6 years of recording. In particular, we focus our attention on the LP
events, as these were numerous throughout the recordings (95,243 observations over 6 years). The monthly
evolution of the classification results is shown in Figure 6. Specifically, the analysis reveals three different
phases that start in May 2006, May 2007, and December 2010, where the mean accuracies are very different from one phase to the other. In particular, the accuracy collapses in May 2007, which reveals a significant
change in the signal shapes compared to the training observations (May 2006). This period also corresponds
to a sharp decrease in the number of LP events.
This was discussed with the Ubinas Volcano Observatory, in Arequipa. A manual revision of the seismic signals for the period of May–July 2007 was performed to determine whether the observatory analysis might
have confused LP events with other signals. Indeed, the classification criteria had been improved since the
MALFANTE ET AL.
11
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
beginning of the creation of the catalog in 2006, through the experience acquired during the other eruptions
of Ubinas Volcano, as well as through the signal classification of Ticsani and Sabancaya Volcanoes. The criteria
used for the reclassification are those that are now used for all of the Peruvian volcanoes. This revision analysis
thus showed a difference between the two classification systems starting from 25 May 2007. Some of the LP
events were indeed mislabeled, essentially as VT events or TRs, or to a lesser extent, as HYB events. Part of the
accuracy collapse observed from 25 May 2007 can thus be explained by this confusion between the classes.
At the same time, the new classification showed a similar decrease in the number of LP events, which would
instead suggest a change in activity, and therefore a change in the characteristics of these signals. Macedo
et al. (2009) observed strong temporal variations of the degassing and seismic activity in a period starting
in November 2006. A downward migration of magma in the crater has been observed on images taken in
December 2006, compared to previous images taken in April and October 2006 (Figure 5 in Macedo et al.,
2009). The receding of magma in the conduit has been observed over several months. Other images of
the crater taken by IGP Arequipa Observatory respectively on 17 April, 8 June, and 26 August 2007 show
a clear drop of the magma level between April and June 2007. The backward migration of magma in the
conduit implies modifications of seismic source positions and possibly mechanisms due to the complexity
of the geometry of the conduit and time modifications of the coupling between magma and the conduit
walls. Magma migration can also affect local stress conditions or conduit cavity properties. This can obviously
explain modifications of the characteristics of the LP events. The decrease of LP accuracy is starting in January
2007 and its strengthening between April and May 2007, which is coincident with observations of volcanic
activity. The interesting point is that these modifications were not detected by manual classification, while
the automatic classification perfectly identified them. This result is particularly important, as the use of automatic methods of classification has revealed inconsistencies in the original manual classification. This thereby
validates the use of such methods to help with the monitoring task carried out by volcano observatories. The
automatic analysis revealed a modification of interpretation of the signals for the LP events, which was not
detected in the manual analysis.
5. Conclusion and Prospects
The present study is related to the classification of volcanic events based on an analysis of the seismic signals. Evaluation of the volcanic risk remains a timely and open issue, which triggers strong interest in the
international community for approaches to automatically process large databases of recordings. Automatic
classification tools for signals that might be either precursors or indicators of volcanic active phases are particularly needed. In this paper, we present our classification architecture that is based on supervised machine
learning. The main idea is to use a labeled data set of targeted signals to build a prediction model. This prediction model can then be used to automatically analyze newly recorded signals. Our main contribution consists
of the gathering of the large number of features that are used to represent the data. We considered 102 features that we tested on 109,609 seismic observations acquired for Ubinas Volcano, which is the most active
volcano in Peru. The features give a general and precise description of transient signals and can be used for
other applications. Our model reaches an accuracy of 93.5% ±0.50% over the labeled set of data, thereby validating the effectiveness of the proposed classification scheme and the chosen features for signal description.
Furthermore, the analysis of the whole data set used for our method revealed a clear change in the LP event
observations linked with a significant evolution of the magmatic activity, which had not been identified by the
manual classification. The learned model actually led to a more precise analysis. Further analysis showed that
the accuracy of the results can be maintained when only a subset of the most important features is used, as
obtained through random forest features ranking. Comparisons of the two learning algorithms (i.e., random
forest and SVM) also confirmed the limited influence these have if the feature space has been well chosen.
The prospects for this study indicate the implementation of such tools in volcano observatories to be run in
real time, the development of a confidence index, and the inclusion of a detection stage to analyze continuous
recordings. More investigation on the feature set could also be conducted, typically features computed from
spectrogram representation of the events could be considered, and deep neural network could be used to
represent the data. Another perspective would be to use semisupervised or unsupervised learning algorithms
to relax the dependency on a labeled data set, which today limits the use of supervised models. The use of
machine learning models can also be used to process continuous signals instead of discrete events. This last
prospect is currently under investigation for future work.
MALFANTE ET AL.
12
Journal of Geophysical Research: Solid Earth
Acknowledgments
This work was supported by a grant
from Labex OSUG@2020
(Investissements d’avenir-ANR10
LABX56) and DGA/MRIS. GIPSA-lab
SIGMAPHY is part of Labex OSUG@2020
(ANR10 LABX56). The authors thank the
IDEX for funding a travel grant
(M. Malfante). The code used in this
work is available at https://github.com/
malfante/AAA. The catalogue of data
can be found on IGP website
https://ovs.igp.gob.pe/catalogosvulcanologicos. Raw data can be
obtained by requesting them to one of
the coauthors from IGP, Orlando
Macedo or Adolfo Inza. This work was
partially supported by the project
VOSICA in the framework of the
Grenoble Alpes Data Institute
(ANR-15-IDEX-02).
MALFANTE ET AL.
10.1029/2018JB015470
References
Apolloni, B., Bassis, S., & Morabito, F. (2009). Neural Nets WIRN09-Proceedings of the 19th Italian Workshop on Neural Nets (pp. 116–123). Vietri
sul Mare, Salerno, Italy.
Bellman, R. (1956). Dynamic programming and Lagrange multipliers. Proceedings of the National Academy of Sciences of the United States of
America, 42(10), 767–769. https://doi.org/10.1073/pnas.42.10.767
Benítez, M. C., Ramírez, J., Segura, J. C., Ibáñez, J. M., Almendros, J., García-Yeguas, A., & Cortés, G. (2007). Continuous HMM-based
seismic-event classification at Deception Island, Antarctica. IEEE Transactions on Geoscience and Remote Sensing, 45, 138–146.
https://doi.org/10.1109/TGRS.2006.882264
Breiman, L. (2001). Random forests. Machine Learning, 45, 5–32. https://doi.org/10.1023/A:1010933404324
Cannata, A., Montalto, P., Aliotta, M., Cassisi, C., Pulvirenti, A., Privitera, E., & Patan, D. (2011). Clustering and classification of infrasonic
events at Mount Etna using pattern recognition techniques. Geophysical Journal International, 185, 253–264. https://doi.org/10.1111/
j.1365-246X.2011.04951.x
Chouet, B. (1986). Dynamics of a fluid-driven crack in three dimensions by the finite difference method. Journal of Geophysical Research,
91(B14), 13,967–13,992. https://doi.org/10.1029/JB091iB14p13967
Chouet, B. A. (1996). Long-period volcano seismicity: Its source and use in eruption forecasting. Nature, 6572, 309–316. https://doi.org/
10.1038/380309a0
Chouet, B. A., & Matoza, R. S. (2013). A multi-decadal view of seismic methods for detecting precursors of magma movement and eruption.
Journal of Volcanology and Geothermal Research, 252(Supplement C), 108–175. Retrieved from https://www.sciencedirect.com/science/
article/pii/S0377027312003435. https://doi.org/10.1016/j.jvolgeores.2012.11.013
Cortés, G., Arámbula, R., Gutiérrez, L., Benítez, C., Ibánez, J., Lesage, P., & Garcia, L. (2009). Evaluating robustness of a HMM-based classification system of volcano-seismic events at Colima and Popocatepetl volcanoes. In Geoscience and Remote Sensing Symposium, 2009 IEEE
International, IGARSS 2009 (Vol. 2, pp. II-1012).
Cortes, C., & Vapnik, V. (1995). Support-vector networks. Machine Learning, 20(3), 273–297. https://doi.org/10.1007/BF00994018
Druitt, T. H., Young, S. R., Baptie, B., Bonadonna, C., Calder, E. S., Clarke, A. B., & Voight, B. (2002). Episodes of cyclic Vulcanian explosive
activity with fountain collapse at Soufrière Hills Volcano, Montserrat. Geological Society, London, Memoirs, 21(1), 281–306.
https://doi.org/10.1144/GSL.MEM.2002.021.01.13
Eronen, A., & Klapuri, A. (2000). Musical instrument recognition using cepstral coefficients and temporal features. In Acoustics, Speech, and
Signal Processing, 2000. ICASSP’00. Proceedings. 2000 IEEE International Conference on (Vol. 2, pp. II753–II756).
Esmaili, S., Krishnan, S., & Raahemifar, K. (2004). Content based audio classification and retrieval using joint time-frequency analysis. In
Icassp-2004, ieee international conference on acoustics, speech, and signal processing (Vol. 5, pp. 9–12).
Esposito, A., Giudicepietro, F., D’Auria, L., Scarpetta, S., Martini, M., Coltelli, M., & Marinaro, M. (2008). Unsupervised neural analysis of
very-long-period events at Stromboli volcano using the self-organizing maps. Bulletin of the Seismological Society of America, 98(5),
2449–2459.
Fagerlund, S. (2007). Bird species recognition using support vector machines. EURASIP Journal on Advances in Signal Processing, 2007(1),
639–645.
Friedman, J., Hastie, T., & Tibshirani, R. (2001). The Elements of Statistical Learning (Vol. 1). Berlin: Springer series in statistics Springer.
Fujinaga, I., & MacMillan, K. (2000). Realtime recognition of orchestral instruments. In Proceedings of the International Computer Music
Conference (ICMC2000) (Vol. 141, pp. 141–143).
Gutiérrez, L., Ibañez, J., Cortés, G., Ramírez, J., Benítez, C., Tenorio, V., & Isaac, Á. (2009). Volcano-seismic signal detection and classification
processing using hidden Markov models. Application to San Cristóbal volcano, Nicaragua. In Geoscience and Remote Sensing Symposium,
2009 IEEE International, igarss 2009 (Vol. 4, pp. IV–522).
Hammer, C., Beyreuther, M., & Ohrnberger, M. (2012). A seismic-event spotting system for volcano fast-response systems. Bulletin of the
Seismological Society of America, 102(3), 948–960.
Han, N. C., Muniandy, S. V., & Dayou, J. (2011). Acoustic classification of Australian anurans based on hybrid spectral-entropy approach.
Applied Acoustics, 72(9), 639–645.
Harrington, R. M., & Brodsky, E. E. (2007). Volcanic hybrid earthquakes that are brittle-failure events. Geophysical Research Letters, 34, L06308.
https://doi.org/10.1029/2006GL028714
Hibert, C., Mangeney, A., Grandjean, G., Baillard, C., Rivet, D., Shapiro, N. M., et al. (2014). Automated identification, location, and
volume estimation of rockfalls at Piton de la Fournaise volcano. Journal of Geophysical Research: Earth Surface, 119, 1082–1105.
https://doi.org/10.1002/2013JF002970
Hibert, C., Provost, F., Malet, J. P., Maggi, A., Stumpf, A., & Ferrazzini, V. (2017). Automatic identification of rockfalls and volcano-tectonic
earthquakes at the Piton de la Fournaise volcano using a random forest algorithm. Journal of Volcanology and Geothermal Research, 340,
130–142.
Huang, C. J., Yang, Y. J., Yang, D. X., & Chen, Y. J. (2009). Frog classification using machine learning techniques. Expert Systems with
Applications, 36(2), 3737–3743.
Ibáñez, J. M., Benítez, C., Gutiérrez, L. A., Cortés, G., García-yeguas, A., & Alguacil, G. (2009). The classification of seismo-volcanic signals
using Hidden Markov Models as applied to the Stromboli and Etna volcanoes. Journal of Volcanology and Geothermal Research, 187(3-4),
218–226. https://doi.org/10.1016/j.jvolgeores.2009.09.002
Iguchi, M., Yakiwara, H., Tameguri, T., Hendrasto, M., & ichi Hirabayashi, J. (2008). Mechanism of explosive eruption revealed by geophysical observations at the Sakurajima, Suwanosejima and Semeru volcanoes. Journal of Volcanology and Geothermal Research, 178(1), 1–9.
https://doi.org/10.1016/j.jvolgeores.2007.10.010 (Dynamics of volcanic explosions: Field observations, experimental constraints and
integrated modelling of volcanic explosions: Field observations, experimental constraints and integrated modelling)
Inza, L. A., Mars, J. I., Métaxian, J. P., O’Brien, G. S., & Macedo, O. (2011). Seismo-volcano source localization with triaxial broad-band seismic
array. Geophysical Journal International, 187(1), 371–384.
Inza, L., Métaxian, J. P., Mars, J., Bean, C., O’Brien, G. S., Macedo, O., & Zandomeneghi, D. (2014). Analysis of dynamics of vulcanian activity of
Ubinas volcano, using multicomponent seismic antennas. Journal of Volcanology and Geothermal Research, 270, 35–52.
Lahr, J., Chouet, B., Stephens, C., Power, J., & Page, R. (1994). Earthquake classification, location, and error analysis in a volcanic environment:
Implications for the magmatic system of the 1989–1990 eruptions at Redoubt Volcano, Alaska. Journal of Volcanology and Geothermal
Research, 62(1-4), 137–151.
Langer, H., Falsaperla, S., Powell, T., & Thompson, G. (2006). Automatic classification and a-posteriori analysis of seismic event
identification at Soufrière Hills volcano, Montserrat. Journal of Volcanology and Geothermal Research, 153(1-2 SPEC. ISS.), 1–10.
https://doi.org/10.1016/j.jvolgeores.2005.08.012
13
Journal of Geophysical Research: Solid Earth
10.1029/2018JB015470
Langet, N. (2014). Détection et caractérisation massives de phénomènes sismologiques pour la surveillance d’événements traditionnels
et la recherche systématique de phénomènes rares (Doctoral dissertation). Ecole doctorale des Sciences de la Terre et de l’environment,
Université de Strasbourg. Retrieved from https://www.theses.fr/2014STRAH017
Langley, P. A. T., Flamingo, L., & Edu, S. (1994). Selection of relevant features in machine learning, 127–131.
Macedo, O., Métaxian, J. P., Taipe, E., Ramos, D., & Inza, A. (2009). Seismicity associated with the 2006-2008 eruption, Ubinas volcano,
262–270.
Maggi, A., Ferrazzini, V., Hibert, C., Beauducel, F., Boissier, P., & Amemoutou, A. (2017). Implementation of a multistation approach for
automated event classification at Piton de la Fournaise Volcano. Seismological Research Letters, 88(3), 878–891. https://doi.org/10.1785/
0220160189
Malfante, M. (2018). Automatic analysis architecture. Retrieved from blabla.
McNutt, S. (1992). Volcanic tremor. Encyclopedia of Earth System Science, 4, 417–425.
McNutt, S. R. (1996). Seismic monitoring and eruption forecasting of volcanoes: A review of the state-of-the-art and case histories,
Monitoring and Mitigation of Volcano Hazards (pp. 99–146). Berlin, Heidelberg: Springer Berlin Heidelberg.
Morgan, S. C., Malone, S. D., Qamar, A. I., Thelen, W. A., Wright, A. K., & Caplan-Auerbach, J. (2008). Seismicity associated with renewed
dome building at Mount St. Helens, 2004-2005: Chapter 2 in A volcano rekindled: The renewed eruption of Mount St. Helens, 2004-2006
(Tech. Rep.) Reston, VA: U.S. Geological Survey. from https://pubs.er.usgs.gov/publication/pp17502 (Report).
Narváez, L., Torres, C. R. A., Gómez, M. D. M., Cortés, J. G. P., Cepeda, H., & Stix, J. (1997). ’Tornillo’-type seismic signals at Galeras volcano,
Colombia, 1992–1993. Journal of Volcanology and Geothermal Research, 77(1), 159–171. https://doi.org/10.1016/S0377-0273(96)00092-3
Neuberg, J., Luckett, R., Baptie, B., & Olsen, K. (2000). Models of tremor and low-frequency earthquake swarms on Montserrat. Journal of
Volcanology and Geothermal Research, 101(1), 83–104. https://doi.org/10.1016/S0377-0273(00)00169-4
Ohminato, T. (2006). Characteristics and source modeling of broadband seismic signals associated with the hydrothermal system at
Satsumaâiwojima volcano, Japan. Journal of Volcanology and Geothermal Research, 158(3), 467–490. https://www.sciencedirect.com/
science/article/pii/S0377027306003325
Ohrnberger, M. (2003). Continuous automatic classification of seismic signals of volcanic origin at Mt . Merapi, Java, Indonesia continuous
automatic classification of seismic signals of volcanic origin at Mt . Merapi, Java, Indonesia. (July 2001).
Quinlan, J. R. (1986). Induction of decision trees. Machine Learning, 1(1), 81–106. https://doi.org/10.1007/BF00116251
Traversa, P., Lengliné, O., Macedo, O., Métaxian, J. P., Grasso, J. R., Inza, A., & Taipe, E. (2011). Short term forecasting of explosions at Ubinas
volcano, Perú. Journal of Geophysical Research, 116, B1130. https://doi.org/10.1029/2010JB008180
Tucker, S., & Brown, G. J. (2005). Classification of transient sonar sounds using perceptually motivated features. IEEE Journal of Oceanic
Engineering, 30(3), 588–600.
White, R. A., Miller, A. D., Lynch, L., & Power, J. (1998). Observations of hybrid seismic events at Soufriere Hills Volcano, Montserrat: July 1995
to September 1996. Geophysical Research Letters, 25(19), 3657–3660. Retrieved from https://doi.org/10.1029/98GL02427
Zaugg, S., Van Der Schaar, M., Houégnigan, L., Gervaise, C., & André, M. (2010). Real-time acoustic classification of sperm whale clicks and
shipping impulses from deep-sea observatories. Applied Acoustics, 71(11), 1011–1019.
MALFANTE ET AL.
14