TY - JOUR
T1 - Data space reduction, quality assessment and searching of seismograms
T2 - Autoencoder networks for waveform data
AU - Valentine, Andrew P.
AU - Trampert, Jeannot
PY - 2012/5
Y1 - 2012/5
N2 - What makes a seismogram look like a seismogram? Seismic data sets generally contain waveforms sharing some set of visual characteristics and features-indeed, seismologists routinely exploit this when performing quality control 'by hand'. Understanding and harnessing these characteristics offers the prospect of a deeper understanding of seismic waveforms, and opens up many potential new techniques for processing and working with data. In addition, the fact that certain features are shared between waveforms suggests that it may be possible to transform the data away from the time domain, and represent the same information using fewer parameters. If so, this would be a significant step towards making fully non-linear tomographic inversions computationally tractable. Hinton & Salakhutdinov showed that a particular class of neural network, termed 'autoencoder networks', may be used to find lower-dimensional encodings of complex binary data sets. Here, we adapt their work to the continuous case to allow the use of autoencoders for seismic waveforms, and offer a demonstration in which we compress 512-point waveforms to 32-element encodings. We also demonstrate that the mapping from data to encoding space, and its inverse, are well behaved, as required for many applications. Finally, we sketch a number of potential applications of the technique, which we hope will be of practical interest across all seismological disciplines, and beyond.
AB - What makes a seismogram look like a seismogram? Seismic data sets generally contain waveforms sharing some set of visual characteristics and features-indeed, seismologists routinely exploit this when performing quality control 'by hand'. Understanding and harnessing these characteristics offers the prospect of a deeper understanding of seismic waveforms, and opens up many potential new techniques for processing and working with data. In addition, the fact that certain features are shared between waveforms suggests that it may be possible to transform the data away from the time domain, and represent the same information using fewer parameters. If so, this would be a significant step towards making fully non-linear tomographic inversions computationally tractable. Hinton & Salakhutdinov showed that a particular class of neural network, termed 'autoencoder networks', may be used to find lower-dimensional encodings of complex binary data sets. Here, we adapt their work to the continuous case to allow the use of autoencoders for seismic waveforms, and offer a demonstration in which we compress 512-point waveforms to 32-element encodings. We also demonstrate that the mapping from data to encoding space, and its inverse, are well behaved, as required for many applications. Finally, we sketch a number of potential applications of the technique, which we hope will be of practical interest across all seismological disciplines, and beyond.
KW - Neural networks, fuzzy logic
KW - Self-organization
KW - Statistical seismology
KW - Time-series analysis
KW - Wave propagation
UR - http://www.scopus.com/inward/record.url?scp=84859713259&partnerID=8YFLogxK
U2 - 10.1111/j.1365-246X.2012.05429.x
DO - 10.1111/j.1365-246X.2012.05429.x
M3 - Article
SN - 0956-540X
VL - 189
SP - 1183
EP - 1202
JO - Geophysical Journal International
JF - Geophysical Journal International
IS - 2
ER -