Abstract
We often apply a number of processing modules to a field dataset to perform various tasks, like denoising, first arrival picking, velocity estimation, and so on. Selecting the architecture, collecting labels, and tuning a deep learning network for a specific seismic processing task have been common practice in the field, yet these, often independent, networks do not take advantage of the fact that most processing tasks utilize the same features embedded in the seismic data that actually can be stored in a common pre-trained network in a self-supervised manner, we refer to as StorSeismic. In this framework, we utilize a Bidirectional Encoder Representations from Transformers (BERT) model to promote pre-training for storing the features of a seismic dataset and then efficiently fine-tune it to adapt to a wide spectrum of seismic processing tasks. Here, we combine field data, with synthetic generated data, in the self-supervised pre-training step to store the seismic features. Then, we use the labeled synthetic data to fine tune the pre-trained network in a supervised fashion to perform various seismic processing tasks, like denoising, low frequency extrapolation, first arrival picking, and velocity estimation, with satisfactory results.
Original language | English (US) |
---|---|
Pages | 1610-1614 |
Number of pages | 5 |
DOIs | |
State | Published - Aug 15 2022 |
Event | 2nd International Meeting for Applied Geoscience and Energy, IMAGE 2022 - Houston, United States Duration: Aug 28 2022 → Sep 1 2022 |
Conference
Conference | 2nd International Meeting for Applied Geoscience and Energy, IMAGE 2022 |
---|---|
Country/Territory | United States |
City | Houston |
Period | 08/28/22 → 09/1/22 |
Bibliographical note
Publisher Copyright:© 2022 Society of Exploration Geophysicists and the American Association of Petroleum Geologists.
ASJC Scopus subject areas
- Geotechnical Engineering and Engineering Geology
- Geophysics