Everything depends on V(x,y,z) (ps.gz 734K) (pdf 531K) (src 702K)
Claerbout J.
Estimating 3-D velocity V(x,y,z) is the most important problem in exploration
geophysics.
It is a very difficult problem. In order to really solve it, SEP has
turned to fundamentals of estimation theory
with topographic data,
regridding, interpolation, truncation, erratic noise, instrument drift, etc.
This return to
fundamentals has proven rewarding,
leading us to the helix discovery.
This discovery is
revitalizing wave equation migration in 3-D,
preconditioning
many estimations
(big speed up),
and
regularizing velocity estimation
(blends measured and prior information).
To enable young people to become productive with 3-D seismic data,
Biondo Biondi and Bob Clapp have built a 3-D seismic software
infrastructure that is able to address real 3-D problems,
such as V(x,y,z) and aliasing in 3-D.
This infrastructure is unique in the academic world.
None of the other academic organizations have enough computing power and
infrastructure to allow routine research activities with 3-D field data.
Wave-equation migration velocity analysis (ps.gz 1418K) (pdf 11489K) (src 32494K)
Biondi B. and Sava P.
In this report,
we introduce a new wave-equation method of migration velocity
analysis (MVA). The method is based on the linear relation that can be
established between a perturbation in the migrated image and the
generating perturbation in the slowness function. Our method consists
of two steps: we first improve the focusing of the migrated image and
then iteratively update the velocity model to explain the
improvement in the focusing of the image.
As a wave-equation method, our version of
MVA is robust and generates smooth slowness functions without model
regularization. We also show that our method has the potential to
exploit the power of residual prestack migration to MVA.
Preconditioning tau tomography with geologic constraints (ps.gz 1443K) (pdf 5017K) (src 4352K)
Clapp R. G. and Biondi B.
Seismic tomography is a non-linear problem with a significant
null-space. Our estimation problem often
converges slowly,
to a geologically unreasonable model, or not at all.
One reason for slow or non-convergence is that we are attempting
to simultaneously estimate reflector position (mapping velocity)
and image our data (focusiing velocity).
By performing tomography in vertical travel-time space, we avoid
estimating mapping velocity, instead concentrating on
focusing velocity. By introducing anisotropic preconditioning
oriented along bedding planes, we can quickly guide the inversion towards
a geologically reasonable model. We illustrate the benefits of
our tomography method by comparing it to more traditional methods
on a synthetic anticline model. In addition, we demonstrate
the method's ability to improve the velocity estimate, and the resulting
migrated image of a real 2-D dataset.
Why tau tomography is better than depth tomography (ps.gz 117K) (pdf 2318K) (src 779K)
Clapp R. G. and Biondi B.
Seismic tomography is a non-linear problem.
A standard technique is to itteratively assume a linear relation between
the change in slowness and the change in travel
times Biondi (1990); Etgen (1990) and then re-linearize
around the new model.
In ray-based methods, this amounts to assuming
stationary ray paths and reflection locations to
construct a back projection
...
Velocity continuation in migration velocity analysis (ps.gz 1619K) (pdf 6105K) (src 4199K)
Fomel S.
Velocity continuation can be applied to migration velocity
analysis. It enhances residual NMO correction by properly taking
into account both vertical and lateral movements of reflectors
caused by the change in migration velocity. I exemplify this fact
with simple data tests.
All stationary points of differential semblance are asymptotic global minimizers: layered acoustics (ps.gz 137K) (pdf 676K) (src 583K)
Symes W. W.
Differential semblance velocity estimators have well-defined and
smooth high frequency asymptotics. A version appropriate for analysis
of CMP gathers and layered acoustic models has no secondary
minima. Its structure suggests an approach to optimal parametrization
of velocity models.
Coherent noise suppression in velocity inversion (ps.gz 245K) (pdf 4070K) (src 240K)
Symes W. W.
Data components with well-defined moveout other than primary
reflections are sometimes called coherent noise.
Coherent noise makes velocity analysis ambiguous, since
no single velocity function explains incompatible moveouts
simultaneously. Contemporary data processing treats the
control of coherent noise influence on velocity as an interpretive
step. Dual regularization theory suggests an
alternative, automatic inversion algorithm for suppression of coherent
noise when primary reflection phases dominate the data. Experiments
with marine data illustrate the robustness and effectiveness of
the algorithm.
Angle-domain common image gathers by wave-equation migration (ps.gz 817K) (pdf 2210K) (src 2237K)
Prucha M. L., Biondi B. L., and Symes W. W.
Shot- and offset-domain common image gathers encounter problems in
complex media. They can place events that come from different
points in the subsurface at one subsurface location based on identical
arrival times and horizontal slownesses. Angle-domain common image
gathers uniquely define ray couples for each point in the subsurface,
therefore each event in the data will be associated with only one
subsurface location. It is possible to generate angle-domain common
image gathers with wave-equation migration methods and these angle-domain
common image gathers may be used for velocity analysis and
amplitude-versus-angle analysis. Applications of these methods to the
Marmousi model are promising.
Subsalt imaging by common-azimuth migration (ps.gz 507K) (pdf 6189K) (src 1728K)
Biondi B.
The comparison of subsalt images obtained by common-azimuth migration
and single-arrival Kirchhoff migration demonstrates the potential
of wave-equation migration when the velocity model causes complex
multipathing.
Subsalt reflectors are better imaged and the typical Kirchhoff
artifacts caused by severe multipathing disappear.
A detailed analysis of common-azimuth images
indicates that the results of common-azimuth
imaging could be improved.
It points to opportunities
to improve the numerical implementation as well as the downward
continuation method.
Extending common-azimuth migration (ps.gz 253K) (pdf 843K) (src 309K)
Vaillant L. and Biondi B.
We present a review of common-azimuth prestack depth migration theory and propose a new extension to the original method. In common-azimuth migration theory, source and receiver raypaths are constrained to lie on the same plane at each depth level. By using data with a broader range of cross-line offsets, we increase the number of raypaths examined and consider more information. Consequently, our extended common-azimuth migration is theoretically better able to model lateral velocity variations due to real
3-D structures and is more compatible with the standard marine acquisition geometry, in which cross-line offsets are concentrated in a narrow band. We first discuss the theory of the process, and then introduce computational issues leading to future implementation.
Comparing Kirchhoff with wave equation migration in a hydrate region (ps.gz 1181K) (pdf 4343K) (src 1236K)
Sinha M. and Biondi B.
Prestack migration is necessary before AVO analysis. Most of the
present migration algorithms not only try to focus the reflections on the
subsurface but also strive to preserve the amplitude for subsequent
amplitude studies. A migration/inversion method developed by
Lumley 1993 estimates the angle dependent
reflectivity at each subsurface point by using least-squares Kirchhoff
migration followed by a linearized Zoeppritz elastic parameter inversion for
relative contrasts in compressional and shear wave impedance. Another
migration algorithm is based on the wave-equation method which uses the Double
...
Angle-gather time migration (ps.gz 1296K) (pdf 5089K) (src 1674K)
Fomel S. and Prucha M.
Angle-gather migration creates seismic images for different
reflection angles at the reflector. We formulate an angle-gather
time migration algorithm and study its properties. The algorithm
serves as an educational introduction to the angle gather concept.
It also looks attractive as a practical alternative to conventional
common-offset time migration both for velocity analysis and for
AVO/AVA analysis.
On Stolt prestack residual migration (ps.gz 1440K) (pdf 4196K) (src 91891K)
Sava P.
Residual migration has proved to be a useful tool in imaging
and in velocity analysis.
Rothman 1983 shows that post-stack residual
migration can be successfully used to improve the focusing of the
migrated sections. He also showed that migration with a given velocity
vm is equivalent to migration with a reference velocity v0
followed by residual migration with a velocity vr that can be
expressed as a function of v0 and vm.
...
Anti-aliasing multiple prediction beyond two dimensions (ps.gz 166K) (pdf 4268K) (src 19192K)
Sun Y.
Theoretically, the Delft method of surface-related multiple elimination
can be applied in three dimensions, as long as the source and receiver
coverage is dense enough. In reality, such a dense coverage is still far
from reach, using the available multi-streamer acquisition system. One way
to fill the gap is to massively interpolate the missing sources and
receivers in the survey, which requires a huge computational cost. In
this paper, I propose a more practical approach for the multi-streamer
system. Instead of using large-volume missing-streamer interpolation,
my method finds the most reasonable proxy from the collected dataset
for each missing trace needed in the multiple prediction. Although this
approach avoids missing-streamer interpolation, another problem pops up in the
multi-streamer case, the aliasing noise caused by the sparse sampling
in the cross-line direction. To solve this problem, I introduce a new concept,
the partially-stacked multiple contribution gather (PSMCG).
Using multi-scale prediction-error filter (MSPEF) theory, this approach
interpolates the PSMCG in the cross-line direction
to remove the aliasing noise.
Acoustic daylight imaging via spectral factorization: Helioseismology and reservoir monitoring (ps.gz 465K) (pdf 762K) (src 1761K)
Rickett J. and Claerbout J.
The acoustic time history of the sun's surface
is a stochastic (t,x,y)-cube of information.
Helioseismologists cross-correlate these noise traces to
produce impulse response seismograms, providing the proof of
concept for a long-standing geophysical conjecture.
We pack the (x,y)-mesh of time series into a single
super-long one-dimensional time series.
We apply Kolmogoroff spectral factorization to the super-trace,
unpack, and find the multidimensional acoustic impulse
response of the sun.
State-of-the-art seismic exploration recording equipment
offers tens of thousands of channels, and permanent recording
installations are becoming economically realistic.
Helioseismology, therefore, provides a conceptual prototype
for using natural noises for continuous reservoir monitoring.
Interpolation with smoothly nonstationary prediction-error filters (ps.gz 723K) (pdf 3470K) (src 2376K)
Crawley S.
Building on the notions of time-variable filtering
and the helix coordinate system,
I develop software for filters that
are smoothly variable in multiple dimensions, but that
are quantized into large enough regions to be efficient.
Multiscale prediction-error filters (PEFs) can estimate dips
from recorded data and use the dip information to fill in
unrecorded shot or receiver gathers.
The data are typically divided into patches with
approximately constant dips,
with the requirement that the patches contain enough data samples to
provide a sufficient number of fitting equations to determine all
the coefficients of the filter.
Each patch of data represents an independent estimation problem.
Instead, I estimate a set of smoothly varying filters in much smaller patches,
as small as one data sample.
They are more work to estimate,
but the smoothly varying filters do give more accurate
interpolation results than PEFs in independent patches, particularly on
complicated data.
To control
the smoothness of the filters.
I use filters like directional derivatives
that Clapp et al. 1998 call ``steering filters''.
They destroy dips in easily adjusted directions.
I use them in residual space to encourage dips in the specified directions.
I describe the notion of ``radial-steering filters'' Clapp et al. (1999),
i.e., steering filters oriented in the radial direction
(lines of constant x/t in (t,x) space).
Break a common-midpoint gather
into pie shaped regions bounded by various values of x/t.
Such a pie-shaped region tends to have constant dip spectrum
throughout the region so it is a natural region
for smoothing estimates of dip spectra
or of gathering statistics
(via 2-D PEFs).
In this paper I use smoothly variable PEFs to interpolate missing traces,
though they may have many other uses.
Finally, since noisy data can produce poor interpolation
results, I deal with the separation of signal and noise along with
missing data.
Directional smoothing of non-stationary filters (ps.gz 664K) (pdf 2384K) (src 5961K)
Clapp R. G., Fomel S., Crawley S., and Claerbout J. F.
Space-varying prediction error filters are an effective tool in solving
a number of common geophysical problems. To estimate these
filters some type of regularization is necessary. An effective method
is to smooth the filters along radial lines in CMP gathers
where dip information
is relatively unchanging.
Texture synthesis and prediction error filtering (ps.gz 695K) (pdf 2155K) (src 1982K)
Brown M.
The spectrum of a prediction-error filter (PEF) tends toward the inverse spectrum of the
data from which it is estimated.
I compute 2-D PEF's from known ``training images'' and use them to synthesize
similar-looking textures from random numbers via helix deconvolution.
Compared to a similar technique employing Fourier transforms, the PEF-based method is
generally more flexible, due to its ability to handle missing data, a fact which I
illustrate with an example.
Applying PEF-based texture synthesis to a stacked 2-D seismic section,
I note that the residual error in the PEF estimation forms the basis for ``coherency''
analysis by highlighting discontinuities in the data, and may also serve as a measure
of the quality of a given migration velocity model.
Last, I relate the notion of texture synthesis to missing data interpolation and show
an example.
Polarity and PEF regularization (ps.gz 692K) (pdf 1327K) (src 759K)
Claerbout J.
We address the puzzle of seismic polarity.
Why do we rarely observe it clearly
and how could we be more systematic about trying to observe polarity?
This puzzle will lead us to long prediction-error filters.
Being long, they require many data samples.
If such a filter is nonstationary,
we might have an inadequate number of fitting equations.
Then we need regularization.
Here we consider some examples
and consider an efficient way to regularize the filter estimation.
...
Spectral factorization revisited (ps.gz 96K) (pdf 491K) (src 20K)
Sava P. and Fomel S.
In this paper, we review some of the iterative methods for the square
root, showing that all these methods belong to the same
family, for which we find a general formula. We then explain how those
iterative methods for real numbers can be extended to spectral
factorization of auto-correlations. The iteration based on
the Newton-Raphson method is optimal from the convergence stand point, though
it is not optimal as far as stability is concerned. Finally, we show
that other members of the iteration family are more stable, though
slightly more expensive and slower to converge.
Helix derivative and low-cut filters' spectral feature and application (ps.gz 7788K) (pdf 20583K) (src 14988K)
Zhao Y.
A helix derivative filter can be used to roughen an image and thus
enhance its details. Unlike the conventional derivative operator,
the helix derivative filter has no direction orientation. I present
the enhanced helix/low-cut derivative filters, in which the zero
frequency response is adjustable. I analyze the quantitative effects
of the adjustable parameters on the filter spectrum and propose
guidelines for choosing parameters. I also show some roughened images
created by the enhanced filters.
Helical meshes on spheres and cones (ps.gz 148K) (pdf 1089K) (src 389K)
Claerbout J.
We embed a helix in the two-dimensional surface of a sphere;
likewise, in the two-dimensional surface of a cone.
This provides a one-dimensional coordinate system
on a two-dimensional surface.
Although mesh points
are exactly evenly spaced along the helix
and approximately evenly spaced in the crossline dimension,
unfortunately,
the angles between neighboring points are continuously changing.
We seem to lose the concepts of two-dimensional autoregression
that we have in cartesian space.
3-D traveltime computation by Huygens wavefront tracing (ps.gz 785K) (pdf 9670K) (src 3988K)
Sava P.
In this paper, I present a 3-D implementation of Huygens wavefront
tracing. The three-dimensional version of the method retains the
characteristics of the two-dimensional one: stability, accuracy,
and efficiency. The major difficulty of the 3-D extension is related
to the handling of triplications. An easy to implement solution is to
approximate the wavefronts at the triplications as planes orthogonal to
the incident ray.
An adaptive finite difference method for traveltime and amplitude (ps.gz 272K) (pdf 2421K) (src 196K)
Qian J. and Symes W. W.
The eikonal equation with point source is difficult to solve with high order accuracy because of the singularity of the solution at the source. All the formally high order schemes turn out to be first order accurate without special treatment of this singularity. Adaptive upwind finite difference methods based on high order ENO (Essentially NonOscillatory) Runge-Kutta difference schemes for the paraxial eikonal equation overcome this difficulty. The method controls error by automatic grid refinement and coarsening based on an a posteriori error estimation. It achieves prescribed accuracy at far lower cost than fixed grid methods. Reliable auxiliary quantities, such as take-off angle and geometrical spreading factor, are by-products.
A second-order fast marching eikonal solver (ps.gz 189K) (pdf 1379K) (src 618K)
Rickett J. and Fomel S.
The fast marching method Sethian (1996) is widely used for solving the
eikonal equation in Cartesian coordinates.
The method's principal advantages are: stability,
computational efficiency, and algorithmic simplicity.
Within geophysics, fast marching traveltime
calculations Popovici and Sethian (1997) may be used
for 3-D depth migration or velocity analysis.
...
Robust and stable velocity analysis using the Huber function (ps.gz 1514K) (pdf 9015K) (src 3949K)
Guitton A. and Symes W. W.
The Huber function is one of several robust error measures which interpolates
between smooth (l2) treatment of small residuals and robust (l1) treatment of
large residuals. Since the Huber function is differentiable, it may be minimized reliably with
a standard gradient-based optimizer. Tests with a linear inverse problem for velocity analysis,
using both synthetic and field data, suggest that (1) the Huber function gives far more robust
model estimates than does least squares, (2) its minimization using a standard quasi-Newton method
is comparable in computational cost to least squares estimation using conjugate gradient
iteration, and (3) the result of Huber data fitting is stable over a wide range of choices for
threshold and total number of quasi-Newton steps.
Extremal regularization (ps.gz 112K) (pdf 1396K) (src 967K)
Symes W. W.
Extremal regularization finds a model fitting the data to a specified
tolerance, and additionally minimizing an auxiliary criterion. It
provides relative model/data space weights when no statistical
information about the model or data is available other than an
estimate of noise level. A version of the Moré-Hebden algorithm
using conjugate gradients to solve the various linear systems
implements extremal regularization for large scale inverse problems.
A deconvolution application illustrates the possibilities and
pitfalls of extremal regularization in the linear case.
Backus revisited: Just in time (ps.gz 71K) (pdf 45K) (src 2K)
Muir F.
The essence of Backus theory Backus (1962) is that it
allows a bunch of layers to be replaced by a single layer. The new
homogeneous medium has elastic properties identical in the long-wavelength
limit so that mass and travel-time are conserved, but wavelet shape is not.
In this paper I show that for normal incidence plane waves the three
elastic layer parameters of thickness, compliance, and density can be replaced
by the two, travel-time and impedance, without losing reflection and
transmission
...
Elastic wave propagation and attenuation in a double-porosity dual-permeability medium (ps.gz 125K) (pdf 1154K) (src 38K)
Berryman J. G. and Wang H. F.
To account for large-volume low-permeability storage porosity and
low-volume high-permeability fracture/crack porosity in oil and gas
reservoirs, phenomenological equations for the poroelastic behavior
of a double
porosity medium have been formulated and the coefficients in these linear
equations identified. This generalization from a single porosity model
increases the number of independent inertial coefficients from three
to six, the number of independent drag coefficients from three to six,
and the number of independent stress-strain coefficients from three
to six for an isotropic applied stress and assumed isotropy of the medium.
The analysis leading to physical interpretations of the inertial
and drag coefficients is relatively straightforward, whereas that
for the stress-strain coefficients is more tedious. In a quasistatic
analysis, the physical interpretations are based upon considerations of
extremes in both spatial and temporal scales. The limit of very short
times is the one most pertinent for wave propagation, and in this case
both matrix porosity and fractures are expected to behave in an undrained
fashion, although our analysis makes no assumptions in this regard.
For the very long times more relevant to reservoir drawdown, the
double porosity medium behaves as an equivalent single porosity medium.
At the macroscopic spatial level, the pertinent parameters (such as the
total compressibility) may be determined by appropriate field tests.
At the mesoscopic scale pertinent parameters of the rock matrix can
be determined directly through laboratory measurements on core, and
the compressiblity can be measured for a single fracture. We show
explicitly how to generalize the quasistatic results to incorporate
wave propagation effects and how effects that are usually attributed
to squirt flow under partially saturated conditions can be explained
alternatively in terms of the double-porosity model. The result
is therefore a theory that generalizes, but is completely consistent
with, Biot's theory of poroelasticity and is valid for analysis
of elastic wave data from highly fractured reservoirs.
A short tour through the Stanford Exploration Project contributions to anisotropy (ps.gz 33K) (pdf 75K) (src 6K)
Alkhalifah T.