THIS PACKAGE IS DEPRECATED WITH CHAINRULES DIRECTLY INTEGRATED INTO JUDI>=V3.1.0
JUDI4Flux: Seismic modeling for deep learning
JUDI4Flux enables compositions of seismic modeling operators with (convolutional) neural networks. JUDI4Flux is an extension of JUDI, a framework for seismic modeling and inversion with automatic code generation and performance optimization based on Devito. JUDI4Flux integrates JUDI's linear and nonlinear modeling operators into the Flux deep learning library, thus allowing the implementation of physicsdriven neural networks. For backpropagation, JUDI4Flux calls JUDI's adjoint PDE solvers, thus making it possible to backpropagate efficiently through single or multiple PDE layers and scale to large problem sizes.
Features:

Compatibility with the Julia Flux deep learning library. Both Flux and JUDI are based on abstract highlevel mathematical expressions that enable clean coding.

Blazingly fast seismic modeling routines using stencilbased finitedifference C code, which is automatically generated and optimized by Devito.

Supported operators: forward/adjoint modeling, linearized Born scattering/RTM
Installation
User installation:
] add https://github.com/slimgroup/JUDI4Flux.jl
Developer installation:
] dev https://github.com/slimgroup/JUDI4Flux.jl
Linear and nonlinear JUDI operators with Flux
JUDI4Flux enables compositions of neural network layers (e.g. convolutional or fullyconnected layers) with operators for seismic modeling. Instead of having to reimplement seismic modeling operators with convolutions from machine learning libraries, this makes it possible to use existing modeling operators, namely JUDI operators for Born and nonlinear modeling. Even more importantly, we can evaluate these operators during backpropagation by calling the corresponding adjoint operators, but fully integrate them into Flux's automatic differentiation (AD) module (Flux.Tracker).
This allows combining JUDI operators with Flux misfit functions to compute gradients for least squares RTM (LSRTM) or full waveform inversion (FWI) and combinations of modeling and neural network layers. For example, the following code show an example of combining the Born scattering operator with a dense neural network layer (check here for the full example):
# Test demigration operator w/ Flux Dense Layer
y = randn(Float32, 100)
W = randn(Float32, 100, length(y))
b = randn(Float32, 100)
# Linearized Born operator
J = judiJacobian(F, q)
# Example image
x = vec(image)
predict(x) = W*(J*x) .+ b
loss(x,y) = Flux.mse(predict(x), y)
# Compute gradient w/ Flux
gs = gradient(() > loss(x, y), params(W, b, x))
gs[x] # evalute gradient of x
Using nonlinear modeling operators and convolutional layers is possible as well! For example, the following code shows how to integrate an extended source modeling JUDI operator into a shallow CNN, consisting of two convolutional layers, with a nonlinear forward modeling layer ℱ inbetween them. As before, we can define a loss function using Flux utilities and compute derivatives with respect to various parameters, such as the squared slowness vector m
. Once again, gradients of layers containing JUDI operators are computed using the corresponding adjoints or JUDI gradients, instead of Flux's Tracker module (full example here):
# Extended source modeling operator
model = Model(n, d, o, m)
Pr = judiProjection(info, recGeometry)
A_inv = judiModeling(info, model; options=opt)
Pw = judiLRWF(info, wavelet)
# Network layers
ℱ = ExtendedQForward(Pr*A_inv*adjoint(Pw))
conv1 = Conv((3, 3), 1=>1, pad=1, stride=1)
conv2 = Conv((3, 3), 1=>1, pad=1, stride=1)
# Network and loss
function predict(x, m)
x = conv1(x)
x = ℱ(x, m) # m is the velocity model
x = conv2(x)
return x
end
loss(x, m, y) = Flux.mse(predict(x, m), y)
# Compute gradient w/ Flux
p = params(x, m, conv1, conv2)
gs = gradient(() > loss(x, m, y), p)
gs[m] # evalute gradient w.r.t. m
gs[conv1.weight] # evalute gradient w.r.t. the convolution kernel
Example applications
A possible application of JUDI4Flux is the implementation of loopunrolled LSRTM algorithms  physicsaugmented convolutional neural networks for seismic imaging. By training a loop unrolled LSRTM network using pairs of true images and observed data, this makes it possible to obtain highfidelity images from noisy simultaneous shot records. The below figure compares RTM, standard LSRTM with gradient descent and loop unrolled LSRTM. Each image is obtained from a single simultaneous shot record only. (Full preprint to be added to arXiv shortly.)
Related work

For a similar framework in Python that interfaces PyTorch (and actually predates
JUDI4Flux
), please check out Alan Richardson very cool package deepwave. Similar toJUDI4Flux
, this package integrates seismic modeling and linearized modeling functions based on finitedifference stencil code into a deep learning framework. Alan's package supports seismic modeling on both CPUs and GPUs. 
DiffEqFlux.jl is a Julia package that integrates the DifferentialEquations.jl package for more generic ODEs/PDEs into Flux.
Author
This package was written by Philipp Witte from the Seismic Laboratory for Imaging and Modeling (SLIM) at the Georgia Institute of Technology.
 Contact Philipp at pwitte3@gatech.edu