At first glance, a passive radar system and a Deep Neural Network (DNN) seem to inhabit different worlds. One deals with wave physics, antennas, and scattering coefficients; the other deals with tensors, activation functions, and gradient descent.
However, if you strip away the domain-specific jargon, you find that the algorithm used to form an image from radar echoes (Back-Projection) and the algorithm used to train a neural network (Back-Propagation) are, at their core, the same mathematical operation.
They are both applications of the Adjoint Operator to solve an Inverse (Reconstruction) Problem.
In this post, we will explore this duality, using the specific example of a passive imaging system to illuminate the inner workings of deep learning.
To understand the backward process, we must first look at the forward process—how data generates an output.
Imagine a passive radar system with multiple receivers (Rx) to detect active targets.
where is the wavelength, and is the distance from the transmitter to the observation point.
where denotes convolution over space .
Following is an interactive demo of Green’s function and the conjugate Green’s function.
Now, look at a single connection in a neural network between an input neuron and an output neuron.
Mathematically, the output is the input modulated by the model weights :
The loss function measures the difference between the predicted output and the true label . Our goal is to minimize this loss by adjusting the weights .
In both fields, we face an inverse problem:
How do we find the target? We take the signal received at the Rx and perform Time Reversal. We mathematically “play the tape backward.”
We project the received signal back into space, applying a negative time delay corresponding to the distance from Rx to the target, and then from the target to Tx.
Now, the imaging problem: we have the received data , and we want to find the target .How do we do it? We mathematically “propagate” the received waves back into space to see where they intersect. This is often called Matched Filtering or Back-projection. Mathematically, we convolve the received signal with the complex conjugate of the Green’s function, :
The physical meaning of is that we are reversing the wave propagation process. By convolving with the conjugate Green’s function, we assume the sensor behaves like a transmitter in reverse, sending a signal back through the same medium.
It is notable that this is an ideal model; in practice we can not perfectly capture the whole received signal in space, and noise will be present. But the core idea remains: by back-projecting the received signals using the conjugate of the forward operator, we can reconstruct the target location.
How do we find the optimal weight? We can use Gradient Descent. However, calculating the gradient of the loss function with respect to each weight directly is computationally expensive. Instead, we use the Back-Propagation algorithm.
Consider a simple neural network with layers:
Then, the loss function is . And we want to compute the gradient of the loss which contains and for all layers .
BP introduces intermediate variables called Error Signals , defined as:
Error signal at the output layer :
Applying the chain rule, we can express the error signal at layer in terms of the error signal at layer :
Because the activation function is applied element-wise, we have:
where if and otherwise. Therefore, we can simplify the expression for the error signal:
To simplify notation, we can express this in vector form:
where denotes the element-wise (Hadamard) product.
Error signal at hidden layers :
Again, applying the chain rule, we have:
According to the definition of , we can rewrite the above equation as:
can be computed using the Jacobian matrix:
Thus, we can express the error signal at layer as:
Since and , we have:
Together with (15) and (10), we can recursively compute the error signals from the output layer back to the input layer. This is the essence of the Back-Propagation algorithm.
After computing the error signals for all layers, we can compute the gradients with respect to the weights and biases:
because , we have . Therefore:
Vectorizing this, we get:
Similarly, for the weights:
Since , we have . Therefore:
In matrix form, this becomes:
Following is an interactive demo of Back-Propagation in Deep Learning. The task is to fit a simple 2D function:
The striking similarity between the equations of wave imaging and neural network training is not a coincidence. Both systems are governed by the same mathematical principle: the duality between a linear operator and its adjoint.
To see this clearly, let us unify our notation. In both cases, we have a Forward Operator that maps an internal state (the target or the weights ) to an observation (the signal or the prediction ).
In a linearized or local sense, both processes can be viewed as an operator acting on a hidden source to produce a measurable result:
| Feature | Passive Radar (Imaging) | Deep Learning (Training) |
|---|---|---|
| Hidden Source | Target Reflectivity | Parameter Update |
| Forward Operator | Green’s Function Convolution | Jacobian Matrix |
| Observation | Received Echoes | Prediction Error |
Mathematically, we represent the forward pass as:
where is our forward transformation.
The “Inverse Problem” asks: given the observation , what was the source ?
Solving this perfectly usually requires computing the inverse , which is often computationally impossible or “ill-posed” (highly sensitive to noise). Instead, both fields use the Adjoint Operator (the complex conjugate transpose in matrix terms).
The reconstructed image is formed by applying the adjoint of the propagation operator:
As derived earlier, corresponds to convolution with the conjugate Green’s function . Physically, this is Time Reversal.
The gradient used to update weights is found by applying the adjoint of the network’s local sensitivity:
Here, the transpose is the adjoint of the forward weight matrix . Just as the radar signal is sent back through the channel, the error signal is “sent back” through the weights.
The effectiveness of both algorithms stems from the Adjoint Identity:
In imaging, this means the energy we collect at the receivers () is mathematically equivalent to the energy at the target location when we project our “knowledge” back into the scene.
In deep learning, this identity ensures that the most efficient way to reduce the loss in the high-dimensional output space is to “project” that error back onto the weights.
The “Back” in both Back-Projection and Back-Propagation refers to the same fundamental movement: reversing the flow of information.
Whether we are “seeing” a stealth aircraft through passive echoes or “learning” to recognize a cat in a photo, we are performing the same mathematical ritual: calculating the forward transformation, measuring the discrepancy, and using the Adjoint Operator to project that discrepancy back to its origin.
Deep learning, in this light, is not just a branch of computer science; it is a form of Computational Imaging where the “scene” being reconstructed is the internal logic of the model itself.
Deep Learning, Imaging — Dec. 19, 2025
Unless otherwise stated, all content on this blog is licensed under CC BY-NC-SA 4.0 |
RSS subscribe |
Friends
鄂ICP备2025091178号 |
对应的中文页面尚未上线