==================== Linear Scale Space ==================== .. sidebar:: **Jan Koenderink** .. image:: /images/koenderink_scsp.png :height: 200 Prof. Jan Koenderink from Utrecht University formulated scale-space theory in the 1980's. The seminal paper is: J.J. Koenderink, *The Structure of Images*, Biological Cybernetics, 1984. From the abstract: "*In practice the relevant details of images exist only over a restricted range of scale. Hence it is important to study the dependence of image structure on the level of resolution. It seems clear enough that visual perception treats images on several levels of resolution.*" It may seem that there are enumerous ways to construct a scale-space. But surprisingly in case we set some simple requirements (basic principles) there is essentially only one scale-space construction possible. .. proof:definition:: (Scale-Space Basic Principles) We are looking for a family of image operators $\op L^s$ where $s$ refers to the scale such that given the 'image at zero scale' $f^0$ we can construct the family of images $\op L^s f^0$ such that: #. $\op L^s$ is a **linear** operator, #. $\op L^s$ is **translation invariant** (all parts of the image are treated equally, i.e. there is no a priori preferred location), #. $\op L^s$ is **rotational invariant** (all directions are treated equally, i.e. there is no a priori preferred direction), #. $\op L^s$ is **separable by dimension**, and #. $\op L^s$ is **scale invariant** (starting with a scaled version of the zero scale image we should in principle obtain the same scale-space with possible reparameterization of the scale). It can be shown that the scale space operator satisfying all the requirements is the Gaussian convolution. .. proof:theorem:: Linear Gaussian Scale-Space The scale-space operator that satisfies all basic principles is the Gaussian convolution: .. math:: f(\v x, s) = (f^0 \ast G^s)(\v x) where $G^s$ is the 2D Gaussian kernel: .. math:: G^s(\v x) = \frac{1}{2\pi s^2} \exp\left(-\frac{\|\v x\|^2}{2 s^2}\right) The function $f$ can be interpreted as a one parameter family of images. For every scale $s>0$ there is an image $f(\cdot,s)$ that we will frequently refer to as $f^s$. Some properties of the Gaussian convolution (and hence of the linear Gaussian scale space construction) are already discussed in a previous chapter (see :numref:`gaussian_derivative_convolutions`). For ease of reference we will repeat some of them here and state some more properties important in the scale-space context. .. proof:theorem:: Linear Diffusion Equation The scale-space function $f$ is the solution of the diffusion equation: .. math:: f_s = s \, \nabla^2 f = s (f_{xx}+f_{yy}) with initial condition $f(\cdot,0)=f^0$. In fact this differential equation was the starting point for Koenderink in deriving linear scale space (see :numref:`linear_diffusion`). From a practical point of view it is an important equation because it tells us that the change in the function $f$ due to a small change in scale (going from scale $s$ to scale $s+ds$) is proportional to the Laplacian differential structure of $f$ at scale $s$. It shows that in a Gaussian scale-space scale derivatives may always be exchanged for spatial derivatives. Because taking a derivative is a linear translation invariant operator that commutes with the Gaussian convolution we have that any spatial derivative $\partial_\ldots f$ adheres to the same scale space equation: .. proof:theorem:: Scale Space Derivatives We can either first calculate a derivative and from that construct the scale-space or take the derivative at every scale in the scale-space: .. math:: \partial_\ldots f^0 \ast G^s = \partial_\ldots \left( f^0 \ast G^s \right) where $\partial_\ldots$ stands for any spatial derivnative. We can equivalently calculate the Gaussian derivative at all scales: .. math:: \partial_\ldots f^0 \ast G^s = f^0 \ast \partial_\ldots G^s This theorem is a specific use of a more general theorem :numref:`derivatives_of_convolution`. .. exec_python:: derivatives scalespace :linenumbers: :code: shutter :code_label: Show code for figure :results: hide :file: python/lecturenotes/normalized_derivatives.py fig_derivatives.savefig('source/images/derivatives_step.png') .. figure:: /images/derivatives_step.png :width: 80% :align: center :name: fig-derivatives **(Normalized) Gaussian Derivatives of Step Function.** On the first row the first order (left) and second order (right) derivative of a step function. On the second row the normalized first and second order derivatives. In :numref:`fig-derivatives` we show a step function and its first and second order derivatives calculated at several scales in the top row. It is evident that the derivatives decrease in amplitude with increasing scale. When comparing derivatives accross scale this is problematic. To solve this problem we will use **scale normalized derivatives**. .. proof:definition:: Scale Normalized Derivatives Let $\partial_\ldots^k$ denote any $k$-th order spatial derivative than the scale normalized Gaussian spatial derivative is given as: .. math:: s^k (f^0 \ast \partial_\ldots^k G^s) The scale normalized derivatives for the step edge are shown in in the bottom row of :numref:`fig-derivatives`. .. exec_python:: flower_gradient scalespace :linenumbers: :code: shutter :code_label: Show code for figure :results: hide # see code for a previous figure fig_flower.savefig('source/images/flower_gradient.png') .. figure:: /images/flower_gradient.png :width: 100% :align: center :name: fig-flower_derivatives In :numref:`fig-flower_derivatives` the standard derivatives are compared with the normalized derivatives. On the first row first the original image followed by $f^s_w$ (the gradient norm) for several values of the scale $s$. On the second row again the original image but now with the normalized gradient $s f^s_w$. .. exec_python:: flower_gradient_vector scalespace :linenumbers: :code: shutter :code_label: Show code for figure :results: hide # see code for a previous figure fig_grad_flower.savefig('source/images/fig_flower_gradient_vector.png') .. figure:: /images/fig_flower_gradient_vector.png :figwidth: 50% :align: right :name: fig-gradient_vector_flower **Gradient Vector dependent on scale.** The images of the flower show that the gradient calculated in one point is dependent on the scale of the Gaussian convolution. In :numref:`fig-gradient_vector_flower` we see a small detail of the flower. In the left column the scale is $s=1$, where in the top row we see the image $f^1$ and in the bottom row $f^1_w$. Also shown is the gradient vector $\nabla f^1$ in one point in the image. In the right column the scale is $s=4$. Observe the change in direction of the gradient vector when compared with the gradient at scale $s=1$. When constructing a scale-space we can calculate each scale by convolving the zero scale image. But given the semi group property of the Gaussian convolution (see :numref:`semigroup_gaussian_derivatives`) the scale space can be constructed incrementally. For ease of reference we repeat that theorem here: .. proof:theorem:: **Semi Group of Gaussian Convolutions** If we take a Gaussian function $G^s$ and convolve it with another Gaussian function $G^t$ the result is a third Gaussian function: .. math:: G^s \ast G^t = G^{\sqrt{s^2+t^2}} So given the image $f^s$ the image at scale $t$ (with $t>s$ can be calculated as: .. math:: f^t = f^0 \ast G^t = f^s \ast G^{\sqrt{t^2-s^2}} The computational complexity of the Gaussian convolution $f\ast G^s$ is proportional to the scale $s$. So the calculation is $\bigO(t)$. The computational complexity of the calculation $(f\ast G^s)\ast G^\sqrt{t^2-s^2}$ is $\bigO(s+\sqrt{t^2-s^2})$. It is not too difficult to prove that in general $\bigO(s+\sqrt{t^2-s^2}) < \bigO(t)$. However, although incrementally building a scale space is more efficient it is not always to be preferred because Gaussian convolutions at small scales are always less accurate than convolutions at larger scale. In :numref:`subsampling_linear_scale_space` we will look at constructing scale-space from a discrete, i.e. sampled image.