Image Discretization ==================== To store an image function $f: \setR^d \rightarrow \setR$ in computer memory we need to make a discrete representation of it. Image discretization involves two separate processes: discretization of the spatial domain (**sampling**) and discretization of the image range (**quantization**). Quantization ------------ The classical grey value image is a mapping from the spatial domain $\setR^2$ to the set of luminance values $\setR$. These luminance values are always positive (there is no such thing as a negative energy that can be observed). Luminances are also restricted to a finite range. There is an energy level that completely saturates the image sensor, such that no larger luminance levels can be observed. Traditionally the luminance range is encoded in 256 levels from 0 to 255, corresponding with an 8 bit discretization of the luminance range. Nowadays grey value luminance ranges are encoded in some sensors in up to 12 bits (especially for medical X-ray images). Color images are most often represented using an 8 bit encoding of the red, green and blue channel each (making up a 24 bit per pixel encoding of the image). We will always assume that the image is properly quantized, meaning that it does not only allow for a faithful representation of the observation but that it also facilitates the necessary image computations. In practice this is dependent on the both the images (CT images have a luminance range that needs more than 10 bits for proper discretization whereas your cheap web-cam only needs about 6 bits (for each of the 3 color channels)) as well as on the type of computations needed (calculating the gradient, based on first order derivatives, requires negative numbers). Word of advice: when developing image processing algorithms start with images quantized with floating point values. Only when speed and or memory is important look for other (integer valued) representations. You would not be the first to be bewildered [#fn]_ by the result of an image expression like .. code:: python g = (f - f.min()) / (f.max()-f.min()) * 255 in case ``f`` is an image of type ``uint8``. Sampling -------- In a previous section on image formation we have seen that an image can only be observed in a finite number of, sample, points. The set of sample points is called the **sampling grid**. In these notes we assume a simple rectangular sampling grid restricted to a rectangular subset of the plane. In some situations we need to distinguish between image functions defined on a continuous domain (i.e. $\setR2$) and discrete, i.e. sampled images. In such situations we will often use lower case letters for the images with a continuous domain and uppercase letters for sampled images. The image $f:\setR^2\rightarrow\setR$ is sampled to obtain the function $F$ defined on a discrete domain $F:\setZ^2\rightarrow\setR$: .. math:: F(i,j) = f(i\Delta x, j\Delta y) where $\Delta x$ and $\Delta y$ are the sampling distances along the $x$ and $y$ axis respectively. The mathematical notion of sampling is a theoretical abstraction. It assumes an image function can be measured in a point (with no area). In practice a sample is always some kind of average over a small area (most often a rectangular part of the CCD or CMOS chip in case of optical or X-ray imaging). A **pixel** stands for a *picture element*. .. ipython:: python :suppress: import numpy as np import matplotlib.pylab as plt f = np.arange(12).reshape(3, 4) k, l = np.meshgrid(np.arange(4), np.arange(3)) plt.imshow(f, interpolation='nearest', cmap='gray'); #plt.axis('off'); plt.scatter(k, l, s=100, c='g'); @savefig pixels_samples.png plt.show(); In the figure above a rectangular image of 12 pixels in total are shown. The sample positions (pixel positions) are indicated as green points overlayed on the standard image display). Note that the sample positions have integer value coordinates. But also note that on the screen the pixel value $F(i,j)$ is used to fill the square $i-0.5