next up previous contents
Next: Noise distributions Up: Basic Concepts Previous: Basic Concepts

Image sampling

The acquisition of a data frame involves a spatial sampling and digitalization of the continuous image formed in the focus plane of a telescope. The image may be recorded analog (e.g. on photographic plates) for later measurements or acquired directly when digital detectors such as diode arrays and CCD's are used. The individual pixel values are obtained by convolving the continuous image with the pixel response function . With a sampling step of and the digital frame is given by

 

where N is the acquisition noise. This convolution is done analog in most detectors except for imaging photon counting systems where it partly is performed digitally. The sampling step and response function are determined normally by the physical properties of the detector and the acquisition setup. The variation of the response function may be very sharp as for most semi--conductor detectors or more smooth as in image dissector tubes. If the original image I is band width limited (i.e. only contains features with spatial frequencies less than a cutoff value ) all information is retained in the digitized frame when the sampling frequency satisfies the Nyquist criterion:

 

In Equation gif it is assumed that R is a Dirac delta function. This means that only features which are larger than can be resolved. A frame is oversampled when while for smaller sample rates it is undersampled.

In astronomy the band width of an image is determined by the point spread function (PSF) and has often no sharp cutoff frequency. Many modern detector systems are designed to have a sampling step only a few times smaller than the typical full width half maximum (FWHM) of seeing disk or PSF. Therefore they will not fully satisfy Equation gif and tend to be undersampled especially in good seeing conditions.

A typical assumption in image processing algorithms is that the pixel response function R can be approximated by a Dirac delta function. This is reasonable when the image intensity does not vary significantly over R as for well oversampled frames where the effective size of R is roughly equal to the sample step. If it is not the case, the effects on the algorithm used should be checked. Interpolation of values between existing pixels is often necessary e.g. for rebinning. Depending on the shape of R and band width of the image different schemes may be chosen to give the best reproduction of the original intensity distribution. In many cases low order polynomial functions are used (e.g. zero or first order) while sinc, spline or gaussian weighted interpolation may be more appropriate for some applications.



next up previous contents
Next: Noise distributions Up: Basic Concepts Previous: Basic Concepts



Pascal Ballester
Tue Mar 28 16:52:29 MET DST 1995