Top Banner
12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy Resolution and Contrast in Confocal … 1/9 www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html Olympus America | Research | Imaging Software | Confocal | Clinical | FAQ’s Home Page Interactive Tutorials Microscopy Primer Physics of Light & Color Microscopy Basic Concepts Special Techniques Fluorescence Microscopy Confocal Microscopy Confocal Applications Digital Imaging Digital Image Galleries Digital Video Galleries Virtual Microscopy Microscopy Web Resources Olympus Product Brochures MICD Microscope Archive Photomicrography Archive Software Downloads Resolution and Contrast in Confocal Microscopy All optical microscopes, including conventional widefield, confocal, and twophoton instruments are limited in the resolution that they can achieve by a series of fundamental physical factors. In a perfect optical system, resolution is restricted by the numerical aperture of optical components and by the wavelength of light, both incident (excitation) and detected (emission). The concept of resolution is inseparable from contrast, and is defined as the minimum separation between two points that results in a certain level of contrast between them. In a typical fluorescence microscope, contrast is determined by the number of photons collected from the specimen, the dynamic range of the signal, optical aberrations of the imaging system, and the number of picture elements (pixels) per unit area in the final image. The influence of noise on the image of two closely spaced small objects is further interconnected with the related factors mentioned above, and can readily affect the quality of resulting images. While the effects of many instrumental and experimental variables on image contrast, and consequently on resolution, are familiar and rather obvious, the limitation on effective resolution resulting from the division of the image into a finite number of picture elements (pixels) may be unfamiliar to those new to digital microscopy. Because all digital confocal images employing laser scanners and/or camera systems are recorded and processed in terms of measurements made within discrete pixels, some discussion of the concepts of sampling theory is required. This is appropriate to the subject of contrast and resolution because it has a direct bearing on the ability to record two closely spaced objects as being distinct. In addition to the straightforward theoretical aspects of resolution, regardless of how it is defined, the reciprocal relationship between contrast and resolution has practical significance because the matter of interest to most microscopists is not resolution, but visibility. The ability to recognize two closely spaced features as being separate relies on advanced functions of the human visual system to interpret intensity patterns, and is a much more subjective concept than the calculation of resolution values based on diffraction theory. Experimental limitations and the properties of the specimen itself, which vary widely, dictate that imaging cannot be performed at the theoretical maximum resolution of the microscope. The relationship between contrast and resolution with regard to the ability to distinguish two closely spaced specimen features implies that resolution cannot be defined without reference to contrast, and it is this interdependence that has led to considerable ambiguity involving the term resolution and the factors that influence it in microscopy. Recent advances in fluorescent protein technology have led to an enormous increase in studies of dynamic processes in living cells and tissues. Such specimens are optically thick and inhomogeneous, resulting in a farfromideal imaging situation in the microscope. Other factors, such as cell viability and sensitivity to thermal damage and photobleaching, place limits on the light intensity and duration of exposure, consequently limiting the attainable resolution. Given that the available timescale may be dictated by these factors and by the necessity to record rapid dynamic events in living cells, it must be accepted that the quality of images will not be as high as those obtained from fixed and stained specimens. The most reasonable resolution goal for imaging in a given experimental situation is that the microscope provide the best resolution possible within the constraints imposed by the experiment. The Airy Disk and Lateral Resolution Imaging a pointlike light source in the microscope produces an electromagnetic field in the image plane
9

Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

Jan 11, 2020

Download

Documents

dariahiddleston
Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

1/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

Olympus America | Research | Imaging Software | Confocal | Clinical | FAQ’s

Home Page

Interactive Tutorials

Microscopy Primer

Physics of Light & Color

Microscopy Basic Concepts

Special Techniques

Fluorescence Microscopy

Confocal Microscopy

Confocal Applications

Digital Imaging

Digital Image Galleries

Digital Video Galleries

Virtual Microscopy

Microscopy Web Resources

Olympus Product Brochures

MIC-­D Microscope Archive

Photomicrography Archive

Software Downloads

Resolution and Contrast in Confocal Microscopy

All optical microscopes, including conventional widefield, confocal, and two-­photon instruments are limitedin the resolution that they can achieve by a series of fundamental physical factors. In a perfect opticalsystem, resolution is restricted by the numerical aperture of optical components and by the wavelength oflight, both incident (excitation) and detected (emission). The concept of resolution is inseparable fromcontrast, and is defined as the minimum separation between two points that results in a certain level ofcontrast between them. In a typical fluorescence microscope, contrast is determined by the number ofphotons collected from the specimen, the dynamic range of the signal, optical aberrations of the imagingsystem, and the number of picture elements (pixels) per unit area in the final image.

The influence of noise on the image of two closely spaced small objects is further interconnected with therelated factors mentioned above, and can readily affect the quality of resulting images. While the effectsof many instrumental and experimental variables on image contrast, and consequently on resolution, arefamiliar and rather obvious, the limitation on effective resolution resulting from the division of the imageinto a finite number of picture elements (pixels) may be unfamiliar to those new to digital microscopy.Because all digital confocal images employing laser scanners and/or camera systems are recorded andprocessed in terms of measurements made within discrete pixels, some discussion of the concepts ofsampling theory is required. This is appropriate to the subject of contrast and resolution because it has adirect bearing on the ability to record two closely spaced objects as being distinct.

In addition to the straightforward theoretical aspects of resolution, regardless of how it is defined, thereciprocal relationship between contrast and resolution has practical significance because the matter ofinterest to most microscopists is not resolution, but visibility. The ability to recognize two closely spacedfeatures as being separate relies on advanced functions of the human visual system to interpret intensitypatterns, and is a much more subjective concept than the calculation of resolution values based ondiffraction theory. Experimental limitations and the properties of the specimen itself, which vary widely,dictate that imaging cannot be performed at the theoretical maximum resolution of the microscope.

The relationship between contrast and resolution with regard to the ability to distinguish two closelyspaced specimen features implies that resolution cannot be defined without reference to contrast, and it isthis interdependence that has led to considerable ambiguity involving the term resolution and the factorsthat influence it in microscopy. Recent advances in fluorescent protein technology have led to anenormous increase in studies of dynamic processes in living cells and tissues. Such specimens areoptically thick and inhomogeneous, resulting in a far-­from-­ideal imaging situation in the microscope. Otherfactors, such as cell viability and sensitivity to thermal damage and photobleaching, place limits on thelight intensity and duration of exposure, consequently limiting the attainable resolution. Given that theavailable timescale may be dictated by these factors and by the necessity to record rapid dynamic eventsin living cells, it must be accepted that the quality of images will not be as high as those obtained fromfixed and stained specimens. The most reasonable resolution goal for imaging in a given experimentalsituation is that the microscope provide the best resolution possible within the constraints imposed by theexperiment.

The Airy Disk and Lateral Resolution

Imaging a point-­like light source in the microscope produces an electromagnetic field in the image plane

Page 2: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

2/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

whose amplitude fluctuations can be regarded as a manifestation of the response of the optical system tothe specimen. This field is commonly represented through the amplitude point spread function, andallows evaluation of the optical transfer properties of the combined system components. Althoughvariations in field amplitude are not directly observable, the visible image of the point source formed in themicroscope and recorded by its imaging system is the intensity point spread function, which describesthe system response in real space. Actual specimens are not point sources, but can be regarded as asuperposition of an infinite number of objects having dimensions below the resolution of the system. Theproperties of the intensity point spread function (PSF;; see Figure 1) in the image plane as well as in theaxial direction (Figure 3) are major factors in determining the resolution of a microscope.

It is possible to experimentally measure the intensity point spread function in the microscope by recordingthe image of a subresolution spherical bead as it is scanned through focus (a number of examples maybe found in the literature). Because of the technical difficulty posed in direct measurement of the intensitypoint spread function, calculated point spread functions are commonly utilized to evaluate the resolutionperformance of different optical systems, as well as the optical-­sectioning capabilities of confocal, two-­photon, and conventional widefield microscopes. Although the intensity point spread function extends inall three dimensions, with regard to the relationship between resolution and contrast, it is useful toconsider only the lateral components of the intensity distribution, with reference to the familiar Airy disk.

The intensity distribution of the point spread function in the plane of focus is described by the rotationallysymmetric Airy pattern. Because of the cylindrical symmetry of the microscope lenses, the two lateralcomponents (x and y) of the Airy pattern are equivalent, and the pattern represents the lateral intensitydistribution as a function of distance from the optical axis. The lateral distance is normalized by thenumerical aperture of the system and the wavelength of light, and therefore is dimensionless. Figure 1(airy disk and intensity function) illustrates diagrammatically the formation and characteristics of the Airydisk, the related three-­dimensional point spread function, and Airy patterns in the fluorescencemicroscope. Following the excitation of fluorophores in a point-­like specimen region, fluorescenceemission occurs in all directions, a small fraction of which is selected and focused by the opticalcomponents into an image plane where it forms an Airy disk surrounded by concentric rings ofsuccessively decreasing maximum and minimum intensity (the Airy pattern).

The Airy pattern intensity distribution is the result of Fraunhofer diffraction of light passing through acircular aperture, and in a perfect optical system exhibits a central intensity maximum and higher ordermaxima separated by regions of zero intensity. The distance of the zero crossings from the optical axis(represented by the variable ν in Figure 1), when the distance is normalized by the numerical apertureand wavelength, occur at values for ν of 3.8, 7.0, 10.2, 13.3....etc. When the intensity on the optical axis isnormalized to one (100 percent), the proportional heights of the first four higher order maxima are 1.7,0.4, 0.2, and 0.08 percent, respectively.

Resolution and Contrast

A useful approach to the concept of resolution is based on consideration of an image formed by two point-­like objects (specimen features), under the assumption that the image-­forming process is incoherent, andthat the interaction of the separate object images can be described using intensity point spread functions.The resulting image is then composed of the sum of two Airy disks, the characteristics of which dependupon the separation distance between the two points. Figure 2 presents a representation of the lateralinteraction between the intensity point spread functions (which are the Airy functions) for two point objectsat various separation distances. When the two objects are widely separated, their images are easilydistinguished on the basis of the variation in image intensity.

As is apparent in Figure 2, when sufficiently separated, the intensity change in the area between theobjects is the maximum possible, cycling from the peak intensity (at the first point) to zero and returning tothe maximum value at the center of the second point. At decreased distance in object space, the intensitydistribution functions of the two points, in the image plane, begin to overlap and the resulting image mayappear to be that of a single larger or brighter object or feature rather than being recognizable as twoobjects. If resolution is defined, in general terms, as the minimum separation distance at which the twoobjects can be sufficiently distinguished, it is obvious that this property is related to the width of theintensity peaks (the point spread function). Microscope resolution is directly related, therefore, to the fullwidth at half maximum (FWHM) of the instrument's intensity point spread function in the component

Page 3: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

3/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

directions.

Some ambiguity in use of the term resolution results from the variability in defining the degree ofseparation between features and their point spread functions that is "sufficient" to allow them to be

distinguished as two objects rather than one. In general, minute features of interest in microscopy

specimens produce point images that overlap to some extent, displaying two peaks separated by a gap.

The greater the depth of the gap between the peaks, the easier it is to distinguish, or resolve, the two

objects. By specifying the depth of the dip in intensity between two overlapping point spread functions, the

ambiguity in evaluating resolution can be removed, and a quantitative aspect introduced.

In order to quantify resolution, the concept of contrast is employed, which is defined for two objects ofequal intensity as the difference between their maximum intensity and the minimum intensity occurring in

the space between them. Because the maximum intensity of the Airy disk is normalized to one, the

highest achievable contrast is also one, and occurs only when the spacing between the two objects is

relatively large, with sufficient separation to allow the first zero crossing to occur in their combined

intensity distribution. At decreased distance, as the two point spread functions begin to overlap, the dip in

intensity between the two maxima (and the contrast) is increasingly reduced. The distance at which two

peak maxima are no longer discernible, and the contrast becomes zero, is referred to as the contrastcut-­off distance. The variation of contrast with distance allows resolution, in terms of the separation oftwo points, to be defined as a function of contrast.

The relationship between contrast and separation distance for two point-­like objects is referred to as the

contrast/distance function or contrast transfer function. Resolution can be defined as the separationdistance at which two objects are imaged with a certain contrast value. It is obvious that when zero

contrast exists (Figure 2(b)), the points are not resolved;; the so-­called Sparrow criterion defines theresolution of an optical system as being equivalent to the contrast cut-­off distance. It is common, however,

to specify that greater contrast is necessary to adequately distinguish two closely spaced points visually,

and the well-­known Rayleigh criterion for resolution states that two points are resolved when the firstminimum (zero crossing) of one Airy disk is aligned with the central maximum of the second Airy disk.

Under optimum imaging conditions, the Rayleigh criterion separation distance corresponds to a contrast

value of 26.4 percent. Although any contrast value greater than zero can be specified in defining

resolution, the 26-­percent contrast of the Rayleigh criterion is considered reasonable in typical

fluorescence microscopy applications, and is the basis for the common expression defining lateral

resolution according to the following equation, in which the point separation (r) in the image plane is thedistance between the central maximum and the first minimum in the Airy disk:

rlateral = 1.22 λ / (2 • NA) = 0.6 λ / NA

where λ is the emitted light wavelength and NA is the numerical aperture of the objective.

Resolution in the microscope is directly related to the FWHM dimensions of the microscope's point spread

function, and it is common to measure this value experimentally in order to avoid the difficulty in

attempting to identify intensity maxima in the Airy disk. Measurements of resolution utilizing the FWHM

values of the point spread function are somewhat smaller than those calculated employing the Rayleigh

criterion. Furthermore, in confocal fluorescence configurations, pointwise illumination scanning and

pointwise detection are employed, so that only the fluorophores in the shared volume of the illumination

and detection point spread functions are able to be detected. The intensity point spread function in the

confocal case is, therefore, the product of the independent illumination intensity and detection intensity

point spread functions. For confocal fluorescence, the lateral (and axial) extent of the point spread

function is reduced by about 30 percent compared to that in the widefield microscope. Because of the

narrower intensity point spread function, the separation of points required to produce acceptable contrast

in the confocal microscope is reduced to a distance approximated by:

rlateral = 0.4 λ / NA

If the illumination and fluorescence emission wavelengths are approximately the same, the confocal

fluorescence microscope Airy disk size is the square of the widefield microscope Airy disk. Consequently,

the contrast cut-­off distance is reduced in the confocal arrangement, and equivalent contrast can be

achieved at a shorter distance compared to the widefield illumination configuration. Regardless of the

instrument configuration, the lateral resolution displays a proportional relationship to wavelength, and is

inversely proportional to the objective lens numerical aperture.

As noted previously, lateral resolution is of primary interest in discussing resolution and contrast, although

Page 4: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

4/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

the axial extent of the microscope intensity point spread function is similarly reduced in the confocal

arrangement as compared to the widefield fluorescence configuration. Reasonable contrast between

point-­like objects lying on the optical axis occurs when they are separated by the distance between the

central maximum and the first minimum of the axial point spread function component. Presented in Figure

3 are the axial intensity distributions for a typical widefield (Figure 3(a)) and confocal (Figure 3(b))

fluorescence microscope. Note the dramatic reduction in intensity of the "wings" in the confocal

distribution as a function of distance from the central maximum.

A variety of equations are presented in the literature that pertain to different models for calculating axial

resolution for various microscope configurations. The ones most applicable to fluorescence emission are

similar in form to the expressions evaluating depth of field, and demonstrate that axial resolution is

proportional to the wavelength and refractive index of the specimen medium, and inversely proportional to

the square of the numerical aperture. Consequently, the numerical aperture of the microscope objective

has a much greater effect on axial resolution than does the emission wavelength. One equation

commonly used to describe axial resolution for the confocal configuration is given below, with ηrepresenting the index of refraction, and the other variables as specified previously:

raxial = 1.4 λ • η / NA2

Although the confocal microscope configuration exhibits only a modest improvement in measured axial

resolution over that of the widefield microscope, the true advantage of the confocal approach is in the

optical sectioning capability in thick specimens, which results in a dramatic improvement in effective axial

resolution over conventional techniques. The optical sectioning properties of the confocal microscope

result from the characteristics of the integrated intensity point spread function, which has a maximum in

the focal plane when evaluated as a function of depth. The equivalent integral of intensity point spread

function for the conventional widefield microscope is constant as a function of depth, producing no optical

sectioning capabilities.

Practical Factors Influencing Contrast

The importance of confocal microscopy in current biological and biomedical research results largely from

the optical sectioning capability that enables the three-­dimensional analysis of thick specimens such as

those encountered in studies involving living cells and tissues. The optical sectioning properties result

from the use of point-­like illumination and detection configurations that restrict the relevant confocal

observation volume to the overlap region of the illumination and detection volumes. Various strategies are

employed to reduce this volume in an attempt to increase resolution of features of interest. Conventional

Abbe theory describes the interaction of the parameters that can be modified to improve lateral resolution,

including decreasing the wavelength of the incident light, increasing the refractive index of the specimen

medium, and increasing the acceptance angle (numerical aperture) of the optical system. The first two of

these are often restricted by the requirements of a given experiment with respect to the fluorophores that

can be employed and by the need to maintain viability of the specimen.

Even with higher numerical aperture objectives, the use of confocal pinholes restricts the illumination and

detection light paths to a much smaller portion of the maximum solid angle of the optical system, resulting

in elongation of the point spread function along the optical axis. Additional research effort has led to a

number of experimental techniques for reducing this restriction on axial resolution, as well as the effects

of various other factors, such as refractive index mismatches, which introduce optical aberrations. Among

the approaches being utilized to minimize this problem is the use of high numerical aperture water

immersion objectives.

The significance of the host of practical limitations to achieving ideal optical performance is that the

classical Abbe resolution calculations should not be misunderstood to describe the resolution of the

microscope. The resolution predicted by the diffraction limit pertains to optimum conditions, including an

unlimited signal-­to-­noise ratio. Fluorescence microscopy is limited in practice by low photon yield, and

requires a ratio of excitation to emitted photons that typically ranges in the millions. Because resolution,

even under high illumination levels, requires a certain contrast level, and contrast is affected by the ratio

Page 5: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

5/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

of signal to noise, the contrast level attained in practice depends ultimately on the number of photonsdetected. The resolution predicted by diffraction limits is never achieved in fluorescence observations ofbiological material, and for performance to be maximized the number of photons detected must beoptimized at every stage in the imaging process.

In addition to the number of photons collected and various optical aberrations, the sampling process itself,which is a fundamental feature of digital microscopy, plays a role in determining contrast, and thereforeresolution, in the fluorescence confocal microscope. As stated previously, the fact that digital confocalimages must be not only recorded, but processed and displayed within discrete picture elementsintroduces imaging variables that may be unfamiliar to microscopists that are new to digital imaging.Furthermore, the pixelation, or division of an image into finite picture elements, takes place at severalstages of the imaging process, and these must interact with each other to transfer image information fromthe specimen to the final visual image display. The possibility of mismatches among these discreteelements at the various stages is another factor that potentially limits image contrast and resolution.

The resolution imposed by the microscope optical system at the specimen level is sometimes describedon the basis of the resel, which is the smallest optically resolvable element. Depending upon the criterionutilized to define a detectable intensity difference between two elements, the size of the resel cancorrespond to the Rayleigh limit, the Sparrow limit, or another arbitrary definition. The effect of pixelation,or digitization, is initially manifested in the imaging sequence at the optical resolution level, through thesampling of the Airy patterns that correspond to point-­like features in the specimen. Partitioning, orpixelation, also takes place at the display stage of the imaging process, and differences in the use ofterminology in the published literature can lead to confusion in discussing the digitization processoccurring at various stages throughout the imaging sequence. The term resel refers specifically to thesmallest area of the specimen (object) that can be distinguished from neighboring areas, without regardto its subsequent detection, processing, or display. In three-­dimensional confocal imaging, a volumeresolution element is sometimes referred to as a voxel, although there is no reason to restrict the conceptof the resel to two dimensions, and the term may be used to describe the minimum spatially-­resolvedelement (in two or three dimensions) at the specimen, determined by the optics of the microscopesystem.

Figure 4 illustrates the mechanism by which the process of sampling the intensity of closely spaced Airypatterns reduces image contrast. The Airy pattern is generally assumed to be a smooth continuousfunction described by an infinite number of samples (or data points), as shown in the typical analogrepresentation of the intensity variation across the pattern. When considered as continuous functions, theAiry patterns exhibit their full intensity variations and produce the maximum theoretical contrast for a givenseparation distance. If divided into a finite number of measurement points or areas, by a scanner or digitalimaging device, for example, the smooth curves are transformed into a series of intensity values, each ofwhich can be stored in a computer memory location. By sampling at discrete intervals, the possibility isintroduced of overlooking the positions that include the minimum and maximum of the function.

Each pixel averages, or summarizes, the intensity response of the optical system within a specified area.Since the Airy function zero crossings occur at points, not areas, a pixel value cannot be zero for any finitepixel size. Similarly, the measured maximum value of the intensity peaks is reduced by the areaaveraging, and the combined effect of increasing the minimum and decreasing the maximum is to reducethe contrast. Consequently, the cut-­off distance is increased and the resolution decreased (for anycontrast criterion) by the pixelation process. Furthermore, if pixel size relative to the resel size is too large,ambiguity is introduced in the positions of the minima and maxima of the intensity response in the imageplane. The severity of the effect of partitioning the intensity into pixels depends directly upon the size ofthe pixel with respect to the Airy disk diameter, which in turn is related to the resel size imposed by thesystem optics at the wavelength of the image-­forming light.

It is unavoidable that any resolution inherent in the optical system, which is not sufficiently sampled by thedetector, is lost. The effect of pixelization is minimized as more pixels are utilized to describe the intensityvariations. Several detector designs acquire pixelated data as an intrinsic property of the detector (CCDcameras, for example), while others require a continuous analog signal to be digitized by an analog-­to-­digital converter or similar digitizer, following detection. The line scan of confocal laser scanning systemsutilizing a single detection channel operate in this manner. The relationship of the pixel size to thediameter of the Airy disk determines the number of pixels that are required to sample two adjacent Airydisks to achieve a certain contrast. The challenge in digital microscopy is to manage the relationshipamong optical resolution, sampling interval in the spatial domain, and the pixel dimensions of the displaydevice in order to maximize the capture and display of true specimen information, while minimizing visualartifacts that may be introduced by interactions among the various sampling (partitioning) stages.

An additional factor of practical interest in determining the contrast and resolution of captured images isthe intensity resolution, which governs the brightness value that is assigned to each image pixel. Byanalogy to the resel, the size of which is determined by the optical characteristics of the system, theminimum detectable difference in intensity that can be resolved depends upon electronic properties of thedetector, in particular its signal-­to-­noise ratio. When transferred to the image-­output stage, each pixel'sbrightness is described by a gray level, and the accuracy with which the brightness is representeddepends upon the relationship between the number of gray levels utilized and the smallest detectableintensity difference measured by the detector. When stored by the computer, each pixel corresponding toa spatial location in the image has an associated intensity value ranging from 0 to 255 for 8-­bit storage(256 gray levels). In confocal microscopy, the use of more than 256 gray levels is seldom justified by thedetection resolution, although there may be some value in discriminating more intensity levels for certaindata processing algorithms.

Page 6: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

6/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

It is important to recognize that pixelation at the image display stage is not unique to digital imaging.

Some form of "granularity" is inherent in every type of image presentation, and manipulations of the size

and organization of the grains are utilized to represent the required range of gray levels. Photographic

film has silver grains of various sizes, television displays are arrays of discrete horizontal lines, each of

which displays intensity variations based on electronic bandwidth, and half-­tone printing techniques group

black and white dots into pixels of various sizes to simulate continuous tonal variations. Video monitors

are able to vary the intensity of each displayed dot and achieve some tonal variation even while utilizing

one dot per pixel. However, in order to represent a sufficient number of gray levels to produce the visual

effect of continuous-­tone images, or to display color variations, a number of dots must be assigned to

represent each pixel. Film-­based photographic methods assign multiple silver grains to each image resel

in order to provide an adequate range of intensities necessary to give the appearance of continuous tonal

variation. In all of these methods, as more fundamental dots or image elements are grouped to achieve

greater tonal range, the appearance of the image becomes more "grainy", with the effect of reducing

apparent resolution. The inverse relationship between pixel size and the ability to display greater gray-­

scale range must be considered and balanced according to the imaging requirements.

The fact that all digital confocal microscopy images are acquired, processed, and displayed in the realm

of discrete partitions, or pixels, as opposed to being treated as a continuous representation of the

specimen data is not a problem of fundamental significance, but rather is a practical matter of imaging

technique. As long as the microscope is operated in accordance with applicable sampling theory, which

governs the sampling interval in space or time that is required to reproduce features of interest with

sufficient contrast, there is no significant limitation. The sampling criterion most commonly relied upon is

based on the well known Nyquist Theorem, which specifies the sampling interval required to faithfullyreconstruct a pure sine wave as a function of its frequency. The problem in practice is that use of the

zoom magnification control on typical confocal microscopes can easily be misused in a manner that

violates the Nyquist criterion.

The application of Nyquist sampling theory is usually explained by considering the specimen features in

the domain of spatial frequency rather than object size. In effect, the number of objects per spatial unit(the frequency) is the inverse of object size and emphasizes the importance of the spacing between

specimen features in image formation. Use of the spatial frequency domain is consistent with the practice

of evaluating performance of optical systems on the basis of their ability to maintain contrast (and

visibility) when transferring image information of different frequencies. All optical systems degrade

contrast in the imaging process, and this effect is more severe for higher spatial frequencies (small

spacing) than for lower frequencies (larger spacing). The contrast transfer function (CTF) of an opticalsystem is constructed by plotting the measured contrast in an image of test patterns consisting of periodic

arrays having alternating dark and light bars at a range of frequencies, or spacing intervals. Figure 5

illustrates a hypothetical CTF for an optical system, and includes curves for the system response to a testtarget having black and white bars (100 percent contrast), and to a target made up of gray and white bars

yielding only 30 percent contrast.

Examination of the contrast transfer curves of Figure 5 clearly illustrates the interdependence of

resolution and contrast, and the problem of some common assumptions that are made, which treat

resolution as a constant of instrument performance. Resolution is a somewhat arbitrary variable that is

meaningful only when considered within the framework of other related factors. When defined as the

highest spatial frequency that produces a certain contrast, it is easily assumed that any features having

frequencies within the stated resolution limit are equally visible, when in fact specimen features that are

originally of high contrast will be more clearly visible than those of lower contrast at every frequency up to

the contrast cut-­off frequency. The curves (Figure 5) illustrate that specimen features initially having only

30-­percent contrast due to staining characteristics or other factors, would not maintain the Rayleigh-­

specified 26-­percent contrast level at spatial frequencies anywhere near the theoretical limit, which

assumes equal contrast at all frequencies up to the resolution limit. The contrast of small features that are

slightly within the resolution limit produce contrast much lower than that of larger features after each has

been degraded by the transfer function of the imaging system.

The visibility of specimen features in the microscope image depends upon the contrast of the features

with respect to their surroundings, the performance of the optical system as reflected in the contrast

transfer function, signal-­to-­noise statistics, and the manner in which the signal is sampled for digitization.

Nyquist found that in order to faithfully reconstruct a pure sine wave, it must be sampled at least twice

during each cycle of the wave, or at two-­times the temporal frequency. The frequencies of interest in

Page 7: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

7/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

imaging are of a spatial nature, but the Nyquist theorem is equally applicable to this type of data. Theminimum sampling frequency employed in microscopy imaging applications is usually 2.3 times themaximum frequency to be reconstructed. The value of 2.3 is used (instead of 2, as suggested by Nyquist)as a practical matter in order to compensate for the low-­pass filtration that is required for perfect imagereconstruction in real optical systems.

Low-­pass filtration that is applied to the sampled data before the image is reconstructed is analogous tothe "filtration" done by the eye and brain to smooth pixelated data such as half-­tone images, or to movingfarther away from a display such as a large screen television in order to eliminate visible scan lines. Low-­pass filtration performs removal of sampling artifacts that are extraneous to the data, and helps to makethe image appear continuous. An ideal filter would permit sampling at 2-­times the highest frequency, butsince no such devices exist, experience has led to the generalization that sampling at 2.3-­times thehighest frequency is sufficient to accommodate the performance of real filters (often an integrator indigital-­to-­analog converters).

In practical operation of the microscope, there is often some uncertainty in estimating the highestfrequency that should be of concern in the specimen. Many specimens have sharply defined edges, andnumerous minute features that contribute extremely high-­frequency components to the signal. The Airyintensity distribution function or Gaussian distribution imposed by optical elements blurs thesecomponents to the extent that detail appears only within the optical resolution limit. Failure to sample all ofthe frequency components below the limit can produce image errors, although it is unnecessary to sampleat the higher frequencies, which will no longer carry specimen information following optical transfer (seeFigure 6). In some cases, specimens should be sampled at more than 2.3-­times the highest informationfrequency to allow for the possibility that the highest frequency was misjudged. In the confocal laserscanning microscope, the highest frequency to be sampled (f) is imposed by the optical system, and for aparticular resolution specification:

f = 1 / r (Resolution)

In order to preserve all of the information transferred by the optical system to the sampling (digitizing)stage of the microscope, having a contrast transfer function such as that illustrated in Figure 5, the pixelsize must be smaller than the inverse of 2.3-­times the cut-­off frequency (fcut-­off):

Pixel Size < 1 / (2.3 fcut-­off)

An accurate representation of specimen features, within the limits imposed by the experiment, requiresuse of a spatial sampling frequency and electronic properties, such as bandpass, which are matched tothe optical resolution of the system. In the confocal microscope, the optical resolution (the resel size) isprimarily determined by the wavelength of the light and the numerical aperture of the objective. The zoomcontrol on the microscope provides the capability to match the pixel size to the optical characteristics, andto meet the Nyquist criterion of acquiring samples at a minimum of twice the highest spatial frequency ofthe specimen signal. In general, the zoom control varies the size of the area scanned on the specimen,but does not change the number of pixels sampled. Consequently, pixel size varies inversely with zoomfactor. There is only one optimal zoom setting for a particular combination of wavelength, numericalaperture, and objective magnification: the one that provides a pixel size that matches the Nyquist criterionwhen referenced to the specimen space. At other than the optimum zoom setting, two different imagingproblems are created, depending upon whether the pixel size results in oversampling or undersampling(Figure 6) of the data.

Page 8: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

8/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

When operating at zoom settings of the microscope that are higher than that corresponding to the Nyquist

condition, smaller pixels are generated at the specimen, oversampling the data within the reduced

scanning area. Sampling with smaller pixel size than that required by the optical resolution not only

subjects the specimen to more severe photobleaching than is necessary, but also decreases the area that

can be scanned in a given period of time. It is sometimes considered useful to sample a specimen feature

at frequencies much higher than that corresponding to the Nyquist limit, in order to obtain information

about the position of a feature. Operating in this manner produces a condition referred to as hyper-­sampling, in which the position of an unresolved object may be located with greater precision thanafforded by the optical resolution of the microscope. Although no details about the specimen feature are

provided beyond that allowed by the resel dimensions, its location can be determined from the intensity

distribution of the blurry image. Provided that the feature is sampled by a large number of pixels, and it is

stable to the high photon flux that this oversampling implies, the centroid of the spot identifies its location

within the limits imposed by the signal noise.

A different type of problem results from undersampling of the specimen data. This situation is caused, as

is the case for oversampling, by a mismatch in the partitioning mechanisms of the imaging sequence. The

mismatch occurs between sampling frequency and resel size (resolution), and involves spatial

frequencies in the data that are higher than can be accurately represented by the sampling rate.

Undersampling typically occurs at lower microscope zoom settings for which the pixel size is larger than

that specified by the Nyquist criterion. In addition to smaller features being completely missed by the

sampling, other features may appear in the image that are not present in the specimen. This

phenomenon is possible due to a process known as aliasing, in which specimen features take on anentirely different character in the image than they actually have in the specimen. In effect, they appear

under an "alias", or false identity and are often masqueraded as spatial frequencies that do not exist in

the specimen.

The process is illustrated in Figure 7 in which specimen features of 3-­per-­micrometer frequency are

sampled at a frequency of 2-­per-­micrometer. The sample partitioning misses entire cycles of the

specimen data, and outputs a 1-­per-­micrometer signal that does not exist in the specimen. The aliased

image results from the undersampling mismatch between the resolved data and the detector

characteristics. When actual features are missed by so-­called "blind spots" produced by too-­large pixel

size, the aliasing may cause features to appear that are larger, smaller, or in different locations than they

are in the specimen. Once the sampling has occurred, there is no way to determine which image features

are real and which are artifacts.

In a situation such as the one illustrated in Figure 7, since specimen features of this size cannot be

accurately rendered at the chosen sampling rate, either the sampling rate must be increased, or some

form of filtration should be employed to eliminate frequencies that are beyond those that can be dealt with

by the system (those exceeding the Nyquist limit). Sampling at greater than 6 cycles per micrometer

would satisfy the Nyquist criterion, and accurately reproduce the specimen features, but if this is not

possible, the most practical solution in most cases is to adjust the microscope zoom setting to provide a

match between the sampling and resolution mechanisms. If a sufficient number of pixels is not achievable

for proper representation of the specimen features, the resolution should be limited, or degraded, optically

and/or electronically before partitioning of the data by sampling.

A popular method of limiting optical resolution is to employ smaller numerical aperture optical

components, or to underfill the rear focal plane of the objective to reduce its effective numerical aperture.

The chief disadvantage of this approach is that it reduces light collection efficiency, and the consequent

increase in specimen exposure time (for image acquisition) risks greater damage to sensitive specimens.

Another mechanism of degrading resolution to match sampling properties is to utilize larger pinhole

diameters (or other optical methods) to increase the effective size of the source disk beyond the

diffraction-­limited spot size. This technique has the advantage of degrading both lateral and axial

resolution in a linear manner, while numerical aperture changes affect axial resolution quadratically. In

either case, the confocal microscope retains its optical sectioning capabilities and characteristic sampling

flexibility. Additionally, it is commonly suggested that when detection systems, such as single-­channel

photomultiplier tubes are used (in which the pixelization of signal is not inherent to the detector), the

sampling resolution should be degraded electronically as well. For a given sampling rate, an integrator

can be employed ahead of the analog-­to-­digital converter to limit sampling of the signal voltage to the

Nyquist-­specified rate.

It is important to be able to operate the microscope with spot sizes consistent with the specimen and

fluorophore properties, and appropriate to the requirements of the experiment. Ideally, the instrument

should enable operation at both the diffraction-­limited configuration, and at larger spot sizes. There are a

number of experimental situations in which it is desirable to intentionally operate the microscope at less

than its diffraction-­limited resolution by breaking the rules pertaining to focal spot size, pinhole diameter,

Page 9: Preview of “Olympus Microscopy Resou in Confocal Microscopy”allows evaluation of the optical transfer properties of the combined system components. Although variations in field

12/17/12 Olympus Microscopy Resource Center | Confocal Microscopy -‐‑ Resolution and Contrast in Confocal …

9/9www.olympusmicro.com/primer/techniques/confocal/resolutionintro.html

Contributing Authors

Kenneth R. Spring -­ Scientific Consultant, Lusby, Maryland, 20657.

Thomas J. Fellers and Michael W. Davidson -­ National High Magnetic Field Laboratory, 1800 East Paul Dirac

Dr., The Florida State University, Tallahassee, Florida, 32310.

and sampling frequency, but the operator should always be aware of the circumstances justifying these

deviations.

BACK TO CONCEPTS IN CONFOCAL

© 2012 Olympus America Inc.

All rights reserved