Review Papers

Review of snapshot spectral imaging technologies

[+] Author Affiliations
Nathan Hagen

Rebellion Photonics Inc., 7547 South Freeway, Houston, Texas 77021

Michael W. Kudenov

North Carolina State UniversityDepartment of Electrical & Computer Engineering, 2410 Campus Shore Drive, Raleigh, North Carolina 27606

Opt. Eng. 52(9), 090901 (Sep 23, 2013). doi:10.1117/1.OE.52.9.090901
History: Received April 15, 2013; Revised July 30, 2013; Accepted August 1, 2013
Text Size: A A A

Open Access Open Access

Abstract.  Within the field of spectral imaging, the vast majority of instruments used are scanning devices. Recently, several snapshot spectral imaging systems have become commercially available, providing new functionality for users and opening up the field to a wide array of new applications. A comprehensive survey of the available snapshot technologies is provided, and an attempt has been made to show how the new capabilities of snapshot approaches can be fully utilized.

Figures in this Article

Spectral imaging sensors sample the spectral irradiance I(x,y,λ) of a scene and thus collect a three-dimensional (3-D) dataset typically called a datacube (see Fig. 1). Since datacubes are of a higher dimensionality than the two-dimensional (2-D) detector arrays currently available, system designers must resort to either measuring time-sequential 2-D slices of the cube or simultaneously measuring all elements of the datacube by dividing it into multiple 2-D elements that can be recombined into a cube in postprocessing. These two techniques are described here as scanning and snapshot.

Graphic Jump LocationF1 :

The portions of the datacube collected during a single detector integration period for (a) scanning and (b) snapshot devices.

The use of imaging spectrometers was rare before the arrival of 2-D CCD arrays in the 1980s, but steadily grew as detector technology advanced. Over the following 30 years, better optical designs, improved electronics, and advanced manufacturing have all contributed to improving performance by over an order of magnitude since that time. But the underlying optical technology has not really changed. Modified forms of the classic Czerny-Turner, Offner, and Michelson spectrometer layouts remain standard. Snapshot spectral imagers, on the other hand, use optical designs that differ greatly from these standard forms in order to provide a boost in light collection capacity by up to three orders of magnitude. In the discussion below, we provide what we believe is the first overview of snapshot spectral imaging implementations. After providing background and definitions of terms, we present a historical survey of the field and summarize each individual measurement technique. The variety of instruments available can be a source of confusion, so we use our direct experience with a number of these technologies [computed tomography imaging spectrometer (CTIS), coded aperture snapshot spectral imager (CASSI), multiaperture filtered camera (MAFC), image mapping spectrometry (IMS), snapshot hyperspectral imaging Fourier transform (SHIFT) spectrometer, and multispectral Sagnac interferometer (MSI)—each described in Sec. 4 below] to provide comparisons among them, listing some of their advantages and disadvantages.

Definitions and Background

The field of spectral imaging is plagued with inconsistent use of terminology, beginning with the field’s name itself. One often finds spectral imaging, imaging spectrometry (or imaging spectroscopy), hyperspectral imaging, and multispectral imaging used almost interchangeably. Some authors make a distinction between systems with few versus many spectral bands (spectral imaging versus imaging spectrometry), or with contiguous versus spaced spectral bands (hyperspectral versus multispectral imaging). In the discussion below, we use spectral imaging to refer simply to any measurement attempting to obtain an I(x,y,λ) datacube of a scene, in which the spectral dimension is sampled by more than three elements. In addition, we use the term snapshot as a synonym for nonscanning—i.e., systems in which the entire dataset is obtained during a single detector integration period. Thus, while snapshot systems can often offer much higher light collection efficiency than equivalent scanning instruments, snapshot by itself does not mean high throughput if the system architecture includes spatial and/or spectral filters. When describing a scene as dynamic or static, rather than specifying the rate of change in absolute units for each case, we simply mean to say that a dynamic scene is one that shows significant spatial and/or spectral change during the measurement period of the instrument, whether that period is a microsecond or an hour. Since snapshot does not by itself imply fast, a dynamic scene can blur the image obtained using either a snapshot or a scanning device, the difference being that whereas motion induces blur in a snapshot system, in a scanning system, it induces artifacts. In principle, blurring and artifacts are on a similar footing, but in practice one finds that artifacts prove more difficult to correct in postprocessing.

When describing the various instrument architectures, “pixel” can be used to described an element of the 2-D detector array or a single spatial location in the datacube (i.e., a vector describing the spectrum at that location). While some authors have tried introducing “spaxel” (spatial element) to describe the latter,1 this terminology has not caught on, so we simply use “pixel” when describing a spatial location whose spectrum is not of interest, and “point spectrum” when it is. While many authors refer to the spectral elements of a datacube as bands, we use “channel” to refer to individual spectral elements and reserve the use of bands for broad spectral regions [such as the visible band or the longwave IR (LWIR) band]. It is useful to have a terminology for a single horizontal plane of the datacube (the image taken over a single spectral channel), so we refer to this as a “channel image”. A single element of the datacube is referred to as a “voxel” (volume element). When describing the dimensions of the datacube, we use Nx, Ny, and Nw as the number of sample elements along the (x,y) spatial and λ spectral axes, respectively, so that the total number of datacube elements is given by NxNyNw. We try to avoid describing the datacube in terms of resolution, since many authors use this to refer to the number of sampling elements, while others refer to it as the number of resolvable elements. At the Nyquist sampling rate, these two differ by a factor of two, but if one always refers to the number of samples, then the meaning is clear. One can also note that it is problematic to talk about a single value for resolution when discussing computational sensors, since the number of resolvable elements for these systems varies with the scene—some scenes are easier to resolve than others.

For time-resolved (video imaging) systems, the data dimensions assume the form (Nx, Ny, Nw, Nt), where Nt is the number of frames captured during a video sequence. We refer to this dataset as a “hypercube”.

The amount of light collected by a given instrument is an important quantity, so we often refer to a given sensor’s throughput or efficiency. Whereas the former refers specifically to the AΩ product (or “étendue”), the latter is a ratio of the sensor’s throughput with respect to a reference sensor that can collect light from the entire datacube during the full measurement period and that also has ideal quantum efficiency. Whereas the ideal instrument for any given application always has an efficiency of 1, its throughput varies with different applications.

Also, many authors make a distinction between spectrometer, spectrograph, and spectroscope, with distinctions varying among researchers. We make no distinction here and generally stick to using spectrometer, except where this clashes with a given field’s nomenclature.

As in so much of optics, one can trace the beginnings of spectral imaging back into the nineteenth century, where one finds the astronomer P. J. C. Janssen using a double-slit monochromator to view the solar corona.2,3 The double-slit monochromator (at the time termed a spectroscope, or in this case a spectrohelioscope) was the only means of obtaining narrow-band spectra, and an image was obtained by spinning the device rapidly while viewing the exit slit with the eye. By adjusting the relative position of the exit slit with respect to the prism dispersion, one could thus view the same scene at different wavelengths. Although an important4 and clever setup, it was regarded by other researchers as clumsy.5 Some three decades later, Fabry and Perot developed their interferometric filter, which for the first time allowed astronomers to both view a full scene over a narrow spectral band and tune the filter wavelength.68 The tunable filter thus represented an important advance, giving scientists access to information that was previously difficult to obtain. This allowed them to build (x,y,λ) representations of the object in view, albeit laboriously.

An additional advantage of the Fabry–Perot interferometer was that it allowed a much higher light throughput than the double-slit monochromator, enabling users to view dimmer objects. This opened up a number of discoveries, but, as a spectral imager, it still suffered from two problems that simple cameras do not: motion artifacts and poor light collection efficiency. These two issues have plagued the field ever since. In order to overcome these problems, astronomers began looking for nonscanning instruments that could obtain the full 3-D dataset in a single measurement period—snapshot spectral imagers. The first published example of a snapshot spectral imaging system was not far behind. Bowen developed his image slicer9 by placing a series of tilted mirrors to slice the image into thin strips and then translate each strip to form a single long slit. Walraven later took this concept and created a design that was easier to manufacture, using only a thick glass plate (with a 45 deg angle cut into one end) cemented to a wedge-cut prism.10 The resulting device is shown in Fig. 2. Once a beam of light enters the glass plate, the beam reflects due to total internal reflection, except in those regions where the plate meets the prism edge, where the beam transmits. The succession of partial reflections and partial transmissions transforms the input beam into a long slit, which can then be used as the input to a slit spectrometer. Over the next few decades after Bowen’s initial invention, a number of astronomers adapted this slicing technique to telescopes around the world,1114 but the method still provided only modest gains over existing instrumentation. It was with the 3-D instrument on the William Herschel Telescope that the image slicing technique provided a large leap forward in performance, allowing for a snapshot measurement of a 16×16×2100 datacube (5.4×105voxels).15 For the first time, precision manufacturing, large-format detector arrays, and computers had advanced to the point that snapshot approaches could display capabilities going well beyond their scanning counterparts.

Graphic Jump LocationF2 :

Various views of a Bowen-Walraven image slicer, illustrating how the glass plate and wedge-cut prism combine to slice the optical beam into a long slit. Shapes shown in yellow indicate the light passing through the slicer; the beam reflecting within the top plate is not shown for clarity. (Adapted from Fig. 1 of 16.)

During this development, in 1958 Kapany introduced the concept of placing a coherent fiber bundle at the image plane and then reformatting the fiber output into a long thin line for easy adaptation to one or more slit spectrometers.17 But it appears not to have been implemented until 1980, when such a device was developed for the Canada-France-Hawaii telescope (CFHT) on Mauna Kea.18 Adaptations by other astronomers soon followed.19,20

A third snapshot spectral imaging technique was developed by Courtes in 1960, in which a lenslet array is used to create an array of demagnified pupils.21,22 These pupils fill only a small portion of the available space at the image, so with the proper geometry one can reimage them through a disperser to fill in the unused space with dispersed spectra. This concept was adapted to the CFHT and data began to be collected in 1987, producing datacubes with Nx×Ny=271 spatial elements and up to Nw=2200 spectral samples.23

These three techniques for snapshot spectral imaging—image slicing, fiber reformatting, and lenslet-based pupil array dispersion—are now widely described in astronomy as integral field spectroscopy (IFS), so we label the three techniques here as IFS-M, IFS-F, and IFS-L. The integral spectroscopy nomenclature appears to originate from the dissertation work of Chabbal,24 but the first publication in which one can find it used in the title is 25.

Outside of astronomy, one finds similar uses of slit spectrometers and tunable filter cameras, but the vast majority of uses did not require imaging. These include measurements such as mapping the spectrum of atomic emission/absorption lines, measuring the chemical composition of the atmosphere,26 and measuring the quantity of physical compounds.27 Outside of astronomy, the coupling of imaging with spectrometry did not take hold until much later, with the beginning of airborne remote sensing. Here spectral imaging was first used for agricultural assessment and management around 1966 (Refs. 28 and 29) and received a large boost with the launch of the Landsat remote sensing satellite in 1972.30 With the launch of Landsat also came the development of field-portable spectral imaging instruments needed to calibrate Landsat’s measurements. As spectral imaging became more widely used, researchers faced the same obstacles of scanning artifacts and poor light throughput as those faced by the astronomers, so that they too began to explore new methods, leading to a variety of new instruments. As single-pixel detectors gave way to linear and then 2-D detector arrays, system design options expanded, allowing for new approaches. The first of these new approaches derived from the natural idea of using multiple beamsplitters, in which the beam is split into independent spectral channels, with each channel directed to an independent camera. While this was a common choice for imaging in three spectral bands, especially for wideband measurements (such as visible plus near-IR), it did not take hold for more than four spectral channels. Hindrances included the difficulty of making large beamsplitters of high-enough quality and the limited ability to reduce the bulk and weight of the resulting system. With the increasing availability of thin-film filters, another natural choice involved using an array of filters coupled to an array of lenses. This, too, did not progress far, perhaps because of the difficulty of making large arrays of lenses with sufficient quality and correction for parallax effects. (31 comments that the first good thin film filters became available in the late 1930s. So we can expect that they did not become commercially available until the 1940s or 1950s.)

With the realization of advanced computers, the option of a computational sensing32 approach became feasible. The first of these new approaches was a computational sensor later named the CTIS. This device used a 2-D disperser to project a spectrally dispersed scene directly onto a detector array, allowing the spectral and spatial elements to multiplex. (One can experience the exact same thing by donning one of the many diffraction grating glasses that are given out at baseball games and fireworks displays.) The resulting data on the detector array are equivalent to tomographic projections of the datacube taken at multiple view angles, so that tomographic reconstruction techniques can be used to estimate the 3-D datacube from the set of 2-D projections. A compressive sensing approach33 to snapshot spectral imaging was also developed, promising to allow snapshot measurement of datacubes with more voxels than there are pixels on the detector array.34 As of this writing, there is still a significant amount of research that needs to be done with these computational techniques, as they have not yet shown performance that can compete with their noncomputational counterparts.35,36

In parallel with the development of snapshot methods, scanning techniques for spectral imaging also advanced. The development of linear digital sensor arrays allowed researchers to collect light from a number of voxels simultaneously. This involved imaging a linear spatial region through a sequential set of spectral filters or, more commonly, by collecting light over a full set of spectral channels while imaging one spatial location and then scanning in two dimensions over the full field of view—a point scanning spectrometer. However, unless used with very wide spectral channels (such as on the Landsat satellites), the poor light collection of these devices often made it difficult to use these instruments outside the lab.

Once 2-D detector arrays became available in the 1980s, these were rapidly adopted in imaging spectrometers, providing a large boost in light collection capacity.37 For the first time, researchers had available systems that could collect light emitted by thousands (and eventually millions) of voxels simultaneously. As array detectors advanced in size and performance, instrument designers took advantage of the new capability by increasing spatial and spectral resolution. Using a configuration in which the 2-D detector array is mapped to a spectral plus one-dimensional spatial (x,λ) dataset—a pushbroom spectrometer—made possible the first imaging spectrometers without moving or electrically tuned parts. If the sensor is placed on a steady-moving platform such as a high-flying aircraft, an Earth-observing satellite, or the object moves on a constant-motion conveyor, the second spatial dimension of the cube is obtained simply by scene motion across the instrument’s entrance slit. The removal of moving parts greatly improved the robustness of these devices and reduced their overall size, allowing for image spectrometry to develop into a mainstream technology for Earth-observing remote sensing. Other architectures are also available for scanning systems, as summarized below. A good review of these architectures and their relative SNRs is 38. Other surveys of scanning-based approaches include Refs. 39, 40, and 41. (Note that 40’s snapshot is equivalent to what we refer to here as a wavelength-scanned system or filtered camera.)

Point Scanning Spectrometer

The input spectrum is dispersed across a line of detector elements, allowing for very fast readout rates. The scene is scanned across the instrument’s input aperture with the use of two galvo mirrors (or just one if the instrument platform is itself moving), allowing for collection of a full 3-D datacube.

Pushbroom Spectrometer

The input aperture is a long slit whose image is dispersed across a 2-D detector array, so that all points along a line in the scene are sampled simultaneously. To fill out the spatial dimension orthogonal to the slit, the scene is scanned across the entrance aperture. This can take the form of objects moving along a conveyor belt, the ground moving underneath an airborne or spaceborne platform, or the scene scanned across the entrance slit by a galvo mirror.

Tunable Filter Camera

A tunable filter camera uses an adjustable filter (such as a filter wheel) or an electrically tunable filter, such as a mechanically tuned Fabry–Perot etalon,42,43 a liquid-crystal tunable filter (LCTF),44 and an acousto-optic tunable filter (AOTF).45 Response/switching times of the various approaches range from 1s for the filter wheel, to 50 to 500 ms for the LCTF and mechanically tuned Fabry–Perot, and to 10 to 50 μs for the AOTF.

Imaging Fourier Transform Spectrometer

An imaging Fourier transform spectrometer scans one mirror of a Michelson interferometer in order to obtain measurements at multiple optical path difference (OPD) values—the Fourier domain equivalent of a tunable filter camera.46,47 A more recent alternative method here is the birefringent Fourier-transform imaging spectrometer developed by Harvey and Fletcher-Holmes, which has the advantage of being less vibration sensitive due to its common path layout.48

Computed Tomography Hyperspectral Imaging Spectrometer

This is a scanning device closely related to the CTIS snapshot technique mentioned above and has the advantage of having a conventional disperser design and of being able to collect many more projection angles so that the reconstructed datacube has fewer artifacts. Its main disadvantage is that the detector is not used efficiently in comparison to alternative methods.49

Coded Aperture Line-Imaging Spectrometer

Although coded aperture spectrometry began as a method of scanning a coded aperture across the entrance slit of a conventional dispersive spectrometer, in Refs. 50 and 51 it was adapted to modern 2-D detector arrays, allowing for improved SNR at the cost of using larger pixel count detector arrays.

Each of these architectures uses passive measurement methods, in which no control is required over the illumination in order to resolve the incident light spectrum. Several of these techniques can be used in reverse to produce spectrally encoded illumination systems. In this approach, the object is illuminated with a well-defined spectral pattern, such that the imaging side no longer needs to have spectral resolution capability. For example, one can illuminate a scene with a broadband light source transmitted through a tunable filter and time the image acquisition to coincide with steps in the filter wavelength to produce a datacube of light emitted or reflected by the scene.

Finally, when comparing scanning and snapshot devices, we can note that the division between the two is not as black and white as one might expect. For example, designers have produced sensor architectures that mix both snapshot and scanning techniques, so that the number of scans required to gather a complete dataset is significantly reduced. The earliest example of this of which we are aware (although it seems likely that astronomers had tried this well before this time), is a patent by Busch,52 where the author illustrates a method for coupling multiple optical fibers such that each fiber is mapped to its own entrance slit within a dispersive spectrometer’s field of view. More recently, we can find examples such as Chakrabarti et al., who describe a grating spectrometer in which the entrance slit is actually four separate slits simultaneously imaged by the system.53 The respective slits are spaced apart such that the dispersed spectra do not overlap at the image plane. This setup can be used to improve light collection by a factor of four, at the expense of either increasing the detector size or reducing the spectral resolution by a factor of four. Ocean Optics’ SpectroCam is another example of a mixed approach, in which a spinning filter disk is combined with a pixel-level spectral filter array (more detail on the latter is given in Sec. 4.8) to improve the speed of multispectral image acquisition.

In addition, the fact that a given instrument is snapshot does not in itself imply that the device is fast. Scanning devices can have very short measurement times, and snapshot devices can potentially have very long ones. The essential difference is that snapshot devices collect data during a single detector integration period, and whether this is short or long depends on the application. For large-format snapshot spectral imagers in particular, the frame readout rate can also be rather long in comparison to the exposure time, so that a video sequence (or hypercube) can be time-aliased due to poor sampling if the two rates are not forced to be better matched.

Before attempting a comparison of existing methods for performing snapshot imaging spectrometry, we go through the menagerie of instrument architectures and attempt to summarize the basic measurement principles of each one in turn. Previous surveys (such as Refs. 54 and 55) have focused on instruments developed for astronomy, while we attempt here to discuss all of the techniques of which we are aware. In order to describe the various instruments, there is an inevitable flood of acronyms. A selection of these are summarized in Table 1, together with the section discussing each, the figure illustrating a typical system layout, and the first reference to the technique.

Table Grahic Jump Location
Table 1Location of summaries discussing each technology.
Integral Field Spectrometry with Faceted Mirrors (IFS-M, 1938)

In astronomy, the most common approaches to snapshot imaging spectrometry are the integral field techniques (based on mirror arrays, fiber arrays, and lenslet arrays: IFS-M, IFS-F, and IFS-L)—so called because each individual measurement of a datacube voxel results from integrating over a region of the field (the object). IFS-M was the first snapshot spectral imaging technology to emerge, beginning with Bowen’s image slicer. As originally conceived, this device was both difficult to manufacture (due to precision alignment of an array of small mirrors) and offered only a modest gain of 5× over existing slit spectrometers. In 1972, Walraven found a way to modify the design to use a prism-coupled plane parallel plate in place of the mirror array.10,16 This made the device much easier to align and assemble, but it was still not very widely used, partly because its use was primarily limited to slow beams with f-numbers >30.65

It was not until the “3-D” instrument was completed on the CFH telescope that image-slicer-type IFS could point to results that were difficult to obtain with existing instruments but that were readily obtained with the 3-D: performing spectroscopy and mapping the spatial distributions of dim extended objects (such as nebulae or distant galaxies).15Figure 3 shows a view of the image slicer subsystem for an IFS-M. The 3-D instrument design, however, had several limitations: in order to keep the pupils separated from one another, the image slicer’s facet tilts had to be large, thus inducing some defocus at the ends of the facet. Furthermore, it was difficult to optimize the volume to increase the number of facets and therefore the number of spatial elements in the datacube. In 1997, Content showed that allowing the microfacets to have curvature greatly reduces these constraints, allowing the designer to reduce the facet tilts and to place the pupils closer together—an approached he named “advanced imaging slicing.”66,67 While adding curvature eases system design constraints, it substantially complicates manufacture of the slicing mirror, and researchers began a period of investigating how to improve manufacturing techniques.68

Graphic Jump LocationF3 :

The system layout (a) for an integral field spectroscopy with faceted mirrors (IFS-M), and closeup (b) of the slicer mirror. For clarity, the layout only shows the chief rays corresponding to each mirror facet, and the spectrometer optics behind each pupil have been omitted. (These back-end optics are located behind each pupil in the array and include a collimating lens, disperser, reimaging lens, and detector array. If the mirror facets are given curvature, then the collimating lens is unnecessary.)

Because of its all-mirror approach, the IFS-M technique is well suited to measurements in the IR. It was also known that although scatter induced by manufacturing artifacts, and by diffraction at facet edges, was a serious problem at visible wavelengths, it was much less so in the near-IR and shortwave IR, and the image slicing method has been shown to excel in these spectral bands.69 As confidence in manufacturing techniques increased, Content later introduced the concept of microslicing (or IFS-μ), a technique that combines design elements of IFS-M with IFS-L. This enabled one to measure many more spatial elements in the datacube, at the expense of reduced spectral sampling.70 The basic idea of microslicing is to use the same slicing mirror as IFS-M, but with larger facets. The slicer allows the various strips in the image to be physically separated, and each is then passed through an anamorphic relay, such that one axis is stretched. This gives some extra space so that further down the optical path, the stretched image is relayed through an anamorphic lens array, which simultaneously unstretches and samples the image. The result is a format in which the various spatial elements of the image take the form of narrow strips that are sufficiently separated at the detector array that they can be spectrally dispersed without adjacent strips overlapping. The concept requires a mechanically elaborate design, but promises to achieve an impressive datacube size of 1200×1200×600 or 1500×1500×200.70

Integral Field Spectrometry with Coherent Fiber Bundles (IFS-F, 1958)

With the invention of coherent fiber bundles, it was quickly realized that one can image through the circular bundle on one end and squeeze the other end into a long thin line (creating what Kapany describes as a light funnel) to fit into the long entrance slit of a dispersive spectrometer.17 It was quite some time after this, however, that a working instrument was deployed. Instead, one first finds examples in the literature of systems that manipulate single fibers to positions within the field of view—for example, placing one fiber at each of a number of stars within the image.71 Rather than spectral imaging per se, this is better termed multiobject spectroscopy. Allowing true spectral imaging through a fiber bundle required first overcoming several hurdles, the first of which was manufacturability. The process of reformatting the exit face of the fiber bundle into a long slit generally produced a lot of broken fibers, and replacing them is laborious. Another drawback with the use of fibers was that it could be quite difficult to couple light efficiently into them, so that a significant amount of light was lost at the incident face of the fiber. Moreover, if the cladding of the fibers is insufficiently thick, the fibers show significant crosstalk, which quickly degrades the measurement quality. Finally, earlier fiber-based systems were also restricted by the limited spectral range transmitted by available fibers. Improvements in manufacturing and assembly techniques have steadily reduced each of these problems. The use of multimode fibers increased coupling efficiency,72 as did the use of lenslets registered to each fiber so that the light is focused onto the fiber cores and not in regions with poor coupling efficiency. (Lee et al. mention that these fiber-coupling lenslets should be used at f/4 to f/6.73) In addition, multimode fibers are also more robust, so that it became easier to format them into a line without breaking. With these and other advances, fiber-based IFS systems were successfully deployed on telescopes.74Figure 4 shows an example layout for an IFS-F system.

Graphic Jump LocationF4 :

The system layout for an integral field spectrometer with coherent fiber bundles (IFS-F): the object is imaged onto the face of a coherent fiber bundle. At the opposite end of the bundle, the fibers are splayed out (reformatted) into a linear array, which is compatible with the input of a standard slit spectrometer. At the input and output faces of the fiber bundle, there may be lenslets coupled to each fiber in order to improve light throughput.

A drawback with the fiber-based approach, it was learned, was that the light emerging from the exit face of the fiber was always faster (lower f-number) than it was on the input face—a phenomenon that was termed focal ratio degradation.19,69,75 Another phenomenon that was discovered when operating fiber-based systems at high spectral resolution was modal noise.76,77 In the midst of all of these developments, it was realized that the fiber-based approach also allowed astronomers to do away with the free-space dispersive spectrometers that have been the mainstay of so much of astronomical spectroscopy. If one can make use of components developed in the photonic industry, which has optimized miniaturization and manufacturing efficiency, then it should be possible to minimize instrument size and cost in modern astronomical telescopes.78 Since photonic components are designed primarily for single-mode inputs, this approach first required the development of a device that can convert a single multimode input into a series of single-mode outputs—what has been called a photonic lantern.79 One of the great advantages of this integrated photonic spectrograph approach80 is that it also enables the ability to use high-precision optical Bragg filters to filter out unwanted light from the atmosphere (often called OH suppression).81

In parallel with the developments in astronomy, advances in the technology also made it possible to build a commercial spectral imager based on the same concepts. The first nonastronomical instrument of which we are aware is that of Matsuoka et al., which delivered datacubes of 10×10×151 dimensions at 0.2 fps.82 This was followed by systems developed by other research groups.8387

Integral Field Spectroscopy with Lenslet Arrays (IFS-L, 1960)

The first discussion of using lenslet arrays for integral field spectroscopy appears to be that of Courtes in 1960, in which he proposes to use a lenslet array placed at the telescope’s focal plane. Such a configuration generates an array of pupil images—each mapped to a different field position.21,88 (The first review of this concept in English appears to be Meaburn’s in 1975, in what he calls an insect-eye Fabry–Perot spectrograph.89) This is the basis of the lenslet-based integral field approach. The lenslet array, placed near the image plane, creates a series of pupil images—one for each lenslet. Since the lenslets are at or near an image plane, each pupil image comprises all of the light integrated across the field positions corresponding to the spatial extent of the lenslet. The advantage here is that one can allow the lenslets to create faster beams (with lower f-number) than the original input, so that void space is created between the pupil images. One can then take advantage of this space by dispersing the light across it, allowing for detection of the spectrum. Figure 5 shows an example layout for an IFS-L system.

Graphic Jump LocationF5 :

The system layout for an integral field spectrometer with lenslet arrays (IFS-L).

The modern form of the IFS-L was first presented by Courtes in 1980,22,23 but the first published data from an instrument did not follow until 1988, when Courtes et al. present a system providing datacube dimensions of 44×35×580.90 As the concept spread, a number of other astronomers began creating designs for different telescopes.73,91,92 Borrowing from terminology used in fiber-based integral field spectrometry, one difficulty with the lenslet approach is focal ratio degradation. The beam behind the lenslet array must have a smaller f-number than the beam in front of it, placing higher étendue requirements on the back-end optics. One way of mitigating this issue is to use pinholes in place of or in tandem with the lenslet array.93 The tradeoff in doing this, of course, is that by spatial filtering one is reducing the system’s light throughput. In fact, if one replaces the lenslets with pinholes (so that one is sampling field positions rather than integrating across them), then the light throughput of the system becomes no better than a scanning approach.

While the IFS-L technique began in astronomy, its success brought it outside notice, and it was eventually adapted to other spectral imaging applications, with the first publication demonstrating a system achieving datacube dimensions of 180×180×20, measured at 30 fps and f/1.8 using a 1280×1024 CCD.94,95

Multispectral Beamsplitting (MSBS, 1978)

The idea of using multiple beamsplitters for color imaging has been around for quite some time.56 In this setup, three cemented beamsplitter cubes split incident light into three color bands, with each band observed by independent cameras [see Figs. 6(a) and 6(b)].96 While one can change the beamsplitter designs to adjust the measured spectral bands, it is not easy to divide the incident light into more than four beams without compromising the system performance. (Murakami et al., for example, limit themselves to four beamsplitters and attempt to use filters to increase the number of spectral channels.97) Thus, four spectral channels appear to be the practical limit of this approach. A closely related method is to use thin-film filters instead of the bulkier beamsplitter cubes/prisms to split the light,98 but this approach is still probably limited to about five or six spectral channels due to space limitations and cumulative transmission losses through successive filters. The space limitation can be overcome by using a single stack of tilted spectral filters operating in double-pass, which allows for the entire set of spectral images to be collected on a single detector array.99,100 (This is an approach we have previously termed filter stack spectral decomposition.101) Although more compact than the previous methods, since the filters are now operating in double-pass mode, transmission losses are doubled as well, so this method is limited to Nw<6.

Graphic Jump LocationF6 :

System layouts for four different multispectral bemsplitter (MSBS) implementations, using (a) monolithic beamsplitter blocks, (b) a sequence of spectral filters/beamsplitters, (c) a volume hologram optical element splitter, and (d) a stack of tilted spectral filters (or filter stack spectral decomposition).

A fifth implementation is to perform spectral splitting with a volume holographic element (VHE). Matchett et al. have shown that they can manufacture a VHE to split an incident beam in three, with each of the three spectrally filtered beams reflected at different angles.102 The VHE has the advantage of being a compact element with good reflection efficiency over a reasonable range of field angles. But it appears to be difficult to design the VHE for more than three channels. For example, Matchett et al. divided the system pupil in four, using a different VHE for each of the four sections, in order to produce a system that can measure 12 spectral channels. Matchett et al. state that this system has achieved 60 to 75% throughput across the visible spectrum.

Computed Tomography Imaging Spectrometry (CTIS, 1991)

As with every other snapshot spectral imaging technology, CTIS can be regarded as a generalization of a scanning approach—in this case a slit spectrometer. If one opens wide the slit of a standard slit spectrometer, spectral resolution suffers in that spatial and spectral variations across the width of the slit become mixed at the detector. However, if instead of a linear disperser one uses a 2-D dispersion pattern, then the mixing of spatial and spectral data can be made to vary at different positions on the detector. This allows tomographic reconstruction techniques to be used to estimate the datacube from its multiple projections at different view angles. Figure 7 shows the CTIS system layout. The CTIS concept was invented by Okamoto and Yamaguchi57 in 1991 and independently by Bulygin and Vishnyakov103 in 1991/1992, and was soon further developed by Descour, who also discovered CTIS’s missing cone problem.35,104,105 The instrument was further developed by using a custom-designed kinoform disperser and for use in the IR bands.106 The first high-resolution CTIS, however, was not available until 2001, providing a 203×203×55 datacube on a 2048×2048 CCD camera.107 Although the CTIS layout is almost invariably shown using a transmissive disperser, Johnson et al. successfully demonstrated a reflective design in 2005.108

Graphic Jump LocationF7 :

The system layout for a computed tomography imaging spectrometer (CTIS).

A major advantage of the CTIS approach is that the system layout can be made quite compact, but a major disadvantage has been the difficulty in manufacturing the kinoform dispersing elements. Moreover, since its inception, CTIS has had to deal with problems surrounding its computational complexity, calibration difficulty, and measurement artifacts. These form a common theme among many computational sensors, and the gap they create between ideal measurement and field measurements forms the difference between a research instrument and a commercializable one. While CTIS has shown a lot of progress on bridging this gap, it has not shown the ability to achieve a performance level sufficient for widespread use.

Multiaperture Filtered Camera (MAFC, 1994)

An MAFC uses an array of imaging elements, such as an array of cameras or a monolithic lenslet array, with a different filter placed at each element in order to collect portions of the full spectral band (see Fig. 8). This first MAFC implementation of which we are aware is the Fourier transform spectrometer approach by Hirai et al.58 Surprisingly, it was not until 2004 that we found an implementation like that shown in Fig. 8a, by Shogenji et al.,109 after which we find other research groups following the same approach.110,111 This layout uses an array of lenses coregistered to an array of spectral filters, with the entire set coupled to a monolithic detector array. (Note that the SHIFT system described in Sec. 4.12 describes a similar, but filterless, Fourier-domain approach.)

Graphic Jump LocationF8 :

Layouts for a multiaperture filtered camera (MAFC) system, using (a) the Shogenji design, (b) IMEC112 design, or (c) the Levoy/Horstmeyer design. (The color balls image is from http://commons.wikimedia.org/wiki/File:Toy_balls_with_different_Colors.jpg#.)

Another approach was first suggested by Levoy113 and implemented by Horstmeyer et al.114 This involves adapting a light field camera with a pupil plane filter: a lenslet array is placed at the objective lens’ image plane, so that the detector array lies at a pupil plane (as imaged by the lenslets). The image behind each lenslet is an image of the filter array, modulated by the scene’s average spectral distribution across the lenslet. While more complex and less compact than the Shogenji design, this second layout has the distinct advantage of being able to use a variety of objective lenses, so that zooming, refocusing, and changing focal lengths are easier to achieve. Mitchell and Stone developed a similar technique in which the linear variable filter is placed at the lenslet plane.115

While the MAFC is arguably the most conceptually simple approach to multispectral imaging, it does place some requirements on the scene’s light distribution in order to work well. For finite-conjugate imaging, it requires that the object irradiance is reasonably uniform in angle, so that the individual filters and lenslets sample approximately the same relative light distribution as do all of their counterparts. Specular objects are thus problematic at finite conjugates, and angular variations in irradiance will be mistakenly measured as spectral variations.

Tunable Echelle Imager (TEI, 2000)

The tunable echelle imager (TEI) can be considered as a modification of an echelle spectrometer to allow imaging. To make this possible, a Fabry-Perot etalon is placed into the optical train so that the input spectrum is sampled by the Fabry-Perot’s periodic transmission pattern.59,116 This produces gaps in the spatial pattern of the dispersed spectrum, allowing one to fill the gaps with a 2-D image (see Fig. 9). The light transmitted by the etalon is passed into a cross-disperser (for example, a grating whose dispersion is out of the plane of the page as shown in Fig. 9 and then into an in-plane disperser). The result is a characteristic 2-D echelle dispersion pattern, where the pattern is no longer composed of continuous stripes, but rather a series of individual images, each one of which is a monochromatic slice of the datacube (i.e., an image at an individual spectral channel). Under assumptions that the spectrum is smooth (i.e., bandlimited to the sampling rate of the instrument), this achieves a snapshot measurement of the datacube. However, the main tradeoff is that the system throughput is quite low: not only does the etalon reflect most of the input light, but the crossed-grating format is also inefficient. Moreover, for cases in which the object’s spectrum does not satisfy the bandlimit assumptions, the measurements are prone to severe aliasing unless scanning is used to measure the gaps in the spectral data.

Graphic Jump LocationF9 :

The system layout for the tunable echelle imager (TEI). The box numbers in the raw data simulation shown here indicate wavelengths of subimages in nm; the dashed boxes indicate a replicate order of the Fabry-Perot etalon. (Figure adapted from Figs. 3 and 6 of 59.)

Spectrally Resolving Detector Arrays (SRDA, 2001)

With the development of Bayer filter array cameras in the late 1970s, it became possible to produce single-pixel-level spectral filtering.117 The generalization from color imaging to multispectral imaging by increasing the number of filters is a small step (see Fig. 10), and there have been numerous such proposals.60,97,118126 The resulting instruments are extremely compact, since all of the spectral filtering is performed at the detection layer, but for several reasons this method has not been widely accepted in the spectral imaging community. The primary reasons are undoubtedly that manufacturing these pixel-level filters is difficult and that each pattern is highly specific, so that one cannot easily adjust the system in order to change spectral range or resolution. The fact that the multispectral approach will generally want to use detector arrays with higher pixel counts exacerbates the manufacturability problem. But another drawback is that, as with any filtering technique, an increase in spectral resolution produces a corresponding loss in light throughput.

Graphic Jump LocationF10 :

The system layout for a pixel-level filter array camera (one implementation of SRDA).

Although Bayer-type filter-array approaches are compact, convenient, and robust to perturbations such as temperature changes and vibration, they do have the disadvantage of requiring that the image is spatially bandlimited to the Nyquist limit of the filter array (a limit that is typically several times stricter than the Nyquist limit of the underlying pixel array). Without satisfying this assumption, the resulting reconstructed spectral images may show substantial aliasing effects, in which spatial variations in the scene will couple into erroneous spectral variations in the measured datacube. These effects can be minimized by defocusing the image in order to satisfy the bandlimit constraint.

The spatial/spectral filter-based approach is one approach toward adapting detector arrays to spectral resolving capability. A number of other approaches are also under development, which do not incorporate filters and thus have the potential for an increased detection efficiency. Although development on spectrally resolved detector arrays with more than three spectral bands has been underway for over 40 years,127,128 doing this for more than two spectral channels in snapshot detection mode has only been pursued recently. The first steps in this direction involved dual-band focal plane arrays [such as midwave IR (MWIR)/LWIR FPAs],129,130 but more recently it has involved elements such as cavity-enhanced multispectral photodetectors,131 elements composed of sandwiched electrodes and multiple detection layers,132 multilayer quantum-well infrared photodetectors (QWIPs),133 and transverse field detectors.134

The cavity-enhanced multispectral photodetector is designed by sandwiching several thin layers of amorphous silicon (used as the detection layers) in a resonance-enhanced cavity.135 Sun et al.131 and Wang et al.136 report on using this approach to measure two narrow spectral bands—one centered at 632 nm and another at 728 nm. Parrein et al. follow a closely related approach in which the detection element consists of layers of thin films sandwiched with multiple transparent collection electrodes.132 This measurement method combines the use of wavelength-dependent absorption depth with interference filters to create a stack of sensors having strong wavelength-dependent signal collection. The implementation of 132 so far allows only three spectral channels to be resolved per pixel, but the approach shows promise to allow resolution of more spectral channels.

Multilayer QWIPs are an alternative approach that has seen considerable research.130 Mitra et al., for example, present an IR detector consisting of a stack of multiple quantum well absorbers coupled through a diffractive resonant cavity.133 So far, this technique has been limited to three spectral channels, though the concept is generalizable to more.

Transverse field detection is a concept recently developed by Longoni et al.,134,137 allowing for depth-resolved detection of absorbed photons within the detection layer. Sensor electrodes spaced along the surface of the detector array are biased to different voltages in order to generate the transverse electric fields needed for each electrode to preferentially collect photocarriers produced at different depths. While this trades off spatial resolution for depth resolution, it provides a flexible method for depth-resolved detection.

In general, for multispectral detection (>3 spectral channels), each of the filterless approaches are still under development, and thus considerable work remains before they can be deployed for practical use.

Image-Replicating Imaging Spectrometer (IRIS, 2003)

Lyot invented his tunable filter in 1938 based on the idea of using polarizers to turn the wavelength dependence of retardation in thick waveplates into a wavelength dependence in transmission.138 Although the instrument was refined by others to use a different layout139 and to allow wider fields of view,140,141 it could never measure more than one spectral channel at once. In 2003, Harvey and Fletcher-Holmes described a generalization of Lyot’s filter in which the polarizers are replaced with Wollaston beamsplitting polarizers.61 By splitting each incident beam in two, this technique allows one to view a second spectral channel in parallel. By incorporating N Wollaston polarizers into the system, one can view 2N scenes simultaneously. The resulting layout, for a setup using three Wollaston polarizers, is shown in Fig. 11.

Graphic Jump LocationF11 :

The system layout for an image-replicating imaging spectrometer (IRIS). (The object simulated here is from http://en.wikipedia.org/wiki/File:Colouring_pencils.jpg.)

The IRIS approach is an elegant solution that makes a highly efficient use of the detector array pixels. So far, the IRIS approach has only been shown to operate with up to eight spectral bands,142 and it seems likely that difficulties of obtaining large-format Wollaston polarizers with sufficient birefringence and that can correct for polarization-dependent chromatic aberrations may limit this approach to about 16 spectral channels.143

Coded Aperture Snapshot Spectral Imager (CASSI, 2007)

CASSI was the first spectral imager attempting to take advantage of compressive sensing theory for snapshot measurement. Compressive sensing developed out of the work of Emmanuel Candes, Terence Tao, and David Donoho, typically involving the use of L1-norm reconstruction techniques to reconstruct data that would be termed insufficiently sampled by the Nyquist limit. The phrase is not intended to refer to a broad category of reconstruction algorithms (such as computed tomography) that can sometimes be said to permit compressive measurement.

The concept for CASSI developed from a generalization of coded aperture spectrometry.34 Coded aperture spectrometers replace the entrance slit of a dispersive spectrometer with a much wider field stop, inside which is inserted a binary-coded mask (typically encoding an S-matrix pattern or a row-doubled Hadamard matrix,144 see Fig. 12). This mask attempts to create a transmission pattern at each column within the slit such that each column’s transmission code is orthogonal to that of every other column. This follows directly from the properties of Hadamard matrices that each column of the matrix is orthogonal to every other column. The encoded light, transmitted by the coded mask within the field stop, is then passed through a standard spectrometer back-end (i.e., collimating lens, disperser, reimaging lens, and detector array). Because the columns of the coded mask are orthogonalizable, when they are smeared together by the disperser and multiplexed on the detector array, they can be demultiplexed during postprocessing. The resulting setup allows the system to collect light over a wide aperture without sacrificing the spectral resolution that one would lose by opening wide the slit of a standard slit spectrometer. The tradeoff is a factor of two in light loss at the coded mask, and in some noise enhancement due to the signal processing.

Graphic Jump LocationF12 :

Top: the system layout for a coded aperture snapshot spectral imager (CASSI), showing only the single-disperser configuration. Bottom: the pattern on the detector array due to imaging a coded aperture mask through a disperser, for an object that emits only three wavelengths (the wavelengths used in the example image here are the shortest, middle, and longest wavelengths detected by the system).

The theory of orthogonal codes only requires that the light is uniformly distributed in one axis; the other axis can be used for imaging. This is analogous to a slit spectrometer, which can image across its entrance slit. Using an anamorphic objective lens, one can achieve this by imaging the entrance pupil onto the field stop in one axis, while imaging the object onto the field stop along the orthogonal axis. Although one can consult references 145, 146, or 147 for further details on coded aperture spectral imaging, none of these sources mention the requirements for anamorphic front optics needed to achieve line-imaging with snapshot coded aperture spectrometry.

Compressive sensing allows one to take a similar procedure and apply it to snapshot spectral imaging, measuring (x,y,λ) in a snapshot and not just (x,λ). The primary differences from the slit imaging case are that the aperture code is no longer an orthogonal matrix but a random binary matrix and that the reconstruction algorithm becomes much more complex. The generalization proceeds as follows. If one replaces the anamorphic objective lens with a standard one, and images the object directly onto the coded aperture mask, then the irradiance projected onto the detector array after passing through the disperser will be a mix of spatial and spectral elements of the datacube (see Fig. 12). The spatial-spectral projection at the detector array is modulated by the binary mask in such a way that each wavelength of the datacube experiences a shifted modulation code. If this code satisfies the requirements of compressive sensing, then this is all one needs in order to use compressive sensing reconstruction algorithms to estimate the object datacube.34 The resulting system layout is not only extremely compact, but also uses only a modest size detector array, so it is capable of imaging at high frame rates. Wagadarikar et al. showed the ability to capture 248×248×33 datacubes at 30 fps, though postprocessing of the raw data to produce a hypercube (video datacube sequence) consumed many hours of computer time.148 More recent algorithms are much faster.36

Though compressive sensing holds out great promise for future instrument development, designers have not yet succeeded in creating an architecture that replicates its basic requirements well. In rough form, one can summarize the requirements as follows. The most basic feature is that the measured object has to be compressible in some space. For example, for an image represented in a wavelet space, it is well known that one can almost always use fewer coefficients than the number of pixels you wish to reconstruct in the pixel domain—this is a compressible space for a typical image. If one were to measure the object such that each measurement value was the projection of the object onto a basis function in this compressive space, then one would need far fewer measurements than if the system measured all of a datacube’s voxels directly. Unfortunately, one generally cannot design an optical system such that the object can be measured in the compressive space directly (such as measuring an object’s wavelet coefficients). Compressive sensing, however, provides an alternative that is almost as good. Measurement vectors (we avoid calling them basis functions because they do not satisfy the general definition of a basis) that are composed of columns within a random incoherent measurement matrix have been shown to replicate the properties of measuring in a compressive basis to very high probability. Using this type of measurement matrix, however, comes with some additional requirements. First, the measurement vectors and the compressible basis functions must be mutually incoherent, which means that any element in one cannot be expressed as a sparse linear combination of elements from the other.149 One can think of this in rough terms as having measurement vectors highly spread out when expressed in the basis vectors of the chosen compressible space, and vice versa. Also, the measurement vectors must satisfy isotropy, which means that they should have unit variance and are uncorrelated.150 Orthogonal matrices are one example of systems that satisfy this property.

Once a measurement is completed, the user applies a reconstruction algorithm to estimate the object datacube. The typical procedure for this is for the user to choose the compressible basis in which to work (that is, one must know a priori a basis in which the object is compressible) and apply an algorithm to estimate both the magnitudes of the coefficients and the set of measurement vectors comprising the space described by the product of the sensing and signal compression matrices. The algorithm estimates the coefficients and basis vectors by choosing an object representation that best approximates the actual measurement while penalizing representations that have a larger number of coefficients (i.e., are less sparse). For CASSI, a common choice of basis has been total variation space in the x-y dimensions (i.e., the object’s gradient image is assumed to be highly compressible). While it is possible to adapt the reconstruction algorithm to search for an optimal compressible basis (so that the user need not know this a priori), this greatly burdens an already computationally intensive problem.

The mathematics underlying this measurement approach has seen a rapid advance in the last decade, but implementing its requirements in hardware has been challenging. One of the main obstacles is that, in order to obtain sufficient compression, the feature sizes used to create coded projections are near the scale of the optical resolution. Not only does this mean that characterizing the system measurement matrix requires a great deal of care, but since the compressive sensing reconstruction methods are currently also sensitive to perturbations of the system matrix, the system is prone to artifacts. Because of these issues, subsequent implementations of CASSI have used linear scanning, such that a 640×480×53 datacube was reconstructed from a set of 24 frames (for a full collection time of about 2 s).151 In comparison with equivalent scanning instruments, this final result is disappointing, as even the authors admitted (see the concluding remarks of 36), so it appears that considerable work remains before we can take advantage of the compressive sensing paradigm.

Image Mapping Spectrometry (IMS, 2009)

Image slicing, as accomplished by the IFS-M technique discussed in Sec. 4.1, is best suited for measurements with low spatial and high spectral resolution. For many applications such as microscopy, however, spatial sampling is the more important quantity, and spectral sampling with only 10 to 40 elements is more common. This makes the IFS-M an impractical approach in this field. While the microslicing implementation (IFS-μ) is capable of achieving a much higher spatial sampling, this comes at the cost of a serious increase in system design complexity. An alternative approach is IMS. Like IFS-M, a microfaceted mirror is placed at an image plane. Unlike image slicing, however, many of the mirror facets share the same tilt angle, so that multiple slices of the image are mapped to each individual pupil. The resulting pattern, as seen by the detector array, resembles that of seeing a scene through a picket fence. If there are nine individual pupils in the system, then the spaces between the fence’s slats are 1/9th the slat widths (see Fig. 13). One can see only thin slices of the scene, but there are nine facet images on the detector array, each representing the image shifted by 1/9th of a slat width relative to the others. Assembling all nine subimages thus allows one to replicate the original scene. The advantage of obtaining these facet images is that one has separated the elements of the scene enough so that there is space to allow dispersion. By allowing each pupil to be shared among many mirror facets, the system design becomes much more compact and allows for higher spatial resolution.

Graphic Jump LocationF13 :

The image mapping spectrometer (IMS) system layout. Three different raw images are shown corresponding to a setup in which (a) the lorikeet is being imaged through the full system (as shown), (b) the lorikeet is being imaged in a system in which the prism array has been removed, and (c) a spectrally and spatially uniform object is being imaged through the full system. (For clarity, in all three examples of raw detector data shown here, the data are shown in color as if imaged by a color detector array, even though a monochromatic array is used in existing instruments.) (The lorikeet image is from http://commons.wikimedia.org/wiki/File:Rainbow_Lorikeet_MacMasters.jpg.)

The first IMS instrument (called an ISS at the time) provided a 100×100×25 datacube using a large-format CCD array,62 but this was later improved to 350×350×46.152

As with image slicing (IFS-M), the primary drawback of the IMS is the need for very high precision for cutting the image mapper, which is the central element of the system. Current ultraprecision lathes have advanced to the point where it is possible to make these elements on monolithic substrates, though considerable care is involved.

Snapshot Hyperspectral Imaging Fourier Transform Spectrometer (SHIFT, 2010)

The SHIFT spectrometer 63,153 performs its spectral measurement in the time domain and acquires image information using a division of aperture approach. Conceptually, the idea is an extension of an earlier formulation—the multiple-image Fourier transform spectrometer (MIFTS) developed by Hirai in 1994.58 However, while the original MIFTS was based on a Michelson interferometer and lens array, the SHIFT spectrometer is based on a pair of birefringent Nomarski prisms behind a lenslet array. As depicted in Fig. 14, an N×M lenslet array images a scene through two linear polarizers surrounding a pair of Nomarski prisms. Thus, N×M subimages are formed on a detector array. Rotating the prisms by a small angle δ relative to the detector array enables each one of the subimages to be exposed to a different OPD. Therefore, a 3-D interferogram cube can be assembled by sequentially extracting each one of the subimages. Fourier transformation, along the OPD axis of the interferogram cube, enables reconstruction of the 3-D datacube. This prism-based design allows for a reduced system volume and an improved robustness to vibration.

Graphic Jump LocationF14 :

The system layout for a snapshot hyperspectral imaging Fourier transform spectrometer (SHIFT).

As with the original MIFTS system, the SHIFT can be considered as the Fourier transform analog of the MAFC, and it offers many of the same advantages, such as compactness. Unlike the MAFC, however, it also offers continuously sampled spectra and is more easily fabricated due to its use of birefringent prisms. On the other hand, it also shares the MAFC’s disadvantage of suffering from parallax effects.

Multispectral Sagnac Interferometer (MSI, 2010)

The MSI64 is an extension of channeled imaging polarimetry154 to imaging spectroscopy. The idea was conceptually demonstrated using the MSI depicted in Fig. 15. In this interferometer, incident light is divided by a beamsplitter into two counter-propagating components. The component that was initially reflected by the beamsplitter begins its propagation in the z direction, where it gets diffracted away from the optical axis by grating G2. Reflection off of mirrors M2 and M1 guides the beam to grating G1, where the beam is diffracted in the opposite direction. The converse path is taken by the component that was initially transmitted by the beamsplitter, thus allowing both beams to exit the interferometer collimated and dispersed laterally. The lateral dispersion induced by the gratings produces a lateral shear S between the emerging wavefronts, where S is linearly dependent on the free-space wavelength λ. Additionally, gratings G1 and G2 are multiple order diffractive structures; i.e., the blaze contains deep grooves that impose more than one wave of OPD. When these high-order sheared beams converge through an objective lens and onto a detector array, a 2-D spatial interference pattern is generated. The spatial frequency of this interference is directly proportional to the diffraction order and, therefore, directly related to the given order’s spectral transmission.

Graphic Jump LocationF15 :

The system layout for a multispectral Sagnac interferometer (MSI).

These interference fringes are measured by the detector array as a superposition of coincident amplitude-modulated spatial carrier frequencies. A 2-D Fourier transformation can be taken of the raw data to window and filter the amplitude-modulated channels in the frequency domain. Inverse Fourier transformation yields the 2-D spatial data that correspond to the unique spectral passbands generated by the gratings.

This system is essentially a multispectral approach in which many unique spectral slices are measured simultaneously on coincident interference fields. Its advantages include inherent spatial coregistration between the bands while offering simple postprocessing. However, its disadvantages lie in its implementation, namely, the spectral bands must correspond to the grating’s orders. Additionally, only one dimension in the Fourier space can be used to modulate spatial and spectral information. Therefore, more work must be done to make this technique a viable competitor to any of the other methods mentioned here.

There are many ways to compare the various snapshot implementations, such as compactness, speed