Fourier Slice Photography
|
|
|
- Kory Chapman
- 9 years ago
- Views:
Transcription
1 Fourier Slice Photography Ren Ng Stanford University Abstract This paper contributes to the theory of photograph formation from light fields. The main result is a theorem that, in the Fourier domain, a photograph formed by a full lens aperture is a 2D slice in the 4D light field. Photographs focused at different depths correspond to slices at different trajectories in the 4D space. The paper demonstrates the utility of this theorem in two different ways. First, the theorem is used to analyze the performance of digital refocusing, where one computes photographs focused at different depths from a single light field. The analysis shows in closed form that the sharpness of refocused photographs increases linearly with directional resolution. Second, the theorem yields a Fourier-domain algorithm for digital refocusing, where we extract the appropriate 2D slice of the light field s Fourier transform, and perform an inverse 2D Fourier transform. This method is faster than previous approaches. Keywords: Digital photography, Fourier transform, projectionslice theorem, digital refocusing, plenoptic camera. 1 Introduction A light field is a representation of the light flowing along all rays in free-space. We can synthesize pictures by computationally tracing these rays to where they would have terminated in a desired imaging system. Classical light field rendering assumes a pin-hole camera model [Levoy and Hanrahan 1996; Gortler et al. 1996], but we have seen increasing interest in modeling a realistic camera with a lens that creates finite depth of field [Isaksen et al. 2000; Vaish et al. 2004; Levoy et al. 2004]. Digital refocusing is the process by which we control the film plane of the synthetic camera to produce photographs focused at different depths in the scene (see bottom of Figure 8). Digital refocusing of traditional photographic subjects, including portraits, high-speed action and macro close-ups, is possible with a hand-held plenoptic camera [Ng et al. 2005]. The cited report describes the plenoptic camera that we constructed by inserting a microlens array in front of the photosensor in a conventional camera. The pixels under each microlens measure the amount of light striking that microlens along each incident ray. In this way, the sensor samples the in-camera light field in a single photographic exposure. This paper presents a new mathematical theory about photographic imaging from light fields by deriving its Fourier-domain representation. The theory is derived from the geometrical optics of image formation, and makes use of the well-known Fourier Slice Theorem [Bracewell 1956]. The end result is the Fourier Slice Photography Theorem (Section 4.2), which states that in the Fourier domain, a photograph formed with a full lens aperture is a 2D slice in the 4D light field. Photographs focused at different depths correspond to slices at different trajectories in the 4D space. This Fourier representation is mathematically simpler than the more common, spatial-domain representation, which is based on integration rather than slicing. Sections 5 and 6 apply the Fourier Slice Photography Theorem in two different ways. Section 5 uses it to theoretically analyze the performance of digital refocusing with a band-limited plenoptic camera. The theorem enables a closed-form analysis showing that the sharpness of refocused photographs increases linearly with the number of samples under each microlens. Section 6 applies the theorem in a very different manner to derive a fast Fourier Slice Digital Refocusing algorithm. This algorithm computes photographs by extracting the appropriate 2D slice of the light field s Fourier transform and performing an inverse Fourier transform. The asymptotic complexity of this algorithm is O(n 2 log n), compared to the O(n 4 ) approach of existing algorithms, which are essentially different approximations of numerical integration in the 4D spatial domain. 2 Related Work The closest related Fourier analysis is the plenoptic sampling work of Chai et al. [2000]. They show that, under certain assumptions, the angular band-limit of the light field is determined by the closest and furthest objects in the scene. They focus on the classical problem of rendering pin-hole images from light fields, whereas this paper analyzes the formation of photographs through lenses. Imaging through lens apertures was first demonstrated by Isaksen et al. [2000]. They qualitatively analyze the reconstruction kernels in Fourier space, showing that the kernel width decreases as the aperture size increases. This paper continues this line of investigation, explicitly deriving the equations for full-aperture imaging from the radiometry of photograph formation. More recently, Stewart et al. [2003] have developed a hybrid reconstruction kernel that combines full-aperture imaging with band-limited reconstruction. This allows them to optimize for maximum depth-of-field without distortion. In contrast, this paper focuses on fidelity with full-aperture photographs that have finite depth of field. The plenoptic camera analyzed in this report was described by Adelson and Wang [1992]. It has its roots in the integral photography methods pioneered by Lippman [1908] and Ives [1930]. Numerous variants of integral cameras have been built over the last century, and many are described in books on 3D imaging [Javidi and Okano 2002; Okoshi 1976]. For example, systems very similar to Adelson and Wang s were built by Okano et al. [1999] and Naemura et al. [2001], using graded-index (GRIN) microlens arrays. Another integral imaging system is the Shack-Hartmann sensor used for measuring aberrations in a lens [Tyson 1991]. A different approach to capturing light fields in a single exposure is an array of cameras [Wilburn et al. 2005]. 3 Background Consider the light flowing along all rays inside the camera. Let L F be a two-plane parameterization of this light field, where
2 Figure 1: We parameterize the 4D light field, L F, inside the camera by two planes. The uv plane is the principal plane of the lens, and the xy plane is the sensor plane. L F (x, y, u, v) is the radiance along the given ray. Figure 2: Reparameterizing the light field by moving the sensor plane from F to F =(α F ). The diagram shows the simplified 2D case involving only x and u. By similar triangles, the illustrated ray that intersects the lens at u, and the F plane at x, also intersects the F plane at u +(x u)/α. L F (x, y, u, v) is the radiance along the ray traveling from position (u, v) on the lens plane to position (x, y) on the sensor plane (see Figure 1). F is the distance between the lens and the sensor. Let us consider how photographs are formed from the light field in conventional cameras and plenoptic cameras. Conventional Camera The image that forms inside a conventional camera is proportional to the irradiance [Stroebel et al. 1986], which is equal to a weighted integral of the radiance coming through the lens: E F (x, y) = 1 L F 2 F (x, y, u, v)cos 4 φdudv, (1) where F is the separation between the lens and the film, E F (x, y) is the irradiance on the film at (x, y), and φ is the angle between ray (x, y, u, v) and the film plane normal. The integration is a physical process that takes place on the sensor surface, such as the accumulation of electrons in a pixel of a CCD that is exposed to light. The derivations below assume that the uv and xy planes are infinite in extent, and that L is simply zero beyond the physical bounds of the lens and sensor. To shorten the equations, they also absorb the cos 4 φ into the light field itself, be defining L(x, y, u, v) = L(x, y, u, v)cos 4 φ. This contraction is possible because φ depends only on the angle that the ray makes with the light field planes. As a final note about Eq. 1, it is worth mentioning that the light field inside the camera is related to the light field in the world via the focal length of the lens and the thin lens equation. To keep the equations as simple as possible, however, the derivations deal exclusively with light fields inside the camera. Plenoptic Camera In the case of a plenoptic camera that measures light fields, image formation involves two steps: measurement and processing. For measurement, this kind of camera uses a flat sensor that provides a directional sampling of the radiance passing through each point on the sensor. Hence, if the sensor is at a depth F from the lens, it samples L F directly. During processing, this light field can be used to compute the conventional photograph at any depth E F, where F need not be the same as F. This is done by reparameterizing L F to produce L F and then applying Eq. 1. A simple geometric construction (see Figure 2) shows that if we let α = F /F, L F (x, y, u, v) =L (α F ) (x, y, u, v) = L F (u +(x u)/α, v +(y v)/α, u, v) = L F (u(1 1/α)+x/α, v(1 1/α)+y/α, u, v). (2) In other words, L F is a 4D shear of L F, a fact that was derived previously by Isaksen et al. [2000] in the first demonstration of digital refocusing. Combining Eqs. 1 and 2 leads to the central definition of this section, which codifies the fundamental relationship between photographs and light fields: Photography Operator Let P α be the operator that transforms a light field at a sensor depth F into the photograph formed on film at depth (α F ). If P α [L F ] represents the application of P α to light field L F, then P α [L F ](x, y) =E (α F ) (x, y) = (3) 1 α 2 F 2 L F (u(1 1/α)+ x/α, v(1 1/α)+ y/α, u, v) du dv. This definition is the basis for digital refocusing, in that it explains how to compute photographs at different depths from a single measurement of the light field inside the camera. The photography operator can be thought of as shearing the 4D space, and then projecting down to 2D. 4 Photographic Imaging in Fourier-Space The key to analyzing Eq. 3 in the Fourier domain is the Fourier Slice Theorem (also known as the Fourier Projection-Slice Theorem), which was discovered by Bracewell [1956] in the context of radio astronomy. This theorem is the theoretical foundation of many medical imaging techniques [Macovski 1983]. The classical version of the Fourier Slice Theorem [Deans 1983] states that a 1D slice of a 2D function s Fourier spectrum is the Fourier transform of an orthographic integral projection of the 2D function. The projection and slicing geometry is illustrated in Figure 3. Conceptually, the theorem works because the value at the origin of frequency space gives the DC value (integrated value) of the signal, and rotations do not fundamentally change this fact. From this perspective, it makes sense that the theorem generalizes to higher dimensions. For a different kind of intuition, see Figure 3 in Malzbender s paper [1993]. It also makes sense that the theorem works for shearing operations as well as rotations, because shearing a space is equivalent to rotating and dilating the space. These observations mean that we can expect that the photography operator, which we have observed is a shear followed by projection, should be proportional to a dilated 2D slice of the light field s 4D Fourier transform. With this intuition in mind, Section 4.1 and 4.2 are simply the mathematical derivations in specifying this slice precisely, culminating in Eqs. 8 and Generalization of the Fourier Slice Theorem Let us first digress to study a generalization of the theorem to higher dimensions and projections, so that we can apply it in our 4D space.
3 A closely related generalization is given by the partial Radon transform [Liang and Munson 1997], which handles orthographic projections from N dimensions down to M dimensions. The generalization presented here formulates a broader class of projections and slices of a function as canonical projection or slicing following an appropriate change of basis (e.g. a 4D shear). This approach is embodied in the following operator definitions. Integral Projection Let IM N be the canonical projection operator that reduces an N-dimensional function down to M- dimensions by integrating out the last N M dimensions: IM N [f](x 1,...,x M )= f(x 1,...,x N ) dx M+1...dx N. Slicing Let SM N be the canonical slicing operator that reduces an N-dimensional function down to an M dimensional one by zero-ing out the last N M dimensions: SM N [f](x 1,...,x M )=f(x 1,...,x M, 0,...,0). Change of Basis Let B denote an operator for an arbitrary change of basis of an N-dimensional function. It is convenient to also allow B to act on N-dimensional column vectors as an N N matrix, so that B [f](x) =f(b 1 x), where x is an N-dimensional column vector, and B 1 is the inverse of B. Fourier Transform Let F N denote the N-dimensional Fourier transform operator, and let F N be its inverse. With these definitions, we can state a generalization of the Fourier slice theorem as follows: THEOREM (GENERALIZED FOURIER SLICE). Let f be an N- dimensional function. If we change the basis of f, integral-project it down to M of its dimensions, and Fourier transform the resulting function, the result is equivalent to Fourier transforming f, changing the basis with the normalized inverse transpose of the original basis, and slicing it down to M dimensions. Compactly in terms of operators, the theorem says: F M I N M B S N M B T B T FN, (4) where the transpose of the inverse of B is denoted by B T, and B T is its scalar determinant. A proof of the theorem is presented in Appendix A. Figure 4 summarizes the relationships implied by the theorem between the N-dimensional signal, M-dimensional projected signal, and their Fourier spectra. One point to note about the theorem is that it reduces to the classical version (compare Figures 3 and 4) for N =2,M =1and the change of basis being a 2D rotation matrix (B = R θ ). In this case, the rotation matrix is its own inverse transpose (R θ = R θ T ), and the determinant Rθ T equals 1. The theorem states that when the basis change is not orthonormal, then the slice is taken not with the same basis, but rather with the normalized transpose of the inverse basis, (B T / B T ). In 2D, this fact is a special case of the so-called Affine Theorem for Fourier transforms [Bracewell et al. 1993]. 4.2 Fourier Slice Photography This section derives the equation at the heart of this paper, the Fourier Slice Photography Theorem, which factors the Photography Operator (Eq. 3) using the Generalized Fourier Slice Theorem (Eq. 4). Figure 3: Classical Fourier Slice Theorem, using the operator notation developed in Section 4.1. Here R θ is a basis change given by a 2D rotation of angle θ. Computational complexities for each transform are given in square brackets, assuming n samples in each dimension. Figure 4: Generalized Fourier Slice Theorem (Eq. 4). Transform relationships between an N-dimensional function G N,anM-dimensional integral projection of it, G M, and their respective Fourier spectra, G N and G M. n is the number of samples in each dimension. Figure 5: Fourier Slice Photography Theorem (Eq. 9). Transform relationships between the 4D light field L F, a lens-formed 2D photograph E α F, and their respective Fourier spectra, L F and E α F. n is the number of samples in each dimension. The first step is to recognize that the photography operator (Eq. 3) indeed corresponds to integral projection of the light field following a change of basis (shear): P α [L F ] 1 α 2 F 2 I4 2 B α [L F ], (5) which relies on the following specific change of basis: Photography Change of Basis B α is a 4D change of basis defined by the following matrices: B α = α 0 1 α 0 0 α 0 1 α Bα 1 = 1 α α α α
4 Directly applying this definition and the definition for I2 4 verifies that Eq. 5 is consistent with Eq. 3. We can now apply the Fourier Slice Theorem (Eq. 4) to turn the integral projection in Eq. 5 into a Fourier-domain slice. Substituting (F 2 S2 4 (B T α / B T α ) F 4 ) for (I2 4 B α), and noting that B T α =1/α 2, we arrive at the following result: P α 1 α 2 F F 2 S Bα T B T α ) F4 1 F F 2 S B T α F 4 (6) namely that a lens-formed photograph is obtained from the 4D Fourier spectrum of the light field by: extracting an appropriate 2D slice (S2 4 B T α ), applying an inverse 2D transform (F 2 ), and scaling the resulting image (1/F 2 ). Before stating the final theorem, let us define one last operator that combines all the action of photographic imaging in the Fourier domain: Fourier Photography Operator P α 1 F 2 S4 2 B T α. (7) It is easy to verify that P α has the following explicit form, directly from the definitions of S2 4 and B α. This explicit form is required for calculations: P α[g](k x,k y) (8) = 1 G(α kx, α ky, (1 α) kx, (1 α) ky). F 2 Applying Eq. 7 to Eq. 6 brings us, finally, to our goal: THEOREM (FOURIER SLICE PHOTOGRAPHY). P α F 2 P α F 4. (9) A photograph is the inverse 2D Fourier transform of a dilated 2D slice in the 4D Fourier transform of the light field. Figure 5 illustrates the relationships implied by this theorem. From an intellectual standpoint, the value of the theorem lies in the fact that P α, a slicing operator, is conceptually simpler than P α, an integral operator. This point is made especially clear by reviewing the explicit definitions of P α (Eq. 8) and P α (Eq. 3). By providing a frequency-based interpretation, the theorem contributes insight by providing two equivalent but very different perspectives on the physics of image formation. In this regard, the Fourier Slice Photography Theorem is not unlike the Fourier Convolution Theorem, which provides equivalent but very different perspectives of convolution in the two domains. From a practical standpoint, the theorem provides a faster computational pathway for certain kinds of light field processing. The computational complexities for each transform are illustrated in Figure 5, but the main point is that slicing via P α (O(n 2 )) is asymptotically faster than integration via P α (O(n 4 )). This fact is the basis for the algorithm in Section 6. 5 Theoretical Limits of Digital Refocusing The overarching goal of this section is to demonstrate the theoretical utility of the Fourier Slice Photography Theorem. Section 5.1 presents a general signal-processing theorem, showing exactly what happens to photographs when a light field is distorted by a convolution filter. Section 5.2 applies this theorem to analyze the performance of a band-limited light field camera. In these derivations, we will often use the Fourier Slice Photography Theorem to move the analysis into the frequency domain, where it becomes simpler. Figure 6: Filtered Light Field Photograph Theorem (Eq. 10). L F is the input 4D light field, and L F is a 4D filtering of it with 4D kernel k. E α F and E α F are the best photographs formed from the two light fields, where the photographs are focused with focal plane depth (α F ). The theorem shows that E α F is a 2D filtering of E α F, where the 2D kernel is the photograph of the 4D kernel, k. 5.1 Photographic Effect of Filtering the Light Field A light field produces exact photographs focused at various depths via Eq. 3. If we distort the light field by filtering it, and then form photographs from the distorted light field, how are these photographs related to the original, exact photographs? The following theorem provides the answer to this question. THEOREM (FILTERED LIGHT FIELD PHOTOGRAPHY). A4D convolution of a light field results in a 2D convolution of each photograph. The 2D filter kernel is simply the photograph of the 4D filter kernel focused at the same depth. Compactly in terms of operators, P α Ck 4 C 2 P α[k] P α, (10) where we have expressed convolution with the following operator: Convolution Ck N is an N-dimensional convolution operator with filter kernel k (an N-dimensional function), such that Ck N [F ](x) = F (x u) k(u) du where x and u are N- dimensional vector coordinates and F is an N-dimensional function. Figure 6 illustrates the theorem diagramatically. It is worth noting that in spite of its plausibility, the theorem is not obvious, and proving it in the spatial domain is quite difficult. Appendix B presents a proof of the theorem in the frequency-domain. At a high level, the approach is to apply the Fourier Slice Photography Theorem and the Convolution Theorem to move the analysis into the frequency domain. In that domain, photograph formation turns into a simpler slicing operator, and convolution turns into a simpler multiplication operation. This theorem is useful because it is simple and general. The next section contains a concrete example of how to use the theorem, but it should be emphasized that the theorem is much more broadly applicable. It will be worth exploiting it in general analysis of light field acquisition, where the system impulse response is the filter kernel, k(x, y, u, v), and light field processing, where the resampling strategy defines k(x, y, u, v). 5.2 Band-Limited Plenoptic Camera This section analyzes digital refocusing from a plenoptic camera, to answer the following questions. What is the quality of the photographs refocused from the acquired light fields? How are these photographs related to the exact photographs, such as those that might be taken by a conventional camera that were optically focused at the same depth?
5 The central assumption here, from which we will derive significant analytical leverage, is that the plenoptic camera captures bandlimited light fields. While perfect band-limiting is physically impossible, it is a plausible approximation in this case because the camera system blurs the incoming signal through imperfections in its optical elements, through area integration over the physical extent of microlenses and photosensor pixels, and ultimately through diffraction. The band-limited assumption means that the acquired light field, ˆL FL, is a simply the exact light field, L FL, convolved by a perfect low-pass filter, a 4D sinc: ˆL FL = Clowpass 4 [L FL ], where (11) lowpass(k x,k y,k u,k v)= 1/( x u) 2 sinc(k x/ x, k y/ x, k u/ u, k v/ u). (12) In this equation, x and u are the linear spatial and directional sampling rates of the integrated light field camera, respectively. The 1/( x u) 2 is an energy-normalizing constant to account for dilation of the sinc. Also note that, for compactness, we use multi-dimensional notation so that sinc(x, y, u, v) = sinc(x)sinc(y)sinc(u)sinc(v) Analytic Form for Refocused Photographs Our goal is an analytic solution for the digitally refocused photograph, Ê F, computed from the band-limited light field, L FL. This is where we apply the Filtered Light Field Photography Theorem. Letting, α = F/F L, [ ] [ Ê F = P α ˆLFL = P α C 4 lowpass [L FL ] ] = C 2 P α[lowpass] [P α [L FL ]] = C 2 P α[lowpass] [E F ], (13) where E F is the exact photograph at depth F. This derivation shows that the digitally refocused photograph is a 2D-filtered version of the exact photograph. The 2D kernel is simply a photograph of the 4D sinc function interpreted as a light field, P α [lowpass]. It turns out that photographs of a 4D sinc light field are simply 2D sinc functions: P α [lowpass] [ = P α 1/( x u)2 sinc(k x/ x, k y/ x, k u/ u, k ] v/ u) =1/Dx 2 sinc(k x/d x,k y/d x), (14) where the Nyquist rate of the 2D sinc depends on the amount of refocusing, α: D x = max(α x, 1 α u). (15) This fact is difficult to derive in the spatial domain, but applying the Fourier Slice Photography Theorem moves the analysis into the frequency domain, where it is easy (see Appendix C). The end result here is that, since the 2D kernel is a sinc, the band-limited camera produces digitally refocused photographs that are just band-limited versions of the exact photographs. The performance of digital refocusing is defined by the variation of the 2D kernel band-width (Eq. 15) with the extent of refocusing Interpretation of Refocusing Performance Notation Recall that the spatial and directional sampling rates of the camera are x and u. Let us further define the width of the camera sensor as W x, and the width of the lens aperture as W u. With these definitions, the spatial resolution of the sensors is N x = W x/ x and the directional resolution of the light field camera is N u = W u/ u. Exact Refocusing easy to verify that Since α =(F/F L) and u = W u/n u,itis α x (1 α) u F F L x(n uf/w u). (16) The claim here is that this is the range of focal depths, F L, where we can achieve exact refocusing, i.e. compute a sharp rendering of the photograph focused at that depth. What we are interested in is the Nyquist-limited resolution of the photograph, which is the number of band-limited samples within the field of view. Precisely, by applying Eq. 16 to Eq. 15, we see that the bandwidth of the computed photograph is (α x). Next, the field of view is not simply the size of the light field sensor, W x, but rather (αw x). This dilation is due to the fact that digital refocusing scales the image captured on the sensor by a factor of α in projecting it onto the refocus focal plane (see Eq. 3). If α>1, for example, the light field camera image is zoomed in slightly compared to the conventional camera. Figure 7 illustrates this effect. Thus, the Nyquist resolution of the computed photograph is (αw x)/(α x) =W x/ x. (17) This is simply the spatial resolution of the camera, the maximum possible resolution for the output photograph. This justifies the assertion that the refocusing is exact for the range of depths defined by Eq. 16. Note that this range of exact refocusing increases linearly with the directional resolution, N u. Inexact Refocusing If we exceed the exact refocusing range, i.e. F F L > x(n uf/w u). (18) then the band-limit of the computed photograph, Ê F, is 1 α u >α x (see Eq. 15), and the resulting resolution is not maximal, but rather (αw x)/( 1 α u), which is less than W x/ x. In other words, the resulting photograph is blurred, with reduced Nyquist-limited resolution. Re-writing this resolution in a slightly different form provides a more intuitive interpretation of the amount of blur. Since α = F/F L and u = W u/n u, the resolution is αw x 1 α u = W x W u/(n u F ) F F. (19) L Since ((N uf )/W u) is the f-number of a lens N u times smaller than the actual lens used on the camera, we can now interpret W u/(n u F ) F F L as the size of the conventional circle of confusion cast through this smaller lens when the film plane is misfocused by a distance of F F L. In other words, when refocusing beyond the exact range, we can only make the desired focal plane appear as sharp as it appears in a conventional photograph focused at the original depth, with a lens N u times smaller. Note that the sharpness increases linearly with the directional resolution, N u. 5.3 Summary It is worth summarizing the point of the analysis in Section 5. On a meta-level, this section has demonstrated the theoretical utility of the Fourier Slice Photography Theorem, applying it several times in deriving Eqs. 10, 13 and 14. At another level, this section has derived two end results that are of some importance. The first is the Filtered Light Field Photography Theorem, which is a simple but general signal-processing tool for analyzing light field imaging systems. The second is the fact that making a simple band-limited assumption about plenoptic cameras
6 Sensor Depth Conventional Plenoptic α =0.82 Inexact refocusing α =0.90 Exact refocusing α =1.0 α =1.11 Exact refocusing α =1.25 Inexact refocusing Figure 7: Photographs produced by an f/4 conventional and an f/4 plenoptic [Ng et al. 2005] camera, using digital refocusing (via Eq. 3) in the latter case. The sensor depth is given as a fraction of the film depth that brings the target into focus. Note that minor refocusing provides the plenoptic camera with a wider effective depth of focus than the conventional system. Also note how the field of view changes slightly with the sensor depth, a change due to divergence of the light rays from the lens aperture. yields an analytic proof that limits on digital refocusing improve linearly with directional resolution. Experiments with the plenoptic camera that we built achieved refocusing performance within a factor of 2 of this theory [Ng et al. 2005]. With N u =12, this enables sharp refocusing of f/4 photographs within the wide depth of field of an f/22 aperture. 6 Fourier Slice Digital Refocusing This section applies the Fourier Slice Photography Theorem in a very different way, to derive an asymptotically fast algorithm for digital refocusing. The presumed usage scenario is as follows: an in-camera light field is available (perhaps having been captured by a plenoptic camera). The user wishes to digitally refocus in an interactive manner, i.e. select a desired focal plane and view a synthetic photograph focused on that plane (see bottom row of Figure 8). In previous approaches to this problem [Isaksen et al. 2000; Levoy et al. 2004; Ng et al. 2005], spatial integration via Eq. 3 results in an O(n 4 ) algorithm, where n is the number of samples in each of the four dimensions. The algorithm described in this section provides a faster O(n 2 log n) algorithm, with the penalty of a single O(n 4 log n) pre-processing step. 6.1 Algorithm The algorithm follows trivially from the Fourier Slice Photography Theorem: Preprocess Prepare the given light field, L F, by pre-computing its 4D Fourier transform, F 4 [L], via the Fast Fourier Transform. This step takes O(n 4 log n) time. Refocusing For each choice of desired world focus plane, W, Compute the conjugate virtual film plane depth, F, via the thin lens equation: 1/F +1/W =1/f, where f is the focal length of the lens. Extract the dilated Fourier slice (via Eq. 8) of the preprocessed Fourier transform, to obtain (P α F 4 )[L], where α = F /F. This step takes O(n 2 ) time. Compute the inverse 2D Fourier transform of the slice, to obtain (F 2 P α F 4 )[L]. By the theorem, this final result is P α [L F ]=E F the photo focused on world plane W. This step takes O(n 2 log n) time. Figure 8 illustrates the steps of the algorithm. 6.2 Implementation and Results The complexity in implementing this simple algorithm has to do with ameliorating the artifacts that result from discretization, resampling and Fourier transformation. These artifacts are conceptually the same as the artifacts tackled in Fourier volume rendering [Levoy 1992; Malzbender 1993], and Fourier-based medical reconstruction techniques [Jackson et al. 1991] such as those used in CT and MR. The interested reader should consult these citations and their bibliographies for further details Sources of Artifacts In general signal-processing terms, when we sample a signal it is replicated periodically in the dual domain. When we reconstruct this sampled signal with convolution, it is multiplied in the dual domain by the Fourier transform of the convolution filter. The goal is to perfectly isolate the original, central replica, eliminating all other replicas. This means that the ideal filter is band-limited: it is of unit value for frequencies within the support of the light field, and zero for all other frequencies. Thus, the ideal filter is the sinc function, which has infinite extent. In practice we must use an imperfect, finite-extent filter, which will exhibit two important defects (see Figure 9). First, the filter will not be of unit value within the band-limit, gradually decaying to smaller fractional values as the frequency increases. Second, the filter will not be truly band-limited, containing energy at frequencies outside the desired stop-band. The first defect leads to so-called rolloff artifacts [Jackson et al. 1991], the most obvious manifestation of which is a darkening of the borders of computed photographs (see Figure 10). Decay in the filter s frequency spectrum with increasing frequency means that the spatial light field values, which are modulated by this spectrum, also roll off to fractional values towards the edges. The second defect, energy at frequencies above the band-limit, leads to aliasing artifacts (postaliasing, in the terminology of
7 Mitchell and Netravali [1998]) in computed photographs (see Figure 10). The non-zero energy beyond the band-limit means that the periodic replicas are not fully eliminated, leading to two kinds of aliasing. First, the replicas that appear parallel to the slicing plane appear as 2D replicas of the image encroaching on the borders of the final photograph. Second, the replicas positioned perpendicular to this plane are projected and summed onto the image plane, creating ghosting and loss of contrast Correcting Rolloff Error Rolloff error is a well understood effect in medical imaging and Fourier volume rendering. The standard solution is to multiply the affected signal by the reciprocal of the filter s inverse Fourier spectrum, to nullify the effect introduced during resampling. In our case, directly analogously to Fourier volume rendering [Malzbender 1993], the solution is to spatially pre-multiply the input light field by the reciprocal of the filter s 4D inverse Fourier transform (see Figure 11). This is performed prior to taking its 4D Fourier transform in the pre-processing step of the algorithm. Unfortunately, this pre-multiplication tends to accentuate the energy of the light field near its borders, maximizing the energy that folds back into the desired field of view as aliasing. 4 F Pα>1 Pα=1 Pα<1 2 F 2 F 2 F Figure 8: Fourier Slice Digital Refocusing algorithm. Row 1: 4D light field captured from a plenoptic camera, with four close-ups. Row 2: 4D Fourier transform of the light field, with four close-ups. Row 3: Three 2D slices of the 4D Fourier transform, extracted by Pα (Eq. 8) with different values of α. Row 4: Inverse 2D transforms of row 3. These images are conventional photographs focused on the closest, middle and furthest crayons. Row 5: Same images as row 4 except computed with brute-force integration in the spatial domain via Eq. 3, for comparison. Suppressing Aliasing Artifacts The three main methods of suppressing aliasing artifacts are oversampling, superior filtering and zero-padding. Oversampling within the extracted 2D Fourier slice (PF F 4 ) [L] increases the replication period in the spatial domain. This means that less energy in the tails of the in-plane replicas will fall within the borders of the final photograph. Exactly what happens computationally will be familiar to those experienced in discrete Fourier transforms. Specifically, increasing the sampling rate in one domain leads to an increase in the field of view in the other domain. Hence, by oversampling we produce an image that shows us more of the world than desired (see Figure 12), not a magnified view of the desired portion. Aliasing energy from neighboring replicas falls into these outer regions, which we crop away to isolate the central image of interest. Oversampling is appealing because of its simplicity, but oversampling alone cannot produce good quality images. The problem is that it cannot eliminate the replicas that appear perpendicular to the slicing plane, which are projected down onto the final image as described in the previous section. This brings us to the second major technique of combating aliasing: superior filtering. As already stated, the ideal filter is a sinc function with a band-limit matching the spatial bounds of the light field. Our goal is to use a finite-extent filter that approximates this perfect spectrum as closely as possible. The best methods for producing such filters use iterative techniques to jointly optimize the band-limit and narrow spatial support, as described in Jackson et al. [1991] in the medical imaging community, and Malzbender [1993] in the Fourier volume rendering community. Jackson et al. show that a much simpler, and near-optimal, approximation is the Kaiser-Bessel function. They also provide optimal Kaiser-Bessel parameter values for minimizing aliasing. Figure 13 illustrates the striking reduction in aliasing provided by such optimized Kaiser-Bessel filters compared to inferior quadrilinear interpolation. Surprisingly, a Kaiser-Bessel window of just width 2.5 suffices for excellent results. The third and final method to combat aliasing is to pad the light field with a small border of zero values before pre-multiplication and taking its Fourier transform [Levoy 1992; Malzbender 1993]. This pushes energy slightly further from the borders, and minimizes the amplification of aliasing energy by the pre-multiplication described in
8 Figure 9: Source of artifacts. Left: Triangle reconstruction filter. Right: Frequency spectrum of this filter (solid line) compared to ideal spectrum (dotted line). Shaded regions show deviations that lead to artifacts. Figure 13: Aliasing reduction by superior filtering. Rolloff correction is applied. Left: Quadrilinear filter (width 2). Middle: Kaiser-Bessel filter, width 1.5. Right: Kaiser-Bessel filter, width Implementation Summary We directly discretize the algorithm presented in 6.1, applying the following four techniques (from and 6.2.3) to suppress artifacts. In the pre-processing phase, Figure 10: Two main classes of artifacts. Left: Gold-standard image produced by spatial integration via Eq. 3. Middle: Rolloff artifacts using Kaiser-Bessel filter. Right: Aliasing artifacts using quadrilinear filter. Figure 11: Rolloff correction by pre-multiplying the input light field by the reciprocal of the resmpling filter s inverse Fourier transform. Left: Kaiser- Bessel filtering without pre-multiplication. Right: With pre-multiplication. 1. We pad the light field with a small border (5% of the width in that dimension) of zero values. 2. We pre-multiply the light field by the reciprocal of the Fourier transform of the resampling filter. In the refocusing step where we extract the 2D Fourier slice, 3. We use a linearly-separable Kaiser-Bessel resampling filter. width 2.5 produces excellent results. For fast previewing, an extremely narrow filter of width 1.5 produces results that are superior to (and faster than) quadrilinear interpolation. 4. We oversample the 2D Fourier slice by a factor of 2. After Fourier inversion, we crop the resulting photograph to isolate the central quadrant. The bottom two rows of Figure 8 compare this implementation of the Fourier Slice algorithm with spatial integration Performance Summary Figure 12: Aliasing reduction by oversampling. Quadrilinear filter used to emphasize aliasing effects for didactic purposes. Top left pair: 2D Fourier slice and inverse-transformed photograph, with unit sampling. Bottom: Same as top left pair, except with 2 oversampling in the frequency domain. Top right: Cropped version of bottom right. Note that aliasing is reduced compared to version with unit sampling. This section compares the performance of our algorithm compared to spatial-domain methods. Tests were performed on a 3.0 Ghz Pentium IV processor. The FFTW-3 library [Frigo and Johnson 1998] was used to compute Fourier transforms efficiently. For a light field with stresolution and uv resolution, spatial-domain integration achieved 1.68 fps (frames per second) using nearest-neighbor quadrature and 0.13 fps using quadrilinear interpolation. In contrast, the Fourier Slice method presented in this paper achieved 2.84 fps for previewing (Kaiser-Bessel filter, width 1.5 and no oversampling) and 0.54 fps for higher quality (width 2.5 and 2x oversampling). At these resolutions, performance between spatial-domain and Fourier-domain methods are comparable. The pre-processing time, however, was 47 seconds. The Fourier Slice method outperforms the spatial methods as the directional uv resolution increases, because the number of light field samples that must be summed increases for the spatial integration methods, but the cost of slicing stays constant per pixel. For a light field with st resolution and uv resolution, spatial-domain integration achieved 1.63 fps using nearestneighbor, and 0.10 fps using quadrilinear interpolation. The Fourier Slice method achieved fps for previewing and 2.73 fps for higher quality. At these resolutions, the Fourier slice methods are an order of magnitude faster. In this case, the pre-processing time was 30 seconds.
9 7 Conclusions and Future Work The main contribution of this paper is the Fourier Slice Photography Theorem. By describing how photographic imaging occurs in the Fourier domain, this simple but fundamental result provides a versatile tool for algorithm development and analysis of light field imaging systems. This paper tries to present concrete examples of this utility in analyzing a band-limited plenoptic camera, and in developing the Fourier Slice Digital Refocusing algorithm. The band-limited analysis provides theoretical backing for recent experimental results demonstrating exciting performance of digital refocusing from plenoptic camera data [Ng et al. 2005]. The derivations also yield a general-purpose tool (Eq. 10) for analyzing plenoptic camera systems. The Fourier Slice Digital Refocusing algorithm is asymptotically faster than previous approaches. It is also efficient in a practical sense, thanks to the optimized Kaiser-Bessel resampling strategy borrowed directly from the tomography literature. Continuing to exploit this connection with tomography will surely yield further benefits in light field processing. A clear line of future work would extend the algorithm to increase its focusing flexibility. Presented here in its simplest form, the algorithm is limited to full-aperture refocusing. Support for different apertures could be provided by appropriate convolution in the Fourier domain that results in a spatial multiplication masking out the undesired portion of the aperture. This is related to work in Fourier volume shading [Levoy 1992]. A very different class of future work might emerge from looking at the footprint of photographs in the 4D Fourier transform of the light field. It is a direct consequence of the Fourier Slice Photography Theorem (consider Eq. 9 for all α) that the footprint of all full-aperture photographs lies on the following 3D manifold in the 4D Fourier space: { (α kx, α k y, (1 α) k x, (1 α) k y) where α [0, ), and k x,k y R } (20) Two possible lines of research are as follows. First, it might be possible to optimize light field camera designs to provide greater fidelity on this manifold, at the expense of the vast remainder of the space that does not contribute to refocused photographs. One could also compress light fields for refocusing by storing only the data on the 3D manifold. Second, photographs focused at a particular depth will contain sharp details (hence high frequencies) only if an object exists at that depth. This observation suggests a simple Fourier Range Finding algorithm: search for regions of high spectral energy on the 3D manifold at a large distance from the origin. The rotation angle of these regions gives the depth (via the Fourier Slice Photography Theorem) of focal planes that intersect features in the visual world. Acknowledgments Thanks to Marc Levoy, Pat Hanrahan, Ravi Ramamoorthi, Mark Horowitz, Brian Curless and Kayvon Fatahalian for discussions, help with references and reading a draft. Marc Levoy took the photograph of the crayons. Thanks also to Dwight Nishimura and Brad Osgood. Finally, thanks to the anonymous reviewers for their helpful feedback and comments. This work was supported by NSF grant (Interacting with the Visual World), and by a Microsoft Research Fellowship. References ADELSON, T., AND WANG, J. Y. A Single lens stereo with a plenoptic camera. IEEE Transactions on Pattern Analysis and Machine Intelligence 14, 2 (Feb), BRACEWELL, R. N., CHANG, K.-Y., JHA, A. K., AND WANG, Y.H Affine theorem for two-dimensional fourier transform. Electronics Letters 29, BRACEWELL, R. N Strip integration in radio astronomy. Aust. J. Phys. 9, BRACEWELL, R. N The Fourier Transform and Its Applications, 2nd Edition Revised. WCB / McGraw-Hill. CHAI, J., TONG, X., AND SHUM, H Plenoptic sampling. In SIG- GRAPH 00, DEANS, S. R The Radon Transform and Some of Its Applications. Wiley-Interscience. FRIGO, M., AND JOHNSON, S. G FFTW: An adaptive software architecture for the FFT. In ICASSP conference proceedings, vol. 3, GORTLER, S. J., GRZESZCZUK, R., SZELISKI, R., AND COHEN, M.F The Lumigraph. In SIGGRAPH 96, ISAKSEN, A., MCMILLAN, L., AND GORTLER, S. J Dynamically reparameterized light fields. In SIGGRAPH 2000, IVES, H. E Parallax panoramagrams made with a large diameter lens. J. Opt. Soc. Amer. 20, JACKSON, J. I., MEYER, C. H., NISHIMURA, D. G., AND MACOVSKI, A Selection of convolution function for fourier inversion using gridding. IEEE Transactions on Medical Imaging 10, 3, JAVIDI, B., AND OKANO, F., Eds Three-Dimensional Television, Video and Display Technologies. Springer-Verlag. LEVOY, M., AND HANRAHAN, P Light field rendering. In SIG- GRAPH 96, LEVOY, M., CHEN, B., VAISH, V.,HOROWITZ, M., MCDOWALL, I., AND BOLAS, M Synthetic aperture confocal imaging. ACM Transactions on Graphics 23, 3, LEVOY, M Volume rendering using the fourier projection-slice theorem. In Proceedings of Graphics Interface 92, LIANG, Z.-P., AND MUNSON, D. C Partial Radon transforms. IEEE Transactions on Image Processing 6, 10 (Oct 1997), LIPPMANN, G La photographie intégrale. Comptes-Rendus, Académie des Sciences 146, MACOVSKI, A Medical Imaging Systems. Prentice Hall. MALZBENDER, T Fourier volume rendering. ACM Transactions on Graphics 12, 3, MITCHELL, D. P., AND NETRAVALI, A. N Reconstruction filters in computer graphics. In SIGGRAPH 98, NAEMURA, T., YOSHIDA, T., AND HARASHIMA, H D computer graphics based on integral photography. Optics Express 8, 2, NG, R., LEVOY, M., BRÉDIF, M., DUVAL, G., HOROWITZ, M., AND HANRAHAN, P Light field photography with a hand-held plenoptic camera. Tech. Rep. CSTR , Stanford Computer Science. OKANO, F., ARAI, J., HOSHINO, H., AND YUYAMA, I Threedimensional video system based on integral photography. Optical Engineering 38, 6 (June 1999), OKOSHI, T Three-Dimensional Imaging Techniques. Acad. Press. STEWART, J., YU, J., GORTLER, S. J., AND MCMILLAN, L A new reconstruction filter for undersampled light fields. In Proceedings of the Eurographics Symposium on Rendering 2003, STROEBEL, L., COMPTON, J., CURRENT, I., AND ZAKIA, R Photographic Materials and Processes. Focal Press. TYSON, R. K Principles of Adaptive Optics. Academic Press. VAISH, V.,WILBURN, B., JOSHI, N., AND LEVOY, M Using plane + parallax for calibrating dense camera arrays. In Proc. of CVPR, 2 9. WILBURN, B., JOSHI, N., VAISH, V.,TALVALA, E.-V., ANTUNEZ, E., BARTH, A., ADAMS, A., LEVOY, M., AND HOROWITZ, M High performance imaging using large camera arrays. To appear at SIG- GRAPH 2005.
10 Appendices: Proofs and Derivations A Generalized Fourier Slice Theorem Theorem (Generalized Fourier Slice). ( ) F M IM N B SM N B T / B T F N Proof. The following proof is inspired by one common approach to proving the classical 2D version of the theorem. The first step is to note that F M IM N = SM N F N, (21) because substitution of the basic definitions shows that for an arbitrary function, f, both (F M IM N )[f](u 1,...,u M ) and (SM N F N )[f](u 1,...,u M ) are equal to f(x 1,...,x N )exp( 2πi (x 1 u x M u M )) dx 1...dx N. The next step is to observe that if basis change operators commute with Fourier transforms via F N B (B T / B T ) F N, then the proof of the theorem would be complete because for every function f we would have (F M IM N B)[f] =(SM N F N )[B [f]] (by Eq. 21) ( ( ) = SM N B T ) / B T F N [f]. (commute) (22) Thus, the final step is to show that F N B (B T / B T ) F N. Directly substituting the operator definitions establishes these two equations: ( ( (B F N )[f](u) = f(x)exp 2πi x T ( B 1 u ))) dx; (23) ) (F N B T [f](u) (24) B T = 1 B T ( f B T x ) exp ( 2πi ( x u )) dx. In these equations, x and u are N-dimensional column vectors, and the integral is taken over all of N-dimensional space. Let us now apply the change of variables x = B T x to Eq. 24, noting that x = B T x, and dx = ( 1/ B T ) dx. Making these substitutions, (F N B T B T ) [f](u) = = ( ( ( ) )) T f (x)exp 2πi B T x u dx ( ( )) f (x)exp 2πi x T B 1 u dx (25) where the last line relies on the linear algebra rule for transposing matrix products. Equations 23 and 25 show that F N B (B T / B T ) F N, completing the proof. B Filtered Light Field Photography Thm. Theorem (Filtered Light Field Photography). P F C 4 k C 2 P F [k] P F, To prove the theorem, let us first establish a lemma involving the closely-related modulation operator: Modulation M N β is an N-dimensional modulation operator, such that M N β [F ](x) =F (x) β(x) where x is an N-dimensional vector coordinate. Lemma. Multiplying an input 4D function by another one, k, and transforming the result by P F, the Fourier photography operator, is equivalent to transforming both functions by P F and then multiplying the resulting 2D functions. In operators, P F M 4 k M 2 P F [k] P F (26) Algebraic verification of the lemma is direct given the basic definitions, and is omitted here. On an intuitive level, however, the lemma makes sense because P α is a slicing operator: multiplying two functions and then slicing them is the same as slicing each of them and multiplying the resulting functions. Proof of theorem. The first step is to translate the classical Fourier Convolution Theorem (see, for example, Bracewell [1986]) into useful operator identities. The Convolution Theorem states that a multiplication in the spatial domain is equivalent to convolution in the Fourier domain, and vice versa. As a result, F N Ck N M N F N [k] FN (27) and F N M N k C N F N [k] FN. (28) Note that these equations also hold for negative N, since the Convolution Theorem also applies to the inverse Fourier transform. With these facts and the lemma in hand, the proof of the theorem proceeds swiftly: P F Ck 4 F 2 P F F 4 Ck 4 F 2 P F M 4 F 4 [k] F4 F 2 M 2 (P F F 4 )[k] P F F 4 C 2 (F 2 P F F 4 )[k] F 2 P F F 4 C 2 P F [k] P F, where we apply the Fourier Slice Photography Theorem (Eq. 9) to derive the first and last lines, the Convolution Theorem (Eqs. 27 and 28) for the second and fourth lines, and the lemma (Eq. 26) for the third line. C Photograph of a 4D Sinc Light Field This appendix derives Eq. 14, which states that a photo from a 4D sinc light field is a 2D sinc function. The first step is to apply the Fourer Slice Photography Theorem to move the derivation into the Fourier domain. [ P α 1/( x u)2 sinc(x/ x, y/ x, u/ u, v/ u) ] =1/( x u) 2 (F 2 P α F 4 )[sinc(x/ x, y/ x, u/ u, v/ u)] =(F 2 P α)[ (k x x, k y x, k u u, k v u)]. (29) Now we apply the definition for the Fourier photography operator P α (Eq. 7), to arrive at [ P α 1/( x u)2 sinc(x/ x, y/ x, u/ u, v/ u) ] (30) = F 2 [ (αk x x, αk y x, (1 α)k x u, (1 α)k y u)]. Note that the 4D rect function now depends only on k x and k y, not k u or k v. Since the product of two dilated rect functions is equal to the smaller rect function, [ P α 1/( x u)2 sinc(x/ x, y/ x, u/ u, v/ u) ] = F 2 [ (k xd x,k yd x)] (31) where D x = max(α x, 1 α u). (32) Applying the inverse 2D Fourier transform completes the proof: [ P α 1/( x u)2 sinc(x/ x, y/ x, u/ u, v/ u) ] =1/Dx 2 sinc(k x/d x,k y/d x). (33)
Aliasing, Image Sampling and Reconstruction
Aliasing, Image Sampling and Reconstruction Recall: a pixel is a point It is NOT a box, disc or teeny wee light It has no dimension It occupies no area It can have a coordinate More than a point, it is
Lecture 14. Point Spread Function (PSF)
Lecture 14 Point Spread Function (PSF), Modulation Transfer Function (MTF), Signal-to-noise Ratio (SNR), Contrast-to-noise Ratio (CNR), and Receiver Operating Curves (ROC) Point Spread Function (PSF) Recollect
Metric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
Theory and Methods of Lightfield Photography SIGGRAPH 2009
Theory and Methods of Lightfield Photography SIGGRAPH 2009 Todor Georgiev Adobe Systems [email protected] Andrew Lumsdaine Indiana University [email protected] 1 Web Page http://www.tgeorgiev.net/asia2009/
Linear Algebra Notes for Marsden and Tromba Vector Calculus
Linear Algebra Notes for Marsden and Tromba Vector Calculus n-dimensional Euclidean Space and Matrices Definition of n space As was learned in Math b, a point in Euclidean three space can be thought of
The continuous and discrete Fourier transforms
FYSA21 Mathematical Tools in Science The continuous and discrete Fourier transforms Lennart Lindegren Lund Observatory (Department of Astronomy, Lund University) 1 The continuous Fourier transform 1.1
Continued Fractions and the Euclidean Algorithm
Continued Fractions and the Euclidean Algorithm Lecture notes prepared for MATH 326, Spring 997 Department of Mathematics and Statistics University at Albany William F Hammond Table of Contents Introduction
Algebra 2 Chapter 1 Vocabulary. identity - A statement that equates two equivalent expressions.
Chapter 1 Vocabulary identity - A statement that equates two equivalent expressions. verbal model- A word equation that represents a real-life problem. algebraic expression - An expression with variables.
Light Field Photography with a Hand-held Plenoptic Camera
Light Field Photography with a Hand-held Plenoptic Camera Ren Ng Marc Levoy Mathieu Brédif Gene Duval Mark Horowitz Pat Hanrahan Stanford University Duval Design Abstract This paper presents a camera that
Correlation and Convolution Class Notes for CMSC 426, Fall 2005 David Jacobs
Correlation and Convolution Class otes for CMSC 46, Fall 5 David Jacobs Introduction Correlation and Convolution are basic operations that we will perform to extract information from images. They are in
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
Solving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
Section 4.4 Inner Product Spaces
Section 4.4 Inner Product Spaces In our discussion of vector spaces the specific nature of F as a field, other than the fact that it is a field, has played virtually no role. In this section we no longer
Section 1.1. Introduction to R n
The Calculus of Functions of Several Variables Section. Introduction to R n Calculus is the study of functional relationships and how related quantities change with each other. In your first exposure to
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
Linear Algebra: Vectors
A Linear Algebra: Vectors A Appendix A: LINEAR ALGEBRA: VECTORS TABLE OF CONTENTS Page A Motivation A 3 A2 Vectors A 3 A2 Notational Conventions A 4 A22 Visualization A 5 A23 Special Vectors A 5 A3 Vector
Big Ideas in Mathematics
Big Ideas in Mathematics which are important to all mathematics learning. (Adapted from the NCTM Curriculum Focal Points, 2006) The Mathematics Big Ideas are organized using the PA Mathematics Standards
Lecture L3 - Vectors, Matrices and Coordinate Transformations
S. Widnall 16.07 Dynamics Fall 2009 Lecture notes based on J. Peraire Version 2.0 Lecture L3 - Vectors, Matrices and Coordinate Transformations By using vectors and defining appropriate operations between
MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform
MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish
Understanding astigmatism Spring 2003
MAS450/854 Understanding astigmatism Spring 2003 March 9th 2003 Introduction Spherical lens with no astigmatism Crossed cylindrical lenses with astigmatism Horizontal focus Vertical focus Plane of sharpest
Current Standard: Mathematical Concepts and Applications Shape, Space, and Measurement- Primary
Shape, Space, and Measurement- Primary A student shall apply concepts of shape, space, and measurement to solve problems involving two- and three-dimensional shapes by demonstrating an understanding of:
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation
2. SPATIAL TRANSFORMATIONS
Digitales Video 1 2. SPATIAL TRANSFORMATIONS This chapter describes common spatial transformations derived for digital image warping applications in computer vision and computer graphics. A spatial transformation
THREE DIMENSIONAL GEOMETRY
Chapter 8 THREE DIMENSIONAL GEOMETRY 8.1 Introduction In this chapter we present a vector algebra approach to three dimensional geometry. The aim is to present standard properties of lines and planes,
Unified Lecture # 4 Vectors
Fall 2005 Unified Lecture # 4 Vectors These notes were written by J. Peraire as a review of vectors for Dynamics 16.07. They have been adapted for Unified Engineering by R. Radovitzky. References [1] Feynmann,
Factoring Patterns in the Gaussian Plane
Factoring Patterns in the Gaussian Plane Steve Phelps Introduction This paper describes discoveries made at the Park City Mathematics Institute, 00, as well as some proofs. Before the summer I understood
Lecture 12: Cameras and Geometry. CAP 5415 Fall 2010
Lecture 12: Cameras and Geometry CAP 5415 Fall 2010 The midterm What does the response of a derivative filter tell me about whether there is an edge or not? Things aren't working Did you look at the filters?
1 VECTOR SPACES AND SUBSPACES
1 VECTOR SPACES AND SUBSPACES What is a vector? Many are familiar with the concept of a vector as: Something which has magnitude and direction. an ordered pair or triple. a description for quantities such
NOTES ON LINEAR TRANSFORMATIONS
NOTES ON LINEAR TRANSFORMATIONS Definition 1. Let V and W be vector spaces. A function T : V W is a linear transformation from V to W if the following two properties hold. i T v + v = T v + T v for all
By choosing to view this document, you agree to all provisions of the copyright laws protecting it.
This material is posted here with permission of the IEEE Such permission of the IEEE does not in any way imply IEEE endorsement of any of Helsinki University of Technology's products or services Internal
SIGNAL PROCESSING & SIMULATION NEWSLETTER
1 of 10 1/25/2008 3:38 AM SIGNAL PROCESSING & SIMULATION NEWSLETTER Note: This is not a particularly interesting topic for anyone other than those who ar e involved in simulation. So if you have difficulty
The Basics of FEA Procedure
CHAPTER 2 The Basics of FEA Procedure 2.1 Introduction This chapter discusses the spring element, especially for the purpose of introducing various concepts involved in use of the FEA technique. A spring
South Carolina College- and Career-Ready (SCCCR) Pre-Calculus
South Carolina College- and Career-Ready (SCCCR) Pre-Calculus Key Concepts Arithmetic with Polynomials and Rational Expressions PC.AAPR.2 PC.AAPR.3 PC.AAPR.4 PC.AAPR.5 PC.AAPR.6 PC.AAPR.7 Standards Know
2x + y = 3. Since the second equation is precisely the same as the first equation, it is enough to find x and y satisfying the system
1. Systems of linear equations We are interested in the solutions to systems of linear equations. A linear equation is of the form 3x 5y + 2z + w = 3. The key thing is that we don t multiply the variables
3. INNER PRODUCT SPACES
. INNER PRODUCT SPACES.. Definition So far we have studied abstract vector spaces. These are a generalisation of the geometric spaces R and R. But these have more structure than just that of a vector space.
Algebra 1 2008. Academic Content Standards Grade Eight and Grade Nine Ohio. Grade Eight. Number, Number Sense and Operations Standard
Academic Content Standards Grade Eight and Grade Nine Ohio Algebra 1 2008 Grade Eight STANDARDS Number, Number Sense and Operations Standard Number and Number Systems 1. Use scientific notation to express
EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines
EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation
INTRODUCTION TO RENDERING TECHNIQUES
INTRODUCTION TO RENDERING TECHNIQUES 22 Mar. 212 Yanir Kleiman What is 3D Graphics? Why 3D? Draw one frame at a time Model only once X 24 frames per second Color / texture only once 15, frames for a feature
Linear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
The Fourier Analysis Tool in Microsoft Excel
The Fourier Analysis Tool in Microsoft Excel Douglas A. Kerr Issue March 4, 2009 ABSTRACT AD ITRODUCTIO The spreadsheet application Microsoft Excel includes a tool that will calculate the discrete Fourier
Numerical Analysis Lecture Notes
Numerical Analysis Lecture Notes Peter J. Olver 5. Inner Products and Norms The norm of a vector is a measure of its size. Besides the familiar Euclidean norm based on the dot product, there are a number
1 Sets and Set Notation.
LINEAR ALGEBRA MATH 27.6 SPRING 23 (COHEN) LECTURE NOTES Sets and Set Notation. Definition (Naive Definition of a Set). A set is any collection of objects, called the elements of that set. We will most
PYTHAGOREAN TRIPLES KEITH CONRAD
PYTHAGOREAN TRIPLES KEITH CONRAD 1. Introduction A Pythagorean triple is a triple of positive integers (a, b, c) where a + b = c. Examples include (3, 4, 5), (5, 1, 13), and (8, 15, 17). Below is an ancient
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
LINEAR ALGEBRA W W L CHEN
LINEAR ALGEBRA W W L CHEN c W W L Chen, 1997, 2008 This chapter is available free to all individuals, on understanding that it is not to be used for financial gain, and may be downloaded and/or photocopied,
FFT Algorithms. Chapter 6. Contents 6.1
Chapter 6 FFT Algorithms Contents Efficient computation of the DFT............................................ 6.2 Applications of FFT................................................... 6.6 Computing DFT
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS Systems of Equations and Matrices Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
2.2 Creaseness operator
2.2. Creaseness operator 31 2.2 Creaseness operator Antonio López, a member of our group, has studied for his PhD dissertation the differential operators described in this section [72]. He has compared
The Matrix Elements of a 3 3 Orthogonal Matrix Revisited
Physics 116A Winter 2011 The Matrix Elements of a 3 3 Orthogonal Matrix Revisited 1. Introduction In a class handout entitled, Three-Dimensional Proper and Improper Rotation Matrices, I provided a derivation
Tennessee Mathematics Standards 2009-2010 Implementation. Grade Six Mathematics. Standard 1 Mathematical Processes
Tennessee Mathematics Standards 2009-2010 Implementation Grade Six Mathematics Standard 1 Mathematical Processes GLE 0606.1.1 Use mathematical language, symbols, and definitions while developing mathematical
LEARNING OBJECTIVES FOR THIS CHAPTER
CHAPTER 2 American mathematician Paul Halmos (1916 2006), who in 1942 published the first modern linear algebra book. The title of Halmos s book was the same as the title of this chapter. Finite-Dimensional
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
Solutions to Math 51 First Exam January 29, 2015
Solutions to Math 5 First Exam January 29, 25. ( points) (a) Complete the following sentence: A set of vectors {v,..., v k } is defined to be linearly dependent if (2 points) there exist c,... c k R, not
3 An Illustrative Example
Objectives An Illustrative Example Objectives - Theory and Examples -2 Problem Statement -2 Perceptron - Two-Input Case -4 Pattern Recognition Example -5 Hamming Network -8 Feedforward Layer -8 Recurrent
Notes on Determinant
ENGG2012B Advanced Engineering Mathematics Notes on Determinant Lecturer: Kenneth Shum Lecture 9-18/02/2013 The determinant of a system of linear equations determines whether the solution is unique, without
4.5 Linear Dependence and Linear Independence
4.5 Linear Dependence and Linear Independence 267 32. {v 1, v 2 }, where v 1, v 2 are collinear vectors in R 3. 33. Prove that if S and S are subsets of a vector space V such that S is a subset of S, then
Introduction to Medical Imaging. Lecture 11: Cone-Beam CT Theory. Introduction. Available cone-beam reconstruction methods: Our discussion:
Introduction Introduction to Medical Imaging Lecture 11: Cone-Beam CT Theory Klaus Mueller Available cone-beam reconstruction methods: exact approximate algebraic Our discussion: exact (now) approximate
Rotation Matrices and Homogeneous Transformations
Rotation Matrices and Homogeneous Transformations A coordinate frame in an n-dimensional space is defined by n mutually orthogonal unit vectors. In particular, for a two-dimensional (2D) space, i.e., n
α = u v. In other words, Orthogonal Projection
Orthogonal Projection Given any nonzero vector v, it is possible to decompose an arbitrary vector u into a component that points in the direction of v and one that points in a direction orthogonal to v
Video Camera Image Quality in Physical Electronic Security Systems
Video Camera Image Quality in Physical Electronic Security Systems Video Camera Image Quality in Physical Electronic Security Systems In the second decade of the 21st century, annual revenue for the global
Common Core Unit Summary Grades 6 to 8
Common Core Unit Summary Grades 6 to 8 Grade 8: Unit 1: Congruence and Similarity- 8G1-8G5 rotations reflections and translations,( RRT=congruence) understand congruence of 2 d figures after RRT Dilations
Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.
Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).
Reflection and Refraction
Equipment Reflection and Refraction Acrylic block set, plane-concave-convex universal mirror, cork board, cork board stand, pins, flashlight, protractor, ruler, mirror worksheet, rectangular block worksheet,
WHEN DOES A CROSS PRODUCT ON R n EXIST?
WHEN DOES A CROSS PRODUCT ON R n EXIST? PETER F. MCLOUGHLIN It is probably safe to say that just about everyone reading this article is familiar with the cross product and the dot product. However, what
Geometric Optics Converging Lenses and Mirrors Physics Lab IV
Objective Geometric Optics Converging Lenses and Mirrors Physics Lab IV In this set of lab exercises, the basic properties geometric optics concerning converging lenses and mirrors will be explored. The
DICOM Correction Item
Correction Number DICOM Correction Item CP-626 Log Summary: Type of Modification Clarification Rationale for Correction Name of Standard PS 3.3 2004 + Sup 83 The description of pixel spacing related attributes
NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS
NEW YORK STATE TEACHER CERTIFICATION EXAMINATIONS TEST DESIGN AND FRAMEWORK September 2014 Authorized for Distribution by the New York State Education Department This test design and framework document
Figure 1.1 Vector A and Vector F
CHAPTER I VECTOR QUANTITIES Quantities are anything which can be measured, and stated with number. Quantities in physics are divided into two types; scalar and vector quantities. Scalar quantities have
Chapter 111. Texas Essential Knowledge and Skills for Mathematics. Subchapter B. Middle School
Middle School 111.B. Chapter 111. Texas Essential Knowledge and Skills for Mathematics Subchapter B. Middle School Statutory Authority: The provisions of this Subchapter B issued under the Texas Education
Settling a Question about Pythagorean Triples
Settling a Question about Pythagorean Triples TOM VERHOEFF Department of Mathematics and Computing Science Eindhoven University of Technology P.O. Box 513, 5600 MB Eindhoven, The Netherlands E-Mail address:
Geometric Transformations
Geometric Transformations Definitions Def: f is a mapping (function) of a set A into a set B if for every element a of A there exists a unique element b of B that is paired with a; this pairing is denoted
IMPROVEMENT OF DIGITAL IMAGE RESOLUTION BY OVERSAMPLING
ABSTRACT: IMPROVEMENT OF DIGITAL IMAGE RESOLUTION BY OVERSAMPLING Hakan Wiman Department of Photogrammetry, Royal Institute of Technology S - 100 44 Stockholm, Sweden (e-mail [email protected]) ISPRS Commission
Investigation of Color Aliasing of High Spatial Frequencies and Edges for Bayer-Pattern Sensors and Foveon X3 Direct Image Sensors
Investigation of Color Aliasing of High Spatial Frequencies and Edges for Bayer-Pattern Sensors and Foveon X3 Direct Image Sensors Rudolph J. Guttosch Foveon, Inc. Santa Clara, CA Abstract The reproduction
Chapter 17. Orthogonal Matrices and Symmetries of Space
Chapter 17. Orthogonal Matrices and Symmetries of Space Take a random matrix, say 1 3 A = 4 5 6, 7 8 9 and compare the lengths of e 1 and Ae 1. The vector e 1 has length 1, while Ae 1 = (1, 4, 7) has length
Chapter 12 Modal Decomposition of State-Space Models 12.1 Introduction The solutions obtained in previous chapters, whether in time domain or transfor
Lectures on Dynamic Systems and Control Mohammed Dahleh Munther A. Dahleh George Verghese Department of Electrical Engineering and Computer Science Massachuasetts Institute of Technology 1 1 c Chapter
Algebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 2012-13 school year.
This document is designed to help North Carolina educators teach the Common Core (Standard Course of Study). NCDPI staff are continually updating and improving these tools to better serve teachers. Algebra
9 Multiplication of Vectors: The Scalar or Dot Product
Arkansas Tech University MATH 934: Calculus III Dr. Marcel B Finan 9 Multiplication of Vectors: The Scalar or Dot Product Up to this point we have defined what vectors are and discussed basic notation
ECE 842 Report Implementation of Elliptic Curve Cryptography
ECE 842 Report Implementation of Elliptic Curve Cryptography Wei-Yang Lin December 15, 2004 Abstract The aim of this report is to illustrate the issues in implementing a practical elliptic curve cryptographic
Projective Geometry: A Short Introduction. Lecture Notes Edmond Boyer
Projective Geometry: A Short Introduction Lecture Notes Edmond Boyer Contents 1 Introduction 2 11 Objective 2 12 Historical Background 3 13 Bibliography 4 2 Projective Spaces 5 21 Definitions 5 22 Properties
This unit will lay the groundwork for later units where the students will extend this knowledge to quadratic and exponential functions.
Algebra I Overview View unit yearlong overview here Many of the concepts presented in Algebra I are progressions of concepts that were introduced in grades 6 through 8. The content presented in this course
Notes on Orthogonal and Symmetric Matrices MENU, Winter 2013
Notes on Orthogonal and Symmetric Matrices MENU, Winter 201 These notes summarize the main properties and uses of orthogonal and symmetric matrices. We covered quite a bit of material regarding these topics,
Rodenstock Photo Optics
Rogonar Rogonar-S Rodagon Apo-Rodagon N Rodagon-WA Apo-Rodagon-D Accessories: Modular-Focus Lenses for Enlarging, CCD Photos and Video To reproduce analog photographs as pictures on paper requires two
Systems of Linear Equations
Systems of Linear Equations Beifang Chen Systems of linear equations Linear systems A linear equation in variables x, x,, x n is an equation of the form a x + a x + + a n x n = b, where a, a,, a n and
9.4. The Scalar Product. Introduction. Prerequisites. Learning Style. Learning Outcomes
The Scalar Product 9.4 Introduction There are two kinds of multiplication involving vectors. The first is known as the scalar product or dot product. This is so-called because when the scalar product of
CHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
Georgia Standards of Excellence Curriculum Map. Mathematics. GSE 8 th Grade
Georgia Standards of Excellence Curriculum Map Mathematics GSE 8 th Grade These materials are for nonprofit educational purposes only. Any other use may constitute copyright infringement. GSE Eighth Grade
In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data.
MATHEMATICS: THE LEVEL DESCRIPTIONS In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data. Attainment target
Exposing Digital Forgeries Through Chromatic Aberration
Exposing Digital Forgeries Through Chromatic Aberration Micah K. Johnson Department of Computer Science Dartmouth College Hanover, NH 03755 [email protected] Hany Farid Department of Computer Science
CHAPTER 3: DIGITAL IMAGING IN DIAGNOSTIC RADIOLOGY. 3.1 Basic Concepts of Digital Imaging
Physics of Medical X-Ray Imaging (1) Chapter 3 CHAPTER 3: DIGITAL IMAGING IN DIAGNOSTIC RADIOLOGY 3.1 Basic Concepts of Digital Imaging Unlike conventional radiography that generates images on film through
For example, estimate the population of the United States as 3 times 10⁸ and the
CCSS: Mathematics The Number System CCSS: Grade 8 8.NS.A. Know that there are numbers that are not rational, and approximate them by rational numbers. 8.NS.A.1. Understand informally that every number
Operation Count; Numerical Linear Algebra
10 Operation Count; Numerical Linear Algebra 10.1 Introduction Many computations are limited simply by the sheer number of required additions, multiplications, or function evaluations. If floating-point
with functions, expressions and equations which follow in units 3 and 4.
Grade 8 Overview View unit yearlong overview here The unit design was created in line with the areas of focus for grade 8 Mathematics as identified by the Common Core State Standards and the PARCC Model
Mechanics 1: Conservation of Energy and Momentum
Mechanics : Conservation of Energy and Momentum If a certain quantity associated with a system does not change in time. We say that it is conserved, and the system possesses a conservation law. Conservation
Similarity and Diagonalization. Similar Matrices
MATH022 Linear Algebra Brief lecture notes 48 Similarity and Diagonalization Similar Matrices Let A and B be n n matrices. We say that A is similar to B if there is an invertible n n matrix P such that
DERIVATIVES AS MATRICES; CHAIN RULE
DERIVATIVES AS MATRICES; CHAIN RULE 1. Derivatives of Real-valued Functions Let s first consider functions f : R 2 R. Recall that if the partial derivatives of f exist at the point (x 0, y 0 ), then we
MATH 132: CALCULUS II SYLLABUS
MATH 32: CALCULUS II SYLLABUS Prerequisites: Successful completion of Math 3 (or its equivalent elsewhere). Math 27 is normally not a sufficient prerequisite for Math 32. Required Text: Calculus: Early
WORK SCHEDULE: MATHEMATICS 2007
, K WORK SCHEDULE: MATHEMATICS 00 GRADE MODULE TERM... LO NUMBERS, OPERATIONS AND RELATIONSHIPS able to recognise, represent numbers and their relationships, and to count, estimate, calculate and check
7 Gaussian Elimination and LU Factorization
7 Gaussian Elimination and LU Factorization In this final section on matrix factorization methods for solving Ax = b we want to take a closer look at Gaussian elimination (probably the best known method
