Abstract
Orthogonality is exploited for fitting analytically-specified freeform shapes in terms of orthogonal polynomials. The end result is expressed in terms of FFTs coupled to a simple explicit form of Gaussian quadrature. Its efficiency opens the possibilities for proceeding to arbitrary numbers of polynomial terms. This is shown to create promising options for quantifying and filtering the mid-spatial frequency structure within circular domains from measurements of as-built parts.
©2013 Optical Society of America
1. Introduction
Freeform optics play established roles for progressive ophthalmic lenses, in conformal and non-imaging applications, and across a diversity of display and projection systems. Several conventions have been developed and implemented for describing their surface shapes in commercial optical design codes. Perhaps the most standard option is a simple polynomial added to a conic section, but a range of others exists, see e.g [1–4]. While diamond turning machines can deliver parts with the desired figure and finish in one step for some applications (e.g. infrared and illumination), higher precision contexts generally require grind-and-polish processes. For this latter category of molds and parts, a specially tailored gradient-orthogonal basis was introduced to characterize shape [5]. While this basis was constructed to be robust and efficient and to facilitate the assessment of manufacturability, it was recently shown to deliver another benefit. In particular, the associated coefficients were found to be especially effective degrees of freedom during design [6]. Although they span the same space of surface shapes as other common representations, solutions with superior optical performance were discovered when optimizing in terms of this new description. Together with its efficiency, the gradient-orthogonal basis therefore offers significant promise for applications beyond those driven by consideration of “grind-and-polish manufacturability”.
Effective means for interconverting alternative descriptions of shape are valuable because no single convention can meet all needs. One of the advantages of an orthogonal basis is that fitting any particular shape in such a basis can often be performed more efficiently by exploiting orthogonality. Since the values of the associated fit coefficients are independent of the particular subset of the basis elements that is in use and follow directly from simple inner products, the resulting fitting process is sometimes referred to as an “orthogonal projection” onto the basis. A projection of this sort is reported here for the gradient-orthogonal basis of [5]. The process is constructed to avoid the need for determining derivatives and is shown to open the possibility of new applications of these ideas for the analysis of mid-spatial frequency structure (MSF) on manufactured parts.
Modern sub-aperture tools lack the automatic smoothing of traditional spherical fabrication processes. This has given unprecedented importance to both quantifying and tolerancing MSF, but these steps continue to be problematic. Although it may seem intuitive to use Fourier methods to determine a power spectral density (PSD), the lack of data beyond the part’s aperture causes the well-known phenomenon of “spectral leakage”. Various window functions have been introduced to control this corruption, but these inherently mask edge effects on the part that are oftentimes of critical interest. “Detrending” is another standard step for reducing the impact of spectral leakage. This involves subtracting a low-order polynomial from the data to help to suppress the artificial discontinuity at the aperture’s edge before applying a window and Fourier transform. The arbitrary aspects of these steps mean that interpretations of the end results are troubled with questions. It is the finite aperture that generates these complications so it is natural to seek a Fourier-like decomposition that is matched to the aperture at its foundation. The robustness and efficiency of the general sorts of processes described below create such an option for fitting manufactured part shapes entirely with polynomials. The coefficients in the resulting fit are shown to provide a new spectrum that enables familiar types of spatial frequency filtering operations and a promising quantification of MSF.
In keeping with [5], the sag of the freeform surfaces considered in this work is expressed in terms of cylindrical polar coordinates as where, with ,
Here, is the curvature of the best-fit sphere, is the semi-diameter of the cylinder that tightly encloses the surface, and is a polynomial of order . Notice that only the lower limits on the indices of summation are specified in Eq. (1.1); their upper limits can be chosen in a number of ways. As described in [6], all terms of up to order are included if these sums span where the Cartesian order is given byThe expression in Eq. (1.1) is constructed so that the displacement along the normal to the best-fit sphere from that sphere over to the freeform surface is well approximated by the component within the braces. That component is referred to simply as the normal departure.When fitting the parameters in Eq. (1.1) to some other representation of sag, say where the function is regarded as a black box, the objective is to determine values for , , and to match to . The first step is to ensure the best-fit sphere has the required sag values at and at . Because evaluates to zero, the alternative sag description is given a constant bias of , where “” means “is defined to be”. This offset in simply translates the surface without changing its shape. If averaging over angle is denoted by
then because the component within braces in Eq. (1.1) averages to zero at (i.e. at ), is chosen to satisfySolving Eq. (1.4) for yields the first of the fit parameters:To complete matching the expression in Eq. (1.1) to , it is convenient to introduceThe fitting process is thereby reduced to performing a linear least-squares fit for and to match these normal departures:Fitting to a linear combination of, say , basis elements generally requires the solution of a set of linear equations involving an Gram matrix [7]. For an orthogonal basis, however, this process simplifies significantly. To exploit orthogonality for Eq. (1.7), the fit must be performed over the entire enclosing disc (i.e. and ) even if the surface’s clear aperture is only a subset of that disc. The simplification starts, for this case, by splitting the process into two phases. In particular, rather than solving for all of the fitting coefficients in a single step, it is significantly more efficient to break the operation into separate stages of azimuthal and radial conversions. First, for any specific value of , it is straightforward to determine and so that
This is simply a Fourier series in each annular zone and, without loss of generality, can be taken to be zero. It follows from Eqs. (1.7) and (1.8) that the fit can then be completed by choosing and to satisfy the relationsThe Fourier series methods used for Eq. (1.8) are summarized in Sec. 2. That brief review includes a discussion of the impact of aliasing and, importantly, it gives an idea of an analogous issue in working with Eqs. (1.9) and (1.10) to complete the projection. A powerful method for solving Eq. (1.9) for has already been developed [8] and its end result is summarized below in Sec.3. As discussed in Sec. 2, standard Fourier methods involve a particularly simple variant of Gaussian quadrature. Similarly, as pointed out in Sec. 4, the optimal solutions for Eqs. (1.9) and (1.10) involve what is known as Chebyshev-Gauss quadrature [9]. Demonstrations are presented in Secs. 5 and 6 before the Concluding Remarks of Sec. 7. Readers who seek further inspiration to take on the mathematical details may wish to jump first to the figures of Sec. 6.
2. Fourier methods for the azimuthal fit
Although the Fourier series is elementary, some of its aspects offer helpful insights for the sections that follow. For clearer access to the points of interest, I initially express the results in terms of complex exponentials; they are later recast in terms of explicitly real-valued functions. The Fourier series for a function of period can be written as
In the notation of Eq. (1.3), these basis elements are readily seen to be orthogonal:Here, and are integers, is the Kronecker delta and, because the integral is over a full period regardless, this equality holds for any value of .When the coefficients, i.e. , are chosen to minimize the mean squared modulus of the difference between the left- and right-hand sides of Eq. (2.1), it follows from Eq. (2.2) that the associated Gram matrix is an identity. These coefficients are therefore given directly by
That is, as a result of orthogonality, the computations typically required for building up a Gram matrix and solving the associated linear equations are avoided entirely. This is an example of a projection of the type described in the Introduction where the optimal value of any fit parameter is independent of the presence of others and, in this case, is given by the integral in Eq. (2.3). Finding analogous results for a projection to is one of the objectives of what follows, but another significant simplification is also sought.Notice that the dominant computation for the fit of Eq. (2.1) by using Eq. (2.3) lies in numerically approximating that integral for all the required values of . Valuable results emerge if a uniformly distributed and weighted sum is used for that approximation. For example, consider the case of sampling at for where (with, say, in mind). The integral in Eq. (2.3) is then approximated by
The key relation between and its approximant, , now follows upon substituting Eq. (2.1) into Eq. (2.4):The sum inside the final braces of Eq. (2.5) is just a geometric series and, provided , it is readily evaluated in closed form and shown to vanish since is an integer. The only non-zero contributions arise when . These occur whenever is a multiple of , in which case the contents of those braces trivially evaluates to. This observation gives the desired relation thatThe mixing of terms in Eq. (2.6) is referred to as aliasing. For band-limited functions, i.e. where vanishes whenever for some , it follows that is precisely equal to over all of the non-zero band provided . That is, if we sample sufficiently densely, all the terms but vanish in Eq. (2.6) for . This requirement that is the familiar Nyquist sampling condition. In most cases of interest to us, decays rapidly (oftentimes exponentially) as grows, so the functions are effectively band limited and aliasing has negligible impact when sufficiently dense sampling is used, i.e. is sufficiently large. The second objective in what follows is to find analogous summation-based processes that reduce the computational burden of the steps involved in the final stage of projecting to, i.e. for working with Eqs. (1.9) and (1.10).
Before proceeding, it is helpful to notice that when is real valued, it follows from Eq. (2.3) that . Now, if is separated into real and imaginary parts as , then and , hence Eqs. (2.1), (2.3) and (2.4) can be expressed in real-valued terms as
Interestingly, Eq. (2.9) is a variant of Gaussian quadrature where, with discrete sample locations and weights (which turn out to be uniform), a weighted sum of sample values exactly matches the integral for a set of functions of interest with one additional constraint. The constraint can be taken to be that the first sample sits at the origin. It is subsequently readily seen that the samples can all be translated and that, for any value of ,where “” means “is identically equal to” and is any one (or linear combination) of and . Equations (2.7) and (2.9) are used for estimating and of Eq. (1.8) and note that, for even more speed, the sums in Eq. (2.9) can be implemented as FFTs [10].When the terms in Eq. (2.7) beyond all vanish, there are evidently coefficients that can be seen to be determined exactly by using Eq. (2.9) with any . (This follows more simply in the complex-valued representation where trigonometric product-to-sum rules are trivial.) In other words, the number of coefficients that can be determined exactly by way of a sum then matches the number of sample values of the original function. This sets an unbeatable efficiency target for the projections sought in the following sections. Because symmetric functions appear below, i.e. , notice that only is non-zero in such cases and can be determined by a collapsed sum over just . Again, the number of fitted coefficients can readily be seen to then match the number of samples.
3. Rotationally symmetric radial fit (i.e. m = 0)
To account explicitly for the role of the best-fit sphere, a pre-factor is included on the term in Eq. (1.1) so this component must be handled separately. It is shown in [8] that the fit required for Eq. (1.9) can be achieved in two remarkably simple steps. The first is to perform a projection in terms of an auxiliary set of orthogonal polynomials, which leads to a set of coefficients given simply by
where, in terms of of Eq. (1.9), is defined byThese follow from Eqs. (4.4) and (B.4) of [8] (where is written there as ) and reveal that we are once again essentially computing Fourier coefficients: c.f. Equation (2.8a) and (3.1). As a result, is a spectrum that typically decays rapidly. [Keep in mind that the selection of the best-fit sphere, together with Eqs. (1.6) and (1.8), mean that has a repeated root at as well as roots at , so of Eq. (3.2) is well defined at these limits and, in fact, approaches zero at .] The second of these two steps converts to the desired fit coefficients, i.e. , by using Eq. (3.2) of [8], namelyfor . Simple expressions for , , and are given in Eqs. (A.14) to (A.16) of [8], which allow these entities to be pre-computed with just a few arithmetic operations each. To suppress aliasing, it is assumed that Eq. (3.3) is taken to sufficiently large values of that has trended to a negligibly small value. All but the terms of interest can then be discarded.For Eqs. (3.1) and (3.2), notice from Eq. (1.8) that
where Eq. (2.4) has been used again for this approximation. More interestingly, the function to be averaged in Eq. (3.1) is evidently symmetric in and has period , so we need only sample in one quadrant:whereFor extra speed, the sum in Eq. (3.5) can be implemented with an FFT, in particular with the variant of the discrete cosine transform called DCT-IV. [Note that this result was given in Eq. (4.5) of [8], but there is an overall factor-of-two error in that equation that resulted from the same normalization error in Eqs. (B.6) and (B.7) of that work. The normalization is correct in the computational results reported there, however.] If can be fitted exactly by using a sum up to in Eq. (1.9), then it follows from the results of Sec.2 regarding summation of Fourier terms that the exact fit coefficients are given by the sum in Eq. (3.5) provided . That is, this remarkably simple scheme almost matches the efficiency described at the end of Sec.2: it is now necessary to have just one more sample than the number of coefficients being fitted. Of course, when working with functions that are not precisely band limited, it is better to add extra samples to further suppress aliasing.4. Radial fit for m > 0
With the estimates derived in Sec.2 for and of Eqs. (1.10a,b), the final step in the projection is to determine and . These equations both have the same form, which can be written generically as
where and stand for either and or and , respectively. Because is a constant throughout this section, it is now suppressed as an index where possible.The determination of in Eq. (4.1) can be expressed in terms of a sequence of elementary operations that involve matrices with just one off-diagonal band. It is convenient therefore to define an upper-triangular matrix, say , by
A lower triangular matrix is also defined here by . Notice that the computation of , where , proceeds from to simply withand ends with . The inverse operation, i.e. , need not explicitly invert the matrix but can proceed by starting with and working down in from to with the recurrence relationSimilarly, starts with and works from to withso starts with and works from to withIn this notation, it is shown in the Appendix that the coefficients in Eq. (4.1) can be determined by a sequence of these simple recurrence-based operations, namely
where the elements of follow from Eq. (A.4):Here, is given by Eq. (3.6) and is the auxiliary polynomial of order introduced in [5]. Simple expressions for the matrix elements and as well as and for Eq. (4.7) are given in the Appendix; those for and appear in an appendix in [5]. An efficient implementation of Eq. (4.8) requires the recurrence relations given in Eqs. (A.2-6) of [5]. Those relations allow each polynomial to be evaluated robustly with just a handful of arithmetic operations regardless of its order. It follows from Eq. (4.1) that the first integrand in Eq. (4.8) is an even function of , hence the second integrand is an even function of with period . The sum over one quadrant in Eq. (4.8) therefore matches a sum over a full period of the summand and thus has all the striking properties discussed in Sec.2. In particular, the coefficients of Eq. (4.1) can be extracted exactly by Eqs. (4.7) and (4.8) provided . That is, the efficiency described in the previous section is matched yet again. As pointed out in the Appendix, Eq. (4.8) is an application of Chebyshev-Gauss quadrature. This is a critical advantage of these projections that enables such impressive performance.5. Demonstration
An example of a freeform surface from the patent literature was considered in Sec. 2.2 of [6]. This surface has a plane of reflection symmetry and its conventional description involves Cartesian polynomial terms of the form where and is always even on account of the symmetry and coordinate choice. The end result of expressing this surface in terms of was also given in Fig. 4 of that work, but without derivation. An efficient means to achieve that conversion has been laid out above and it is now helpful for code verification to note some of the intermediate results. To give extra digits for numerical checks, all units in this section are expressed in femtometres (fm). For confirmation of the sag itself, now written with Cartesian arguments as , some sample values are
The symmetry in this example means that it is convenient to choose the coordinate relation so that , hence . With this, for each of the azimuthal averages [such as in Eqs. (1.5) or the determination of for Eqs. (3.4) and (4.1)] it is simplest to sample uniformly at for . It turns out that is sufficient for this tenth-order surface. When and is used in Eqs. (4.1) and (4.8), Eq. (4.7) reproduces the coefficients in Fig. 4 of [6] to sufficient accuracy to match the sag at the sub-nm level. This involves sampling at the 66 locations shown in Fig. 1 (associated with and where and ) and yields for and . Of those 55 fitted coefficients, only the 34 of order up to ten need to be retained and, as discussed in [6], these correspond to the original tenth-order Cartesian terms given in the patent. (The 21 higher-order terms that were dropped make only sub-nm contributions.)
The effects of aliasing are suppressed by using larger values for and (hence also larger given ). Because the projection developed above is so efficient, there is little need to run with minimal values of these parameters for simple sag conversions: computing more coefficients allows confirmation of the decaying spectrum and then only the terms of significance are ultimately retained. Increasing has negligible impact in this example, but accuracy is further improved by using (so ). With this choice, when , the vectors written in the Appendix as , , , and are found to be
Notice that the last of these columns holds the spectral coefficients in the column of Fig. 4 in [6]. Twice the number of coefficients is given now and each is to six additional significant digits (i.e. rounded off in fm instead of nm). Similarly, as an additional check, these intermediate results for areAgain, the last column in Eq. (5.3) gives the coefficients for in Fig. 4 of [6]. As discussed in Sec.3, the column is handled differently and, in this case, I note just the final fit coefficients:Importantly, the result from Eq. (1.5) must be retained to full precision for use in Eqs. (1.6), (3.2), and (3.4) when determining the values in Eq. (5.4); it is only after this conversion that the best-fit curvature can be rounded to lower precision in keeping with the power tolerance associated with the sag.The intermediate results given in Eqs. (5.2–4) are offered as an aid in the identification of coding errors. Ultimately, however, there is an acid test for verifying any code that implements such projections: plot and analyze the difference between the original sag expression and the new one. Similarly, there are analogous intermediate checks such as plotting the difference between the left and right sides of Eq. (A.9), etc. Notice too from Eqs. (5.2–4) how rapidly the coefficients decay and therefore what it means to say that such cases are “effectively band limited” (hence that aliasing can readily be suppressed). It is this property that makes it possible to use highly efficient sums in place of integrals in Eqs. (2.9), (3.5), and (4.8).
6. Fitting mid-spatial frequencies
As pointed out in Sec.3 of [5], it is sometimes helpful to express the elements within the brackets on the second line of Eq. (1.1) in terms of their amplitude and phase, say
and the mean square gradient of the normal departure [i.e. the component within braces in Eq. (1.1)] is then given by just the sum over and of the square of . What is more, as presented in [6], alone appears in some low-order testability estimates; is of secondary importance in that regard. These properties are consistent with the observation that if we re-fit a part after rotating it through an angle about the centre of the aperture, i.e. is replaced by , then is invariant while simply changes by . This is the familiar property of translation of a function causing a phase change in the Fourier domain. It is natural to seek analogues of other basic Fourier properties in this context and to carry over as much intuition as possible from that realm. I begin this for specific sag functions while focusing primarily on the reduced form of Eqs. (1.6) and (1.7). In this way, the distractions of cosine factors and base spheres are largely suppressed and the results are always expressed in terms of a unit disk.6.1 Projections of pure sinusoids
Consider, for example, a sag function that is perfectly sinusoidal. Because the orientation of the variations influences only , it can be chosen arbitrarily and the invariant values of are of primary interest. To be consistent with the example in Sec.5, the sag is oriented to be symmetric in and . As a result, and it follows that with . Start then with a sag of unit peak value of the form
which has full cycles between with a phase determined by . When , the sag in Eq. (6.2) is symmetric in . Because changing the sign on alone is equivalent to replacing by, it follows upon considering the impact of this change on that (hence also ) must then vanish for all odd values of . Given that the Cartesian order of the contribution associated with either , , or is given by Eq. (1.2), only terms of even Cartesian order appear in such cases. Alternatively, for the sag is anti-symmetric in and is then zero for all even values of . Further, if the sum of the squares of all of Cartesian order is written as then alternate entries vanish in these two cases, i.e. vanishes when is either even or odd. Of course, the definitive property of means that the sum over of is just the weighted mean square gradient of the fitted polynomial over the unit disc. I use the acronym PSS to refer to this partially summed spectrum, viz. .To clarify some of these observations, specific results are presented in Fig. 2 for the case where, to flatten the best-fit sphere of Eq. (1.5) and suppress the cosine effects of Eq. (1.6), I choose . (When the sag variation is much smaller than the aperture size, the part is essentially flat, so this aperture size is used in all the simple sinusoidal cases considered here.) Notice for the anti-symmetric case, i.e. for as presented in the left column of Fig. 2, that the polynomial is dominated by the tilt, coma, and trefoil terms, i.e. , , and , respectively. The symmetric case, where , is dominated by spherical aberration, astigmatism of second and fourth order, and tetrafoil, i.e. , , , and , respectively. In both cases, the magnitude of the largest tabulated coefficients is of the order of unity and the overall fit error in both cases is below 10−6 times the peak-to-valley of the sag. The PSS is evidently almost all at third order for the odd case, and at second and fourth order for the even case.
When expressed in terms of a normalized coordinate, say , and averaged uniformly over an integral number of periods, the rms slope of is . For integral values of therefore, although it corresponds to a differently weighted average, the number given as “rss” in the plots in the third row can be expected to be approximately equal to . This estimate also applies when averaging over any region much larger than a period, so the resulting approximation should be valid for all much larger than unity.
It follows from Eq. (6.2) that the results for a sinusoid of different phase can be found by simply taking a linear combination of the coefficients in the second row of Fig. 2. For example, a phase of corresponds to an equal mixture of these even and odd cases, which interleaves the results so that the entries are then all non-zero and scaled by . Similarly, the PSS is found by interleaving the two plotted partially summed spectra. Keep in mind too that rotating the sag (i.e. giving the colored contours some other orientation in Fig. 2) means that then becomes non-zero in keeping with the comments made about rotation following Eq. (6.1), but remains untouched. For any fixed frequency , therefore, the entities and for the case give access to a general characterization of such a sinusoidal sag of arbitrary orientation and phase. Some results are presented in Fig. 3 for the cases and .
The spectral amplitudes, i.e. , are plotted in the second row of Fig. 3 because there are too many coefficients to tabulate. The color scale for these plots is logarithmic and ranges from red for the peak value down through seven orders of magnitude to purple. The coefficients of smaller value are not visible in this representation. In fact, rapidly falls by orders of magnitude below the blue diagonal zone in those plots. It is striking that the contours of coincide closely with contours of and that the highest ridge of that distribution occurs near . As a result, the PSS —found by summing along contours of — peaks strongly near . More importantly, the majority of the PSS falls within a narrow interval around this peak. To emphasize this and clarify the trend, the PSS for with is plotted in Fig. 4. The rss values shown in Figs. 3 and 4 confirm that, as anticipated in the discussion of Fig. 2, the rms gradient is approximated well by .
6.2 Frequency filtering the Q-spectra for a sinusoid
It turns out that 60% of the blue area in Fig. 4 falls within the last four lobes of the distribution, i.e. , so the rss of these components alone is over 75% of the total. More generally, sinusoidal variations with a period of are principally localized in their PSS around the region . By combining these two relations, it follows that the abscissa for the PSS can be loosely interpreted as a spatial frequency through
A more complete appreciation of the associated frequency resolution follows upon separately summing different bands of these spectra. As an example, three sub-bands of the spectrum in the right-hand column of Fig. 3 are shown in Fig. 5. Notice that the spacing between the ridges that run from top to bottom near the centre of each plot is consistent with Eq. (6.3) given that the dominant components in each are near equal to 30, 60, and 75, respectively. It is striking, however, that spatial variations are judged by this basis to be of lower frequency when they are nearer and parallel to the edge. That is, the dominant variations near the left and right edges of these plots appear in the two lower frequency bands (which fortuitously means that a part’s typical edge effects can be more readily captured in such spectra). Most of the original sinusoid is clearly in the rightmost plot in Fig. 5, however, and the rms slope of these filtered components doubles at each of the two steps from left to right in the figure. Together, these components constitute the associated sinusoidal form to within ± 0.0017; the balance falls in and that residual drops by more than a factor of ten with each few additional orders in .
These Q-spectra can be filtered not just in terms of , but also in (for separation of different angular frequencies) and (for different radial frequencies). An example is given in Fig. 6 where the filtering is based on angular frequency. Because the results in Fig. 6 are a combination of filtering in and , they are also effectively filtering in . That is, as can be appreciated by tracking the variations along either some diametral lines or concentric circles, both the radial and angular frequencies change as expected in each of the two steps from left to right. The capabilities of such processes are explored by analyzing more complex data in the next section.
6.3 Demonstration with synthetic data
A better appreciation of the ideas in the previous subsection can be obtained by analyzing some synthetic data that mimics as-built part shapes. For this, a superposition of a sum of thousands of sinusoidal terms with a randomized, decaying Fourier spectrum is overlaid with some additional structure. The fitting process for this case determines and over from 0 up to and from 0 up to . To achieve this, samples are taken at with where is given in Eq. (3.6) and . Since there is no plane of symmetry, the sampling is no longer in a half-plane as in Fig. 1. Instead, the azimuthal averages used to determine and for Eq. (4.1) involve sampling at for where and . That is, 302 samples are taken on each of 77 rings so, in this case, sag samples are used to determine coefficients. This is close to the one-for-one efficiency standard set by the discrete Fourier methods. As an aside, since the Fourier series on each ring is independent, an attractive and generally equivalent alternative is to sample on the ring at where is the golden ratio. The distinction is displayed in Fig. 7 and is related to Fibonacci lattices [11]. Of course, the number of samples around these rings can be boosted to a power of two to increase the benefit delivered by FFT algorithms, if desired [10].
The spectrum for this example is plotted in Fig. 8 and the data itself for the normal departure appears in Fig. 9(a). At a demanding yet realistic level, if the units are taken to be millimeters and the departure needs to be characterized down to the nanometer level, the fit must be accurate to better than one part in 106. In this case, it is evident in the plot at left in Fig. 8 that only about a half of the 23,000 fitted coefficients are large enough to be significant for these purposes, viz. those within . (Remember that these fits must be taken to sufficiently high orders to suppress corruption by aliasing, thus subsequently disregarding higher orders is an essential aspect.)
It is instructive now to perform some filtering. First, it is clear from the plot at left in Fig. 8 that the spectrum is dominated by the terms satisfying . Figure 9(b) reveals how well the shape is captured by those first fit coefficients. The residual is almost indistinguishable from what is plotted in Fig. 9(c), which is the sum of the components in the spectrum within . Striking structures become visible at the higher spatial frequencies presented in Figs. 9(d), 9(e), and 9(f). Such MSF patterns are oftentimes found on parts manufactured with modern sub-aperture tools: spoking is evident in Fig. 9(d) and two sets of raster marks at 60° to each other are clear in Fig. 9(e) while Fig. 9(f) suggests some high-order rings on the surface. As a further demonstration, the spoke and high-frequency ring components are more specifically isolated in Fig. 10, but the more impressive and central result is the separation of spatial frequencies demonstrated in Fig. 9: the capability to decompose the data in Fig. 9(a) into the five bands shown is both striking and useful.
More realistically, metrology is corrupted by noise of various types. An indication of the impact that noise has on these fitting processes is presented in the PSS in Fig. 8. It turns out that the spectral coefficients are all impacted at about the same absolute level. Further, it is pleasing that summing the difference between the clean and noisy spectra for coefficients above the noise floor yields a map whose peak value is comparable to the noise level. Another type of noise is encountered when interpolating from pixelated metrology to obtain the desired sample values. To explore this, I used a pixel grid with bi-linear interpolation and the resulting PSS is shown as the purple curve at right in Fig. 8. In this case, summing all terms within gave an absolute error less than 0.0005. This means that all the sub-bands in Fig. 9 except (f) can be extracted in this way. Interestingly, Fibonacci sampling yielded equivalent results, but gave slightly larger coefficients at high azimuthal orders for this interpolated case.
For parts with circular apertures, sampling processes of the type shown in Fig. 7 appear to be promising partners for coordinate measuring machines. When the metrology yields a pixelated map on the other hand, although interpolation is then a workable option, the process can in practice be complicated by the presence of regions of invalid data. In such cases and, more generally, for intentionally non-circular apertures there are useful modifications based on least-squares methods for the fitting processes developed above. While the factoring into azimuthal and radial stages remains a critical step, the details are not discussed further here.
7. Concluding remarks
The goal of this work was to develop a robust projection for expressing nominal shapes in terms of the gradient orthogonal basis . Although the natural projection would involve derivatives for determining gradients, this could sometimes be problematic and was found to be unnecessary. Despite the fact that the underlying orthogonality is based on an integral, the primary results, namely Eqs. (3.3), (3.5), (4.7) and (4.8), are of comparable efficiency to discrete Fourier transforms. That is, they yield almost one fitted coefficient for every sample point evaluated within the circular aperture. Further, the first stage of this projection can be implemented by using FFTs, and the second uses Chebyshev-Gauss quadrature to deliver its remarkable overall efficiency. It is also worth noting that the projection developed in the Appendix to an intermediate basis can equally be followed by a change of basis using Eqs. (B.7-9) of [5] together with Salzer’s recurrence-based methods of [12]. In this way, the results can be expressed in terms of other orthogonal polynomials if required, e.g. Zernike polynomials (although there can be stability problems at extreme orders in that case). That is, the methods developed above allow freeform shapes that are defined over a disc to be expressed in terms of a variety of standard orthogonal polynomials.
Because they readily extend to many thousands of terms, the methods developed here create possibilities for a novel characterization of MSF on as-built parts —regardless of whether the parts are nominally rotationally symmetric or intentionally freeform. This has long been thought to be beyond the reach of polynomials alone. It is also significant that early innovations for coping with the evolving challenges of tolerancing MSF led to a realization of the importance of an rms gradient metric in some applications [13]. It is noteworthy then that the projection to gives direct access to the rms gradient, and that this is connected to the rms radius of a ray spot diagram. Further, the power of the spatial frequency-filtering processes demonstrated in Sec. 6 is perhaps the most promising surprise to emerge from this work. It establishes that, on top of their benefits in nominal shape characterization and design, tailored orthogonal polynomials also have the potential to serve as a useful diagnostic and qualification tool for MSF-related aspects of optics production.
Appendix: Derivation of the radial fit
The weighted average used in the definition of gives the radial analogue to the azimuthal average of Eq. (1.3):
where the normalization factor in front of the integral ensures . The goal in this appendix is to determine the coefficients of Eq. (4.1). This is broken down into three steps. In all the cases considered, is a symmetric function of so the integral in Eq. (A.1) is just one half of the same integral taken over , hence it has the form required for the application of Chebyshev-Gauss quadrature [9] which enters this work in Eq. (4.8).The first step is to project to the auxiliary basis of [5], in particular to , and this amounts to choosing to minimize the mean square error given by
The element in row and column of the associated Gram matrix isand the ’th element of the vector on the right-hand side of the so-called normal equations isSince is just scaled version of the Jacobi polynomial , it follows from Eq. (A.9) of [5] (with replaced throughout by ) that is a linear combination of and . It immediately follows from the orthogonality of the Jacobi polynomials that the Gram matrix of Eq. (A.3) is tri-diagonal.By using Eq. (A.10) of [5] (again, with replaced throughout by ), the above-diagonal elements of the Gram matrix, say , are found to satisfy
Similarly, the diagonal elements, say , can be expressed in terms of asAs in Eq. (A.18) of [5], the elements of the Cholesky decomposition of this matrix are found by starting with and working progressively up from to by usingIn terms of the notation of Eqs. (4.2-7), the Gram matrix of Eq. (A.3) is equal to and the solution for in Eq. (A.2) can therefore now be written as . As an aside, I note that it can be useful to extend these results to include : first, when fitting to other polynomials such as Zernikes, the projection to can be a helpful first step for all and, second, is itself gradient orthogonal so it can be scaled to replace in non-standard applications. Accordingly, note that , , , , while for all they are given byThe second step is to change the basis from to . That is, we now seek such that
By taking and in Eq. (22.7.19) of [9] and using Eq. (A.4) of [5], it follows thatwhere It follows from Eq. (A.10) that the coefficients in Eq. (A.9) are related by . The final step is to change basis from to , and Eq. (B.5) of [5] states that this step can be expressed as . Finally, these three stages can be combined to arrive at the desired result for this appendix: .For verifying any code that implements these relations, it is helpful to have a sample of low-order values for these entities that will typically be pre-computed and stored:
References and links
1. A. Yabe, “Representation of freeform surfaces suitable for optimization,” Appl. Opt. 51(15), 3054–3058 (2012), doi:. [CrossRef] [PubMed]
2. I. Kaya and J. P. Rolland, “Hybrid RBF and local ϕ-polynomial freeform surfaces,” Adv. Opt. Technol. 2(1), 81–88 (2012), doi:. [CrossRef]
3. P. Jester, C. Menke, and K. Urban, “Wavelet Methods for the Representation, Analysis and Simulation of Optical Surfaces,” IMA J. Appl. Math. 77, 357–363 (2012).
4. R. Steinkopf, L. Dick, T. Kopf, A. Gebhardt, S. Risse, and R. Eberhardt, “Data handling and representation of freeform surfaces,” Proc. SPIE 8169, 81690X, 81690X-9 (2011), doi:. [CrossRef]
5. G. W. Forbes, “Characterizing the shape of freeform optics,” Opt. Express 20(3), 2483–2499 (2012). [CrossRef] [PubMed]
6. C. Menke and G. W. Forbes, “Optical design with orthogonal representations of rotationally symmetric and freeform aspheres,” Adv. Opt. Technol. 2(1), 97–109 (2012), doi:. [CrossRef]
7. W. Press, S. Teukolsky, W. Vetterling, and B. Flannery, Numerical Recipes: The Art of Scientific Computing (Cambridge University Press, 1992) Section 15.4.
8. G. W. Forbes, “Robust, efficient computational methods for axially symmetric optical aspheres,” Opt. Express 18(19), 19700–19712 (2010), doi:. [CrossRef] [PubMed]
9. M. Abramowitz and I. Stegun, Handbook of Mathematical Functions (Dover, 1978). See 25.4.38.
10. See Section 12.3 of [7]. Or see Sec. 12.4 in http://apps.nrbook.com/empanel/index.html#
11. J. H. Hannay and J. F. Nye, “Fibonacci numerical integration on a sphere,” J. Phys. Math. Gen. 37(48), 11591–11601 (2004), doi:. [CrossRef]
12. G. W. Forbes, “Robust and fast computation for the polynomials of optics,” Opt. Express 18(13), 13851–13862 (2010), doi:. [CrossRef] [PubMed]
13. J. K. Lawson, J. M. Auerbach, R. E. English Jr, M. A. Henesian, J. T. Hunt, R. A. Sacks, J. B. Trenholme, W. H. Williams, M. J. Shoup III, J. H. Kelly, and C. T. Cotton, “NIF optical specifications: the importance of the RMS gradient,” Proc. SPIE 3492, 336–343 (1999), doi:. [CrossRef]