optimum waist of localized basis functions in truncated series employed in some optical applications
TRANSCRIPT
Optimum waist of localized basis functionsin truncated series employedin some optical applications
Farshid Ghasemi* and Khashayar MehranyDepartment of Electrical Engineering, Sharif University of Technology,
P.O. Box 11365-8639 Tehran, Iran
*Corresponding author: [email protected]
Received 9 September 2009; revised 16 January 2010; accepted 23 January 2010;posted 1 February 2010 (Doc. ID 116797); published 3 March 2010
The waist parameter is a particularly important factor for functional expansion in terms of localizedorthogonal basis functions. We present a systematic approach to evaluate an asymptotic trend forthe optimum waist parameter in truncated orthogonal localized bases satisfying several generalconditions. This asymptotic behavior is fully introduced and verified for Hermite–Gauss and Laguerre–Gauss bases. As a special case of importance, a good estimate for the optimum waist in projection ofdiscontinuous profiles on localized basis functions is proposed. The importance and application of theproposed estimation is demonstrated via several optical applications. © 2010Optical Society of America
OCIS codes: 000.4430, 050.1755, 350.4238, 060.2310, 130.2790, 130.5296.
1. Introduction
Exploitation of a truncated orthogonal series ex-pansion as an approximation for the solution of aphysical problem is a very well-known techniqueamong physicists and engineers. The perpetualquestion propounded in this approach is which basisfunctions with what parameters should be used toachieve the highest accuracy by the least computa-tional effort.Once the set of basis functions is chosen, there is
usually a set of free parameters left to be determined.In particular, the waist parameter for localized basisfunctions canplayan important role in the accuracy ofthe expansion for nonsmooth functions, and its opti-mum determination has attracted much attention indifferent applications [1–7]. To determine theoptimum waist for projecting a specific function onaspecific set of localizedbasis functions, different sug-gestions have been proposed in the literature. For in-stance, Borghi et al. in [1] have proposed a rule of
thumb for finding the optimum waist of two-dimensional (2D) Laguerre–Gauss (LG) basis func-tions in expanding the circ function. Liu et al. in [2]have studied the applicability of this rule in find-ing the optimum waist for one-dimensional (1D)Hermite–Gauss (HG) basis functions employed for ex-panding a step function. Some have also used struc-ture-related parameters, e.g. normalized frequencyof an optical waveguide, to determine the basis func-tion parameters [8–11]. Some others have furthercontemplated on the problem for a class of signalsand have minimized an upper bound for the squarederror (e.g., see [3] and references therein).
Accurate determination of the most favorableparameters for basis functions, however, asks for nu-merical optimization techniques. The optimizationprocess is, however, nonlinear and rather complexand, hence, calls for complicated optimization algo-rithms, e.g., iterative methods [5,6]. The difficultygets even worse when the error function exhibitsmany local minima, increasing the probability thatthe optimization algorithm gets stuck in a certainminimum. To avoid such difficulties, some effortshave been made to analytically solve the nonlinear
0003-6935/10/081210-09$15.00/0© 2010 Optical Society of America
1210 APPLIED OPTICS / Vol. 49, No. 8 / 10 March 2010
optimization problem, but the results are restrictedto those circumstances where only a few basisfunctions are retained in the calculations [7]. Unfor-tunately, the optimized parameters may be consider-ably different when the number of retained basisfunctions is increased. Therefore, the optimizedparameters may turn out to be nonoptimum whenmore basis functions are kept. In this context, a goodestimate for the asymptotic behavior of the optimumparameter with respect to the number of basis func-tions is complementary to the already existing opti-mization formulations and obviates the need tofollow a complicated optimization process.To this end, the behavior of optimum waist para-
meter for a predetermined set of localized basis func-tions is analytically studied in this paper, where thefollowing assumptions are made. The derivative ofeach basis function with respect to the free param-eter is supposed to be in the same space spannedby the basis functions. The obtained results will,therefore, not be valid for discontinuous basis func-tions whose derivatives include the delta function.Furthermore, despite the generality of the proposedmethod, basis functions are assumed to form anorthogonal set and, thus, their derivative matrix(as defined in Appendix A) is a band matrix. This as-sumption is found very helpful in reducing the com-plexity of formulation. It is worth mentioning thatthe validity, as well as the applicability, of the pro-posed analytic approach in choosing the optimal freeparameter is demonstrated for the HG and the LGbases. The rules of thumb as suggested by [1,2]are then extended and a systematic approach to eva-luateing the asymptotic optimum parameter is pre-sented. Finally, different practical applications arepresented and the applicability and significance ofthe proposed method are exhibited.
2. Formulation
Assume f ðrÞ as a complex function in a completesquare-integrable linear space S⊆L2ðRDÞ, which isspanned by the set of basis functions ψmðr;wÞ. D isthe dimension of the Euclidean space. The sequenceof coefficients fαmgM−1
m¼0 is to be determined in such amanner that
PM−1m¼0 αmψmðr;wÞ becomes the best ap-
proximation for f ðrÞ within S, minimizing thesquared error defined by
e2 ¼Z ����f ðrÞ − XM−1
m¼0
αmψmðr;wÞ����2dr: ð1Þ
Since the set of expansion coefficients fαmgM−1m¼0 guar-
antees the best achievable approximation within S,the following set of linear algebraic equations musthold [12]:
GtMA ¼ BM;
GM ¼ ½gi;k�;
gi;kðwÞ ¼ hψ i;ψki ¼Z
ψ iðr;wÞψ�kðr;wÞdr;
BM ¼ ½bi�; bi ¼ hf ;ψ ii;A ¼ ½αi�; i ¼ 0;…;M − 1: ð2Þ
Here, the index i runs from 0 to M − 1 and thesuperscripts t, T, and � indicate the transpose, theconjugate transpose, and the complex conjugate,respectively. The minimum error e2 can then be sim-plified as
e2 ¼ hf ; f i − BtMG−1
M B�M : ð3Þ
Because w is a free parameter, it can be optimizedto render theminimum error. The value ofw can thenbe found by setting ∂e2=∂w ¼ 0. It will be assumedthat the norm of each basis function hψmðr;wÞi is in-dependent of w. Using the chain rule and after somealgebraic manipulations, ∂e2=∂w can be written as
∂e2
∂w¼ −Bt
MðDTMG−1
M þG−1MDMÞB�
M − BtrDT
r G−1M B�
M
− BtMG−1
MDrB�r ; ð4Þ
whereDM, Dr, and Br are defined as in Appendix A. Itis, however, assumed that the derivative of each basisfunction, ∂ψm=∂w, resides in S. On the other hand, itcan be shown that, when the basis is orthonormal,the following relation holds between GM and DM(see Appendix B):
G−1MDM þ DT
MG−1M ¼ 0: ð5Þ
By substituting the preceding equation intoEq. (4),∂e2=∂w can be further simplified as
∂e2
∂w¼ −2RefBt
MG−1MDrB�
rg
¼ −2Re�XM−1
i¼0
X∞k¼M
hf ;ψ iihf ;ψki�ðg−1M Þidi;k
�: ð6Þ
Here, ðg−1M Þi refers to ði; iÞ entry of G−1M (the inverse of
a truncated Grammatrix). The zeros of this equationwill be employed in Section 3 to discuss the displace-ment of the optimum waist parameter.
3. Asymptotic Behavior of Optimum Waist Parameter
The next problem is to determine the asymptotictrend of optimum w as the number of retained basisfunctions, M, tends to infinity. To study the asymp-totic behavior of the optimum waist parameter, weshould inspect the displacement of the zeros of∂e2=∂w, hereafter denoted by wopt
M , as a function ofM. It is, therefore, necessary to study the dependenceof Dw and also the inner products hf ;ψ ii on w, whosecharacteristics govern the zeros of ∂e2=∂w in Eq. (6).
10 March 2010 / Vol. 49, No. 8 / APPLIED OPTICS 1211
First, it can be shown that, if the waist parameterw just scales the basis functions linearly along theradial direction in RD (as for HG basis) then the en-tries of Dw, i.e., di;k in Eq. (6), are proportional to 1=w.Furthermore, if the basis functions dependmerely onthe radial distance, j�r j (as for LG basis), then all theentries become proportional to D=w. These two factsare later employed in this section to simplify the dif-ference equation governing the zeros of ∂e2=∂w.Second, we study the zeros of the other agent in-
volved in Eq. (6), i.e., FMðwÞ≜hf ;ψMi, as the first steptoward extraction of the asymptotic trend for thesought-after roots of Eq. (6). To this end, FMðwÞ isapproximated around one of its zeros as
FMðwÞ ≈ aMðw −wMÞ; ð7ÞwherewM denotes an arbitrary zero of FMðwÞ and aMis the first-order coefficient in Taylor expansion. Itshould be also noted that FMðwÞ might have severalzeros and wM should, therefore, have another indexdiscriminating various zeros for a fixedM. Here, thisindex is considered to be fixed and is omitted for thesake of brevity.Next, to obtain the relation among zeros for succes-
sive truncation orders, i.e., wM, wMþ1, wMþ2, etc., wecontemplate the derivative of this function withrespect to w:
∂
∂wFMðwÞ ¼ ∂
∂whf ;ψMi ¼
Xi
dM;iFiðwÞ: ð8Þ
Now, by using the fact that dM;i is proportional toD=w and substituting Eq. (7) into Eq. (8), the follow-ing two difference equations corresponding to thezeroth and first-order terms are obtained as follows:
1DaM ¼
Xi
~dM;iai; ð9aÞ
0 ¼ Pi
~dM;iaiwi ; ð9bÞ
where ~dM;i is defined as the value of dM;i for w ¼ 1.Once the trend of wM is extracted by solving the
above-mentioned difference equations, the trendof wopt
M follows the same line. This point is furtherdiscussed in Section 4.
4. Special Bases
In this section, the approach outlined in Section 3 isapplied to the HG and the LG basis functions.
A. Hermite–Gauss Basis
HG functions are defined as
hmðx;wÞ ¼ 1ffiffiffiffiffiffiffiffiffiffiffiffiw
ffiffiffiπpp 1ffiffiffiffiffiffiffiffiffiffiffiffi2mm!
p Hm
�xw
�
× exp�−x2
2w2
�ðx ∈ ℝÞ; ð10Þ
where Hm is the Hermite polynomial [13]. In thisbasis, Eqs. (9) result in fourth-order difference equa-tions. Since even and odd orders become decoupled,without loss of generality, we can assume that M iseven. By changing the variables as M → 2m anda2m → a0
m, Eq. (9a) now turns to
a0m ¼ −1
2
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi2mð2m − 1Þ
pa0m−1
þ 12
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffið2mþ 1Þð2mþ 2Þ
pa0mþ1; ð11Þ
and Eq. (9b) yields
wMþ2
wM−2¼
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiMðM − 1Þp
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiðM þ 1ÞðM þ 2Þp aM−2
aMþ2
¼ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiMðM − 1Þp
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiðM þ 1ÞðM þ 2Þp a0m−1
a0mþ1
: ð12Þ
To acquire the asymptotic behavior of wM, therefore,a0m−1=a
0mþ1 is needed. To this end, the Birkhoff–
Adams theorem [14] can be employed. Regardingthe definitions in [14], Eq. (11) has two normal solu-tions. Since we are interested in the asymptoticbehavior of a0
m, we will retain only the first two domi-nant terms of the solution with respect to 1=m. Inthis way, the solution for a0
m relies on the initial con-ditions of Eq. (11), which, in turn, depend on thegiven function f . However, by plugging the solutionsinto Eq. (12), it would be observed that the initialconditions do not play any role in the first two domi-nant terms of a0
m−1=a0mþ1. The final result then reads
as
a0m−1
a0mþ1
∼ 1þO
�1
m2
�∼ 1þO
�1
M2
�; ð13Þ
where ∼ denotes the asymptotic behavior. Substitut-ing Eq. (13) into Eq. (12) then leads to
wMþ2
wM−2∼
�1 −
12M
�∼
ffiffiffiffiffiffiffiffiffiffiffiffiffiM − 2
pffiffiffiffiffiffiffiffiffiffiffiffiffiffiM þ 2
p : ð14Þ
In the last step of deriving Eq. (14), we have cast theasymptotic behavior in the form of f ðM þ 2Þ=f ðM − 2Þ, so that the contributions of wMþ2 andwM−2 can be discriminated. Finally, for some propor-tionality constant μ, we have
wM ≈μffiffiffiffiffiM
p : ð15Þ
Now, if the derivative matrix Dw is banded, whichis the case for HG and LG bases, then ∂e2=∂w can beexpressed a
1212 APPLIED OPTICS / Vol. 49, No. 8 / 10 March 2010
∂e2
∂w¼ −2Re
� X1≤εi≤W=2
X0≤εk≤W=2
hf ;ψM−εiihf ;ψMþεki�ðg−1M ÞM−εidM−εiMþεk
�; ð16Þ
where εi and εk do not exceed the half-bandwidth ofDw, W=2. By replacing the linear approximation ofhf ;ψ ii around its zero, Eq. (7), into Eq. (16), an ap-proximate quadratic equation for the zeros of the er-ror derivative is achieved. Whatever this equation is,its zeros will obey the same behavior as wM. The rea-son is as follows: for each of the terms in the form ofaMþεðw −wMþεÞ, the zero position is scaled bywMþ1þε=wMþε (ε is either εi or εk). Since ε ≪ M,wMþ1þε=wMþε is independent of ε, up to the secondorder of approximation regarding 1=M. Similarly,aMþ1þε=aMþε, dMþ1−εi;Mþ1þεk=dM−εi;Mþεk , and ðg−1M ÞM−εi=ðg−1Mþ1ÞMþ1−εi are identical for all ε ≪ M and followthe same vein. This means that, with increasing Mby one unit, all the zeros and the coefficients areidentically scaled and, as a result, the zeros of thementioned approximate quadratic equation will bescaled in the same manner as wM.
B. Laguerre–Gauss Basis
LG functions are defined as
φmðxÞ ¼ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi2
n!ðnþ kÞ!
sð2xÞk=2Lk
mð2xÞ
× expð−xÞ ðx ∈ ℝþÞ;
φðDÞm ðr;wÞ ¼ cD
1
wD=2φm
��jrjw
�D; 1
�: ð17Þ
Lkm is the associated Laguerre polynomial [15] and D
is the space dimension. To have an orthonormalbasis, c1 ¼ 1, c2 ¼ 1=
ffiffiffiπp, and c3 ¼ ffiffiffiffiffiffiffiffiffiffiffiffiffiffi
3=ð4πÞp. It must
be noted that this definition of basis only encom-passes the radial functionality in higher dimensions.In other words, it is a complete basis for rotationallysymmetric functions in L2ðRDÞ.For the LG basis, the steps to be taken are very
similar to those for the HG basis. In this case, Eqs. (9)result in two second-order equations:
aM ¼ −D2
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiMðM þ kÞp
aM−1 þ D2
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiðM þ 1ÞðM þ kþ 1ÞpaMþ1 ; ð18aÞ
0 ¼ −D2
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiMðM þ kÞp
aM−1wM−1 þ D2
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiðM þ 1ÞðM þ kþ 1ÞpaMþ1wMþ1 : ð18bÞ
By employing the Birkhoff–Adams theorem and re-taining the first two dominant terms in 1=M, we willobserve the following trend:
wM ≈μ
M1=D; ð19Þ
which, by similar discussions, is expected for theoptimum waist parameter, as well. As proposed byEq. (19), the k defined in Eq. (17) does not appearin asymptotic trend of the optimum waist.
5. Practical Applications
In this section, general aspects of the asymptotic op-timum waist are numerically investigated in severaloptical applications. These numerical observationsare instructive for making judicious estimations ofthe optimum waist in practical applications.
A. Application in Permittivity Profile Expansion
As the first example, we have chosen a three-dimensional (3D) spherical defect to be expandedin the 3D LG basis with k ¼ 0. Employing the expan-sion of the structure is a common practice in someversions of the Galerkin method, such as the loca-lized function method (LFM) [16]. To express thedefect structure, we considered a function with a unitvalue inside the unit sphere and vanishingelsewhere.
In Fig. 1, the approximation in Eq. (19) with μ ¼ 1,i.e., w ¼ 1=
ffiffiffiffiffiM3
p, is compared to the optimum waist
obtained by direct search. As it can be seen, choosingthe distance of discontinuity from origin (r ¼ 1) as μin Eq. (19), is a good approximation. Similarly, it wasobserved that, for expanding functions with a singlediscontinuity at r0 (step, circ, and spherical defect),μ ¼ r0 gives a good approximation for 1D HG andLG bases, as defined in Eqs. (10) and (17).
It is worth highlighting that the error increasessharply when the waist becomes smaller than the op-timum value (Fig. 2). This fact has been repeatedlyobserved for LG and 1D HG bases with differenttruncation orders, and for different functions with
10 March 2010 / Vol. 49, No. 8 / APPLIED OPTICS 1213
finite support. The reason is that, at the optimumwaist, the width of the function and the width ofthe highest-order basis function are comparativelythe same. Hence, by increasing the waist from its
optimum value, some parts of the higher-order basisfunctions will gradually fall outside the supportof the function. Thus, the weight of these basisfunctions must be small to obtain the optimum
Fig. 1. Expansion of a sphere defect using the 3D LG basis. (a) Solid curve, error by w ¼ 1=ffiffiffiffiffiM3
p; dashed curve, error by optimum waist.
(b) Enlarged view of (a). (c) Solid curve, w ¼ 1=ffiffiffiffiffiM3
p; dots, optimum waist. (d) Enlarged view of (c).
Fig. 2. Error versus waist: (a) order ¼ 4, (b) order ¼ 100.
1214 APPLIED OPTICS / Vol. 49, No. 8 / 10 March 2010
expansion. In other words, the effective number ofbasis functions in the expansion will be lowered.On the other hand, when w < wopt, the effectivewidths of all basis functions become smaller thanthe support of the function. In this way, the end partof the support cannot be approximated by any basisfunction, leading to a sharp increase in the error.This fact is found to be useful in determining theoptimum waist when expanding discontinuousfunctions.In Fig. 3, the error of the optimum waist is com-
pared to that of two fixed nonoptimum waists. Whenthe fixed waist is smaller thanwopt in a specific order,as for w ¼ 0:25 and orders of 10 to 40, the error shar-ply detaches from the minimum error. On the other
hand, when the fixed waist is larger than the opti-mum one, as for w ¼ 0:5 and orders of 10 to 100,the error slightly degrades. This is in accordancewith what is already demonstrated and describedin Fig. 2.
B. Application in Bragg Fiber Structure
In this section, the expansion of a Bragg fiber struc-ture [17] in the 2D LG basis with k ¼ 0 is studied.The considered structure [Fig. 4(a)] contains six dis-continuities and its functionality can be expressed asthe sum of six simple step functions. Although thetotal expansion is not simply a superposition of sin-gle expansions for single steps, we know that theerror for reconstruction of a single step drasticallyblows up when the waist becomes smaller than itsoptimum value. Like the previous example, for a sin-gle discontinuity at r ¼ 3, the suggestion ~w ¼ 3=
ffiffiffiffiffiM
pgives a satisfactory approximation of the optimumwaist. If we take the waist smaller than ~w, the errorin reconstruction of the farthest discontinuity shar-ply increases and dominates other errors. On theother hand, if the waist is more than ~w, the recon-struction error increases for all the discontinuities.Hence, it appears that this ~w must work well asan estimation for the optimum waist. The error ofthis estimation is compared to the error floor inFig. 4.
C. Application in the Beam Propagation Method
Optimal free parameter selection in orthonormalbases can be accomplished by minimizing the upperbound of the quadratic truncation error for a class of
Fig. 3. Error versus order: solid curve, optimum waist; dashedcurve, w ¼ 0:5, dashed-dotted curve, w ¼ 0:25.
Fig. 4. Expansion of Bragg fiber structure in the 2D LG basis. (a) Bragg structure. (b) Dashed curve, error by optimumwaist; solid curve,error by w ¼ 3=
ffiffiffiffiffiM
p. (c) Enlarged view of (b) for small orders.
10 March 2010 / Vol. 49, No. 8 / APPLIED OPTICS 1215
functions. Den Brinker et al. in [3] used this approachto find the optimal waist parameter for several bases,including HG and LG. This suggestion, althoughinteresting for signal processing applications, hasseveral major shortcomings. First, the relation pro-posed in [3] calls for computing the norm of the de-rivative of the function. This norm does not exist for adiscontinuous function. Even for slightly smoothenedjumps, this norm is large and results in an obviouslynonoptimum suggestion for the waist parameter.Second, the upper bound minimized in this approachis not tight enough for many instances of practicalinterest. Although [18] has used this optimal waistsuccessfully, it should be noted that the function thisreference expands is the real mode of a slab wave-guide. For such bell-shaped functions as guidedmodes in waveguides, the HG expansion essentiallyconverges fast and is less vulnerable to the choice ofoptimum waist.As an example, we deal with the analysis of the re-
flection from a slab waveguide using BPM [19]. Thispaper has embedded 1D HG expansion in the beampropagation method (BPM) formulation. Since the
incident field is assumed to be a HG beam, it is sui-tably reconstructed by few HG functions. But thestructure has the shape of a smoothened step andits expansion entails many more basis functions.Here we examine the optimum waist for expandingthe structure.
The function to be expanded is a raised cosine.Comparing to [19], we have considered a widercos2 section, so that the suggestion of [3] can producea reasonable estimation. The function is defined as
f ðxÞ ¼(
1 jxj < 1
cos2�π2
�jxj − 1
��1 < jxj < 2 : ð20Þ
There are two discontinuities at jxj ¼ 1 and jxj ¼ 2 inthe second derivative of this function. The second de-rivative of the expansion includes the HG function oforders 0 to M þ 1 with only M degrees of freedom intheir coefficients. However, if we were willing toexpand the second derivative efficiently with M þ 2basis functions, the optimum waist was ~w ¼ 2=ffiffiffiffiffiffiffiffiffiffiffiffiffiffiM þ 2
p≈ 2=
ffiffiffiffiffiM
p. The optimum waist as suggested
by [3] is ~w ¼ 0:5. In Fig. 5, the relative error is plottedagainst the highest retained order. For orders up to15, these two approximations generate almost thesame error. But in higher accuracies ðerror <−50dBÞ, the difference is conspicuous. For order ¼100, they differ by almost 13dB. Also included in thisfigure is the upper boundminimized by [3]. Using thenotations of [3], the plotted curve represents
ffiffiffiffiffiffiffiffiffiffiffiFðwÞM
r≅
ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiPM−1n¼0 na2
nðwÞM‖f‖2
s; ð21Þ
with w ¼ 0:5. As can be seen, this upper bound is nottight enough. The optimum waist is demonstrated inFig. 6. The relation α=
ffiffiffiffiffiM
pis fitted (by least squares)
to the optimumwaist for orders from 20 to 100, whichhas led to αopt ¼ 1:93.
D. Application in Studying Cutoff of Planar Waveguides
As the last example, we make use of the resultsreported in [20] to demonstrate the benefits of theintroduced optimum waist. This paper computesthe cutoff frequency of 1D waveguides by expandingthe solution of the TE wave equation in a 1D HG ba-sis. Under such a condition, the Galerkin method andthe variational formulation yield the same system ofequations [21]. Hence the free parameters must beoptimized in such a way that the cutoff frequencyis minimized.
Fig. 5. Expansion of raised-cosine slab in 1D-HG basis. Error isplotted versus order. Solid curve, optimum waist; dashed curve,w ¼ 2=
ffiffiffiffiffiM
p; dashed-dotted curve, w ¼ 0:5; circles, minimized
upper bound [3].
Fig. 6. Expansion of raised-cosine slab in the 1D HG basis: dots,optimum waist; solid curve, 1:93=
ffiffiffiffiffiM
p.
Table 1. Optimum Waist in Fig. 3 of [20] (Middle Row) andEstimated Optimum Waist (Lower Row)
M 25 50 75 100
W ¼ 1α at minimum 0.143 0.100 0.080 0.067
1ffiffiffiffiffiffi2M
p 0.141 0.100 0.081 0.071
1216 APPLIED OPTICS / Vol. 49, No. 8 / 10 March 2010
We do not know the explicit form of the field andthus we cannot apply the proposed method and besure that the optimal choice of the free parametersis made. It is, however, possible to make use of theproposed method and find reasonable results. Fortu-nately, we know that it is infinitely differentiableeverywhere, except at the boundaries. At the bound-aries, which are positioned at x ¼ 0 and x ¼ 1, thesecond derivative is discontinuous. This knowledgehelps us to propose the optimum waist. Sharmaand Meunier, in [20], have expanded the field versusmerely even orders of HG. So by retaining M basisfunctions, even orders ranging from 0 to 2ðM − 1Þare employed. Hence, it can be expected that ~w ¼1=
ffiffiffiffiffiffiffiffi2M
prenders a good estimation of the optimum
waist. In Fig. 3 of [20], the cutoff frequency is plottedagainst the parameter α, which is equivalent to 1=win our nomenclature. Table 1 contains approximatevalues of w ¼ 1=α for which the cutoff frequencyhas assumed its minimum. This data is extractedfrom Fig. 3 in [20]. By taking advantage of the apriori estimation ~w ¼ 1=
ffiffiffiffiffiffiffiffi2M
p, the need to numeri-
cally search for the minimum point is obviated.Putting away the nonlinear minimization procedurethanks to this a priori approximation, only a simplelinear eigenvalue problem remains to be solved.
6. Summary and Conclusion
We have introduced a systematic approach to extractthe asymptotic behavior of the optimum parameterfor the basis functions complying with several gener-al constraints. The result has been worked out for thewaist of HG and LG bases. For functions includingdiscontinuity in their profiles or their derivatives,an explicit estimation of the optimum waist has beenproposed. This estimation extends the knowledgeand observations on individual discontinuities tomultiple discontinuity functions.Although our discussion has been focused on ex-
pansion of known functions, partial information fromthe essence of the function to be expanded can be ade-quate to conjecture the optimum waist. Especiallywhen the dominant error of the whole problem ema-nates from expansion error (e.g., as for variationalformulations), the knowledge about discontinuity po-sitions in the solution or its higher-order derivativescan be of assistance.As demonstrated by miscellaneous examples, the
estimation is beneficiary in many practical cases.However, it fails for spectrally sparse or band-limitedfunctions. Another difficulty is expected to arise if in-dex-hopping occurs. We have not observed a majorchange in the index of the optimum zero in waist op-timization for the HG (or LG) basis. However, this isnot the case if some other parameters, such as thebasis chirp factor (as defined in [22]) are to be opti-mized. The chirp factor can be considered as a basisparameter, like waist, and can be treated in the sameway. Our primitive experiments on the chirped HGbasis have shown remarkable alterations in opti-mum index.
Appendix A
We assumed that the set of basis functions fψmg∞m¼0is complete in S. Hence if we assume that the deriva-tive of each basis function, ∂ψm=∂w, lies in S, thefollowing differentiation operator can be defined ina matrix form:
∂
∂wΨ ¼ DwΨ; Ψ ¼ ½ψ i�; ðA1Þ
whereΨ is a column vector containing all basis func-tions. Dw is an infinite matrix and is partitioned asfollows:
Dw ¼DM Dr
Ds Dp
: ðA2Þ
Here, DM is an M-by-M block of the infinite matrixDw. On the other hand, B is an infinite column vectorwhose entries represent hf ;ψMi and can be similarlypartitioned as
B ¼BM
Br
; ðA3Þ
Here, BM is the M-by-1 subvector of the infinitecolumn vector B.
Appendix B
The partial derivative of the Gram matrix, ∂G=∂w,can be expressed as
G ¼ hΨ;Ψti; ðB1Þ
⇒∂G∂w
¼ DwGþGDTw: ðB2Þ
The basis is orthonormal so that the norm of basisfunctions ψmð�r;wÞ is independent of w. Hence, thepartial derivative of the Gram matrix, which is inde-pendent of w, is null. Consequently, multiplyingEq. (B2) from both sides by G−1 results in
G−1Dw þ DTwG−1 ¼ 0: ðB3Þ
It is worth noting that, for an orthonormal basis, theGram matrix is unitary, and the above-mentionedequation reveals the antisymmetric essence of a par-tial derivative matrix, i.e., the fact that dmn ¼ −d�
nm.Furthermore, G−1 is unitary and we can truncate thematrices:
G−1MDM þ DT
MG−1M ¼ 0: ðB4Þ
References1. R. Borghi, F. Gori, and M. Santarsiero, “Optimization of
Laguerre–Gauss truncated series,” Opt. Commun. 125,197–203 (1996).
2. Y. Liu and B. Lu, “Truncated Hermite–Gauss series expansionand its application,” Optik (Jena) 117, 437–442 (2006).
10 March 2010 / Vol. 49, No. 8 / APPLIED OPTICS 1217
3. A. C. den Brinker and H. J. W. Belt, “Optimal free parametersin orthonormal approximations,” IEEE Trans. Signal Process.46, 2081–2087 (1998).
4. P. Lazaridis, G. Debarge, and P. Gallion, “Discrete orthogonalGauss–Hermite transform for optical pulse propagationanalysis,” J. Opt. Soc. Am. B 20, 1508–1513 (2003).
5. F. Chiadini, G. Panariello, and A. Scaglione, “Variationalanalysis of matched-clad optical fibers,” J. Lightwave Technol.21, 96–105 (2003).
6. T. Rasmussen, J. H. Povlsen, A. Bjarklev, O. Lumholt, B.Pedersen, and K. Rottwitt, “Detailed comparison of twoapproximate methods for the solution of the scalar waveequation for a rectangular optical waveguide,” J. LightwaveTechnol. 11, 429–433 (1993).
7. I. A. Erteza and J. W. Goodman, “A scalar variational analysisof rectangular dielectric waveguides using Hermite–Gaussianmodal approximations,” J. Lightwave Technol. 13, 493–506(1995).
8. R. L. Gallawa, I. C. Goyal, and A. K. Ghatak, “Fiber spot size: asimple method of calculation,” J. Lightwave Technol. 11, 192–197 (1993);
9. Z. H.Wang and J. P. Meunier, “Comments on ‘Fiber spot size: asimple method of calculation,’” J. Lightwave Technol. 13,1593 (1995).
10. R. L. Gallawa, I. C. Goyal, and A. K. Ghatak, “Optical wave-guide modes: an approximate solution using Galerkin’s meth-od with Hermite–Gauss basis functions,” IEEE J. QuantumElectron. 27, 518–522 (1991).
11. J. P. Meunier, Z. H. Wang, and S. I. Hosain, “Evaluation ofsplice loss between two non identical single mode gradedfibers,” IEEE Photonics Technol. Lett. 6, 998–1000 (1994).
12. D. G. Dudley, Mathematical Foundations for ElectromagneticTheory (Oxford U. Press, 1994).
13. G. B. Arfken and H. J. Weber, Mathematical Methods forPhysicists (Academic, 2005).
14. S. Elaydi, An Introduction to Difference Equations(Springer, 2005).
15. M. Martinelli and P. Martelli, “Laguerre mathematicsin optical communications,” Opt. Photonics News 19, 30–35(2008).
16. T. M. Monro, D. J. Richardson, N. G. R. Broderick, and P. J.Bennett, “Holey optical fibers: an efficient modal model,” J.Lightwave Technol. 17, 1093–1102 (1999).
17. S. Guo and S. Albin, “Comparative analysis of Bragg fibers,”Opt. Express 12, 198–207 (2004).
18. V. García-Muñoz and M. A. Muriel, “Hermite–Gauss seriesexpansions applied to arrayed waveguide gratings,” IEEEPhotonics Technol. Lett. 17, 2331–2333 (2005).
19. R. McDuff, “Matrix method for beam propagation usingGaussian Hermite polynomials,” Appl. Opt. 29, 802–808(1990).
20. A. Sharma and J. P. Meunier, “Cutoff frequencies in planaroptical waveguides with arbitrary index profiles: an efficientnumerical method,” Opt. Quantum Electron. 34, 377–392(2002).
21. A. Sharma and J-P. Meunier, “On the scalar modal analysis ofoptical waveguides using approximate methods,” Opt. Com-mun. 281, 592–599 (2008).
22. P. Lazaridis, G. Debarge, and P. Gallion, “Exact solutionsfor linear propagation of chirped pulses using a chirpedGauss–Hermite orthogonal basis,” Opt. Lett. 22, 685–687(1997).
1218 APPLIED OPTICS / Vol. 49, No. 8 / 10 March 2010