arxiv:1801.04506v4 [gr-qc] 10 jun 20194inter-university centre for astronomy and astrophysics, post...

8
Random projections in gravitational wave searches of compact binaries Sumeet Kulkarni, 1 Khun Sang Phukon, 2 Amit Reza, 3 Sukanta Bose, 4, 5 Anirban Dasgupta, 3 Dilip Krishnaswamy, 6 and Anand S. Sengupta 3 1 Indian Institute of Science Education and Research, Homi Bhabha Road, Pune 411008, India 2 Department of Physics, Indian Institute of Technology, Kanpur 208016, India 3 Indian Institute of Technology Gandhinagar, Gujarat 382355, India 4 Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington State University, 1245 Webster, Pullman, WA 99164-2814, U.S.A. 6 IBM Research, Bangalore 560045, India Random projection (RP) is a powerful dimension reduction technique widely used in analysis of high dimensional data. We demonstrate how this technique can be used to improve the computational efficiency of gravitational wave searches from compact binaries of neutron stars or black holes. Improvements in low-frequency response and bandwidth due to detector hardware upgrades pose a data analysis challenge in the advanced LIGO era as they result in increased redundancy in template databases and longer templates due to higher number of signal cycles in band. The RP-based methods presented here address both these issues within the same broad framework. We first use RP for an efficient, singular value decomposition inspired template matrix factorization and develop a geometric intuition for why this approach works. We then use RP to calculate approximate time-domain match correlations in a lower dimensional vector space. For searches over parameters corresponding to non-spinning binaries with a neutron star and a black hole, a combination of the two methods can reduce the total on-line computational cost by an order of magnitude over a nominal baseline. This can, in turn, help free-up computational resources needed to go beyond current spin-aligned searches to more complex ones involving generically spinning waveforms. I. INTRODUCTION The direct detections of gravitational waves (GWs) from the mergers of black holes and neutron stars [16] by Advanced LIGO (aLIGO) [7] and Advanced Virgo (AdV) [8] detectors in the first and second observing runs (O1 and O2, respectively) have launched the era of GW astronomy [9, 10]. In the coming years, the global network of ground-based detectors, comprising aLIGO, AdV, KAGRA [11] and LIGO-India [12] will not only increase the detection rate and facilitate the search for their possible electromagnetic counterparts [1315] but also produce an unprecedentedly large amount of data, which can pose an interesting computational challenge for GW data analysis. At present, theoretically modeled compact binary co- alescence (CBC) waveforms are used as templates to matched-filter [16] the detector data in these searches [17, 18]. A brute force computation of this cross-correlation with a suitable grid of templates spanning astrophysi- cal ranges of search parameters can be expensive (but see [1921]). As these detectors are paced through planned upgrades, one expects better sensitivity at low frequencies and an increase in the detector bandwidth. The combined effects of these changes will not only in- crease the volume of the search parameter space but also result in denser template banks, thereby increasing their redundancy. More cycles of the signal will fall in band and increasing their duration. These highlight the need for designing efficient and scalable methods for matched- filtering-based templated CBC searches [2227]. In a seminal work, Cannon et al. [2830] showed how singular value decomposition (SVD) can mitigate the redundancies in CBC template banks by effectively re- ducing the number of filters or templates, owing to their strong correlation for similar parameter values, with neg- ligible effect on search performance. We show, however, that the computational cost of SVD factorization does not scale favorably with increase in bank size. Further, it may not be possible to factorize very large banks in toto as it requires prohibitively large random access memory. Random Projections (RP), conceived by the pioneer- ing work of Johnson and Lindenstrauss [31], is a com- putationally efficient technique for dimension reduction and finds applications in many areas of data science [32]. In this Letter , we apply this technique to address two key challenges in future CBC searches: handling redun- dancies in large template databases; and efficiently cor- relating noisy data against long templates. The primary impact of this work is multi-fold: (1) Ef- ficient template matrix factorization can be used to ad- dress the redundancy problem. This is similar in spirit to the SVD factorization that is at the heart of the “GstLAL”-based inspiral pipeline [27, 30, 33], but our RP method scales well for very large number of tem- plates embedded in high-dimensional Euclidean space. Such factorizations can be done off-line, in advance of a CBC search. Nonetheless there can be situations when the factors need to be updated on-line, e.g., owing to the non-stationarity of data. Our adaptations will benefit both scenarios. (2) We show the explicit connection be- tween the new factorization scheme and the extant SVD method. This bridges the two approaches and makes it readily usable. (3) The computational challenges aris- ing from correlating noisy data against long templates (also known as the curse of dimensionality) is addressed by casting the match calculation in a lower-dimensional space. (4) Finally, we show that RP-based template matrix factorization and match computation in reduced dimension can be combined effectively for efficient CBC searches. Currently the GstLAL-based inspiral pipeline utilizes time-slicing of templates to improve computational effi- ciency, and also involves spin-aligned templates. Since it is for the first time that the RP is being introduced in GW searches, our primary objective here is to elucidate how its core ideas can help them. This is why we demon- strate application of RP in the simple case of a single slice of data and non-spinning inspiral templates. This simplification notwithstanding, the RP-based methods arXiv:1801.04506v4 [gr-qc] 10 Jun 2019

Upload: others

Post on 01-Sep-2021

3 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: arXiv:1801.04506v4 [gr-qc] 10 Jun 20194Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington

Random projections in gravitational wave searches of compact binaries

Sumeet Kulkarni,1 Khun Sang Phukon,2 Amit Reza,3 Sukanta Bose,4, 5

Anirban Dasgupta,3 Dilip Krishnaswamy,6 and Anand S. Sengupta3

1Indian Institute of Science Education and Research, Homi Bhabha Road, Pune 411008, India2Department of Physics, Indian Institute of Technology, Kanpur 208016, India

3Indian Institute of Technology Gandhinagar, Gujarat 382355, India4Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India

5Department of Physics & Astronomy, Washington State University, 1245 Webster, Pullman, WA 99164-2814, U.S.A.6IBM Research, Bangalore 560045, India

Random projection (RP) is a powerful dimension reduction technique widely used in analysis ofhigh dimensional data. We demonstrate how this technique can be used to improve the computationalefficiency of gravitational wave searches from compact binaries of neutron stars or black holes.Improvements in low-frequency response and bandwidth due to detector hardware upgrades pose adata analysis challenge in the advanced LIGO era as they result in increased redundancy in templatedatabases and longer templates due to higher number of signal cycles in band. The RP-basedmethods presented here address both these issues within the same broad framework. We first useRP for an efficient, singular value decomposition inspired template matrix factorization and developa geometric intuition for why this approach works. We then use RP to calculate approximatetime-domain match correlations in a lower dimensional vector space. For searches over parameterscorresponding to non-spinning binaries with a neutron star and a black hole, a combination ofthe two methods can reduce the total on-line computational cost by an order of magnitude overa nominal baseline. This can, in turn, help free-up computational resources needed to go beyondcurrent spin-aligned searches to more complex ones involving generically spinning waveforms.

I. INTRODUCTION

The direct detections of gravitational waves (GWs)from the mergers of black holes and neutron stars [1–6]by Advanced LIGO (aLIGO) [7] and Advanced Virgo(AdV) [8] detectors in the first and second observingruns (O1 and O2, respectively) have launched the era ofGW astronomy [9, 10]. In the coming years, the globalnetwork of ground-based detectors, comprising aLIGO,AdV, KAGRA [11] and LIGO-India [12] will not onlyincrease the detection rate and facilitate the search fortheir possible electromagnetic counterparts [13–15] butalso produce an unprecedentedly large amount of data,which can pose an interesting computational challengefor GW data analysis.

At present, theoretically modeled compact binary co-alescence (CBC) waveforms are used as templates tomatched-filter [16] the detector data in these searches [17,18]. A brute force computation of this cross-correlationwith a suitable grid of templates spanning astrophysi-cal ranges of search parameters can be expensive (butsee [19–21]). As these detectors are paced throughplanned upgrades, one expects better sensitivity at lowfrequencies and an increase in the detector bandwidth.The combined effects of these changes will not only in-crease the volume of the search parameter space but alsoresult in denser template banks, thereby increasing theirredundancy. More cycles of the signal will fall in bandand increasing their duration. These highlight the needfor designing efficient and scalable methods for matched-filtering-based templated CBC searches [22–27].

In a seminal work, Cannon et al. [28–30] showed howsingular value decomposition (SVD) can mitigate theredundancies in CBC template banks by effectively re-ducing the number of filters or templates, owing to theirstrong correlation for similar parameter values, with neg-ligible effect on search performance. We show, however,that the computational cost of SVD factorization doesnot scale favorably with increase in bank size. Further, itmay not be possible to factorize very large banks in toto

as it requires prohibitively large random access memory.

Random Projections (RP), conceived by the pioneer-ing work of Johnson and Lindenstrauss [31], is a com-putationally efficient technique for dimension reductionand finds applications in many areas of data science [32].In this Letter , we apply this technique to address twokey challenges in future CBC searches: handling redun-dancies in large template databases; and efficiently cor-relating noisy data against long templates.

The primary impact of this work is multi-fold: (1) Ef-ficient template matrix factorization can be used to ad-dress the redundancy problem. This is similar in spiritto the SVD factorization that is at the heart of the“GstLAL”-based inspiral pipeline [27, 30, 33], but ourRP method scales well for very large number of tem-plates embedded in high-dimensional Euclidean space.Such factorizations can be done off-line, in advance of aCBC search. Nonetheless there can be situations whenthe factors need to be updated on-line, e.g., owing to thenon-stationarity of data. Our adaptations will benefitboth scenarios. (2) We show the explicit connection be-tween the new factorization scheme and the extant SVDmethod. This bridges the two approaches and makes itreadily usable. (3) The computational challenges aris-ing from correlating noisy data against long templates(also known as the curse of dimensionality) is addressedby casting the match calculation in a lower-dimensionalspace. (4) Finally, we show that RP-based templatematrix factorization and match computation in reduceddimension can be combined effectively for efficient CBCsearches.

Currently the GstLAL-based inspiral pipeline utilizestime-slicing of templates to improve computational effi-ciency, and also involves spin-aligned templates. Sinceit is for the first time that the RP is being introduced inGW searches, our primary objective here is to elucidatehow its core ideas can help them. This is why we demon-strate application of RP in the simple case of a singleslice of data and non-spinning inspiral templates. Thissimplification notwithstanding, the RP-based methods

arX

iv:1

801.

0450

6v4

[gr

-qc]

10

Jun

2019

Page 2: arXiv:1801.04506v4 [gr-qc] 10 Jun 20194Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington

2

introduced here can be readily applied to time-sliceddata and spin-aligned templates. (A detailed study ofthat application and the computational advantage sogained will be presented in a future work.)

II. COMPACT BINARY SEARCHES

Consider a CBC search involving a bank of NT tem-plates over a given parameter space. Following the con-vention in Ref. [28], let H denote the 2NT ×Ns templatematrix with 2NT rows of real-valued unit-norm whitenedfilters, each sampled over Ns time-points. The templatematrix may be viewed as 2NT row-vectors embeddedin Ns-dimensional Euclidean space RNs . The complexmatched-filter output of the αth template at a specific

point in time, against the whitened data ~S is the innerproduct:

ρα =(H(2α−1) − iH(2α)

)~ST , (1)

where Hα denotes the αth row of H and ~ST is the trans-pose of ~S. The signal-to-noise ratio maximized over theinitial phase φ0, is given by |ρα|. In our notation, the

Hα’s and signal ~S are assumed to be row vectors. Theoverlap between two templates, when maximized overextrinsic parameters (e.g., the time t0 and phase φ0 atarrival or coalescence of the signal in band), producesthe match. The match between templates with similarintrinsic parameters (such as the compact object massesand spins), can be very high - signifying the rank defi-ciency of the template matrix. A typical off-line CBCsearch involves calculating the cross-correlation between~S and every row of H for a series of relative time-shifts,or values of t0, thereby generating a time-series of ραvalues, for every α. The use of a large number of tem-plates (NT ), each sampled over a large number of points(Ns) amplifies the search’s computational cost.

The rank deficiency of H is exploited in the truncatedSVD approach, where every row is approximated as alinear combination of only ` of the 2NT right singularvectors with the most dominant singular values. Further,these “basis” vectors are used as eigen-templates againstwhich the data are cross-correlated. The left singularvectors of H and the singular values are combined into acoefficient matrix that is used to reconstruct the approx-imate signal-to-noise ratio (SNR). The truncation of thebasis leads to errors in the approximation of the tem-plate waveforms, which further translates to imperfectreconstructions of the SNR. The fractional SNR loss canbe measured as a function of the discarded (2NT − `)singular values.

The SVD factorization of the template matrix H hasa time-complexity proportional to O(N2

TNs) assumingNT ≤ Ns. Thus, such factorizations fast become com-putationally unviable with increasing size of a templatebank. Since the entire template matrix can become toolarge to be saved in single machine memory, a suitableparallel scheme is required to apply SVD to larger banks.SVD-based on-line CBC searches [30, 33] work aroundthis problem by splitting the bank into smaller sub-banksthat are more amenable to such factorization separately.While the optimal way of partitioning the bank is anopen problem, the act of splitting the bank prevents ex-ploitation of the linear dependency of templates across

FIG. 1. The factor by which the number of SVD basisvectors increases due to partitioning of template bank of sizeNT into sub-banks of 500 templates each is shown as β onthe vertical axis. Results from six different template-banksizes are shown. For example, the bank with 4209 templatesis divided into eight sub-banks of 500 templates each anda ninth one of 209 templates. There β reaches a high of∼ 2 when one tolerates an average fractional loss in SNR of〈δρ/ρ〉 ∼ 0.1. On the other hand, for any of these templatebanks, as one approaches machine-precision accuracy in SNRreconstruction, β → 1 as expected. A practical operatingpoint would be 〈δρ/ρ〉 ∼ 10−3. The trend from the sixexamples shown here indicates that β can be quite large forsearches in aLIGO data where NT ∼ 105.

the sub-banks. This is seen in Fig. 1, where we plotβ, which is defined as the ratio of the number of basisvectors summed across all the sub-banks to the numberof basis vectors from the SVD factorization of the fullbank, at a given average fractional loss in accuracy ofthe reconstructed SNR. By splitting the bank, one ef-fectively ends up requiring many more eigen-templatesagainst which the data are filtered. When extrapolatedto realistic template bank sizes of NT ≈ 106, β can beas large as ∼ 102 at 〈 δρρ 〉 = 10−3.

The SVD-inspired RP-based factorization presentedbelow addresses this issue and is scalable for large tem-plate banks. We also apply RP to calculate the matchcorrelations in a lower-dimensional space Rk wherek ≤ Ns. These correlations could be either betweentemplates or between basis vectors within the SVDparadigm. The full potential of the RP-based meth-ods introduced here can be realized by combining themtogether. We demonstrate its feasibility with an exam-ple.

III. RANDOM PROJECTION

The core theoretical idea behind the RP techniqueis the Johnson-Lindenstrauss (JL) lemma [31], whichstates that a set of 2NT vectors in RNs can be mappedinto a randomly generated subspace R` of dimension` ∼ O

(log(2NT ) /ε2

)or greater, while preserving all

pairwise L2 norms to within a factor of (1 ± ε), where0 < ε < 1, with a very high probability. Here, ε isthe mismatch or distortion tolerated in the pairwise L2

norms between any two filters after projection. Thus,

Page 3: arXiv:1801.04506v4 [gr-qc] 10 Jun 20194Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington

3

RP also approximately preserves any statistic of thedataset that is characterized by such pairwise distances.The RP of H onto R` produces HΩ; the accuracy ofthis data-oblivious transformation depends on the tar-get dimensions and sampling distribution of the Ns × `projection matrix Ω. While it is enough to sample theentries independently and identically distributed froma sub-Gaussian distribution, here we choose them in-dependently from a Gaussian distribution with meanzero and variance 1/`, i.e., N (0, 1/`), thus producinga Gaussian quasi-orthonormal random matrix [34, 35]such that 〈ΩΩT 〉 = I. Results obtained from RP-basedprocessing can vary depending on the actual choice ofthe distribution (from which elements of Ω are drawn),and in a statistical sense, these results arising from dif-ferent choices of Ω are expected to be equivalent due tothe quasi-orthonormality of the projection. (See Supple-mental Material for a geometric explanation.)

IV. RP-BASED TEMPLATE MATRIXFACTORIZATION

The key idea behind an `-truncated SVD approxima-tion of H is to reconstruct the rows of the templatematrix using the top-` right-singular vectors. This ap-proximation works well because H has a fast-decayingspectrum, as shown in Fig. 2. In making the trunca-tion, one effectively reduces H to its `-rank approxi-mation H(`) [36, 37]. 1 Further, for a bank of nor-malized templates, it is easy to show that the averagefractional loss in SNR due to the truncation is given as〈δρ/ρ〉 ≤ ‖H−H(`)‖2F / ‖H‖2F , where 〈 〉 denotes averageover the bank of templates and ‖ · ‖F is the Frobeniusnorm [36]. However, the existing SVD algorithms do notscale well with increasing dimensions and redundanciesof the template database.

Randomized-SVD (RSVD) [37] is a RP-based matrix-factorization technique to obtain an `-rank matrix fac-torization H(`) such that, for some specified η > 0,‖H−H(`)‖F ≤ minX : rank(X)≤` ‖H−X‖F (1+η) with

high probability. 2 In one implementation, the RSVDalgorithm proceeds by first projecting the individualrow-vectors in the template matrix H to R` by usingΩNs×` ∈ N (0, 1/`), thereby yielding H2NT×` = H Ω.The latter can be used to perform an SVD-like factor-ization directly in R` through a series of operations likethe ones described below. In passing, we note that whileH is an object in a lower-dimensional Euclidean spacerelative to H, it is not constituted of time-decimatedtemplates.

Figure 2 compares the singular values obtained by theRP-based factorization against those from a direct SVDfactorization. As seen there, it is typically sufficient totake ` NT . (Since NT ≤ Ns, as mentioned above,it follows that ` Ns as well.) In fact, the numericalvalue of ` chosen in RSVD may be smaller than thetheoretical JL bound prescribed for preserving pairwiseL2 distances between the 2NT rows to a ε-distortionfactor. Working with the reduced sized matrix H leads

1 Note that H(`) has the same dimensions as H.2 Note that the values of η and ε can be different.

FIG. 2. Comparison of singular values σ for a templatematrix H of size (2NT × Ns) ≡ 9130 × 65536, normalizedby the maximum singular value σmax, as obtained fromSVD and RSVD factorization. RSVD is performed in tar-get dimensions R` where ` = 200, 4000 or 8000. The spec-trum of eigenvalues is seen to fall steeply. This exampletemplate bank was constructed using non-spinning signalmodel for component mass parameters (m1,2) in the range2.5M ≤ m1,m2 ≤ 17.5M. As seen here, the top-` eigen-values obtained by RSVD agree very well with the spectrumobtained by traditional SVD factorization.

to significant computational savings, while producinga decomposition that closely approximates the optimal`-rank factorization of H. The optimum choice of `depends on the shape of the eigenvalue spectrum. Inthe Monte-Carlo simulations presented in SupplementalMaterial, we choose ` = 200. The corresponding averageSNR loss for a set of 500 CBC signals added to simulatedaLIGO noise is 〈δρ/ρ〉 = 2 × 10−4 in that study (seeFig. S3 in Supplemental Material).

RSVD thus proceeds by obtaining a set of orthogo-nal bases for the column space of H by using a thin-QR decomposition[36]: H = Q R, where Q is an or-thonormal matrix with dimensions 2NT × `. The ap-proximate rank-` decomposition is then obtained asH(`) = Q (QTH) = Q B, where B`×Ns ≡ QTH is amatrix that defines the orthonormal projection of thetemplate waveforms into the compressed subspace. It isclear that one can use the ` rows of B as the surrogatetemplates, which in turn can be used to correlate against

the detector data ~S. These can be further combined withQ to reconstruct ρ in RNs . We can thus use the QB de-composition itself to improve the efficiency of both thetime and frequency domain searches by constructing Happropriately, with templates from the correspondingdomains.

Instead of randomly projecting the column space of H,the method can be generalized by applying RP on boththe row and column spaces [37]. This bilateral RSVDmethod is particularly useful when both Ns and NT arevery large.

Page 4: arXiv:1801.04506v4 [gr-qc] 10 Jun 20194Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington

4

V. RECONSTRUCTION OF SNR

The rank-` matrix factorization of H using RSVD isgiven by H(`) = QB. Thus, the SNR ρ′α, for any givent0, can be reconstructed in RNs as

ρ′α =(H

(`)(2α−1) − iH

(`)(2α)

)~ST

=∑ν=1

(Q(2α−1)ν − iQ(2α)ν

) (Bν ~S

T). (2)

Using Pythagoras theorem, and the fact that ||H||2F =2NT , it is easy to show that the average fractional lossof SNR is given by⟨

δρ

ρ

⟩≤ ||H||

2F − ||H(`)||2F||H||2F

= 1−∑`µ=1 σ

2NT, (3)

where σµ are the eigenvalues of H(`). For the example

discussed in Fig. 2,∑`µ=1 σ

2µ/(2NT ) < 1 but approaches

unity monotonically with increasing `. The right-handside of Eq. (3) can be calculated efficiently by evalu-ating the Frobenius norm of B directly (i.e., withoutexplicitly finding the eigenvalues of H(`) first). Thus,the QB decomposition can indeed serve as a stand-inreplacement for the SVD factorization. (For an efficientmethod of explicitly calculating the SVD factors fromthe RP-based factorization see Supplemental Material.)

Ideally one would like to use 〈δρ/ρ〉 as the controlparameter and solve Eq. (3) for the optimum value of `.However, this is a hard problem and in practice the valueis set by a process of trial and error, which thankfully canbe done off-line even when the computation in Eq. (2)is conducted on-line.

A naive implementation of matched-filter in time-domain can be very expensive, with a complexity ofO(N2

s ) per template for Ns time-shifts. Of course, theFast Fourier transform can reduce this to O(Ns logNs).It is however more efficient instead to first project thetwo aforementioned whitened time-series vectors in RNs

to a random k-dimensional (k Ns) subspace and thencalculate the match (using circular cross-correlations),as seen in Fig. S2 of Supplemental Material. In fact, forthe template part, one can directly project the rows ofthe B matrix (which serve as surrogate templates) to Rk

(k ≤ Ns). In this context, RP reduces the complexity ofcalculating the matches by a factor Ns/k. (See Supple-mental Material for how FFT-like algorithms enable itsfast computation [36].)

VI. COMPUTATIONAL COMPLEXITYANALYSIS

The straightforward SVD factorization of H requiresO(N2

T Ns) floating-point operations, assuming NT ≤ Ns.In comparison, the cost of the RP matrix factorizationis O

(`NTNs + (`2NT − 2

3`3) + `NTNs + `Ns

). In this

last expression, we have included partial contributionsfrom first projecting the template matrix to R`, then tak-ing the thin-QR decomposition of H using Householder’smethod [38], followed by the cost of constructing B andcalculating its Frobenius norm, respectively. For prac-tical cases, one expects ` Ns, due to which the costof factorizing H can be orders of magnitude less thana full SVD factorization. This advantage is not justrealized off-line, but can also directly impact the totalon-line cost of the searches owing to a lower value of` alone: Figure 1 shows that for moderate sized banksone effectively ends up using ∼ 3− 4 times fewer surro-gate templates in the on-line portion of the search fromthe new RP-based factorization. This improvement isexpected to be higher for larger banks.

For on-line searches, the number of floating point op-erations per second (flops) in our method is Nflops =(2` kfs + 2`NT fs + k fs). The first term is the numberof floating-point operations required for computing thecross-correlation between the surrogate templates (rowsof B) and the data vector; the second term is the cost ofreconstruction of the SNR for every template; and thethird term is the cost of projecting the data vector intothe lower-dimensional space. In the SVD-only method,the expression for Nflops is analogous, except that in-stead of the last term above, it has a down-samplingcost that is similarly insignificant as our projection cost.The primary difference between the two methods is thatowing to our use of RSVD and RP, ` and k are lessthan the number of basis templates and the number oftime samples of data used, respectively, in the SVD-onlymethod. For the crucial last couple of seconds of thecross-correlation analysis for CBC signals we have eval-uated that our method is an order of magnitude fasterthan the SVD-only method.

VII. CONCLUSION

In summary, here we introduced random projection-based techniques that hold promise for factorization oflarge template matrices and cross-correlation of tem-plates in a scalable and computationally efficient way,which can aid more complex searches, such as of CBCswith generic spins, and, hence, improve the chances fornew discoveries.

ACKNOWLEDGMENTS

We would like to thank Surabhi Sachdev for carefullyreading the manuscript and making useful comments.This work is supported in part by DST’s SERB grantsEMR/2016/007593 and DST/ICPS/CLUSTER/DataScience/General/T-150, NSF grant PHY-1506497, andthe Navajbai Ratan Tata Trust. A large set of data anal-ysis studies were performed on the Sarathi computingcluster at IUCAA.

Page 5: arXiv:1801.04506v4 [gr-qc] 10 Jun 20194Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington

5

Supplemental Material:Random projections in gravitational wave searches of compact binaries

Sumeet Kulkarni,1 Khun Sang Phukon,2 Amit Reza,3 Sukanta Bose,4,5

Anirban Dasgupta,3 Dilip Krishnaswamy,6 and Anand S. Sengupta,3

1Indian Institute of Science Education and Research, Homi Bhabha Road, Pune 411008, India2Department of Physics, Indian Institute of Technology, Kanpur 208016, India

3Indian Institute of Technology Gandhinagar, Gujarat 382355, India4Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India

5Department of Physics & Astronomy, Washington State University, 1245 Webster, Pullman, WA 99164-2814, U.S.A.6IBM Research, Bangalore 560045, India

The purpose of this document is three-fold. First,we provide a geometric explanation for why RSVD pre-serves the top singular subspaces. Second, we present anefficient method for explicitly calculating the SVD fac-tors from the RP-based factorization. Third, we demon-strate how random projection can be used to computethe match between two normalised templates directly inthe target space.

I. GEOMETRIC EXPLANATION FOR TOPSINGULAR SUBSPACE PRESERVATION

UNDER RSVD.

Here we provide a geometric explanation behind thestatement that RSVD preserves the top singular sub-spaces. Figure S1 depicts this intuition.

We denote the L2 norm of a vector as ‖ · ‖. Recallthat given H, finding out the top-k singular vectors isakin to the question of finding out orthogonal directionsu1, . . . uk such that if U = [u1, . . . , uk], then

‖UUTH‖2F = ‖UTH‖2F =

k∑i=1

‖uTi H‖2 (S1)

is maximized (squared lengths of vectors in subfigure(a)).

Given the length-preserving properties of random pro-jection, the sum of the squared lengths of uTi H is equiv-alent, up to an approximation factor of 1± ε, to the sumof squared lengths of (uTi H)Ω ((subfigure (b)), i.e.,

(1− ε)∑i≤k

‖uTi H‖2 ≤∑i≤k

‖(uTi H)Ω‖2.

Hence, we instead find orthonormal vectors a1, . . . , akthat are the top-k left singular directions of HΩ, suchthat

∑i≤k ‖aTi (HΩ)‖2 is maximum (subfigure (c)), thus

achieving∑i≤k ‖aTi (HΩ)‖2 ≥

∑i≤k ‖uTi (HΩ)‖2.

Again by using the length preservation of randomprojection, we have that ‖aTi H‖2 ≥ (1 − ε)‖(aTi H)Ω‖2(subfigure (d)). Putting the above steps together, wefind

k∑i=1

‖aTi H‖2 ≥ (1− ε)k∑i=1

‖(aTi H)Ω‖2

≥ (1− ε)k∑i=1

‖(uTi H)Ω‖2

≥ (1− ε)2k∑i=1

‖(uTi H)‖2 . (S2)

Recall that by definition of singular vectors, we alreadyhave

k∑i=1

‖aTi H‖2 ≤k∑i=1

‖uTi H‖2. (S3)

Combining Eqs. (S2) and (S3), one can obtain the fol-lowing inequality:

(1− ε2

)≤∑ki=1 ‖aTi H‖2∑ki=1 ‖uTi H‖2

≤ 1 . (S4)

Let A = [a1, . . . , ak] and recall that U = [u1, . . . , uk].Given that U is an orthonormal matrix, the matrixUUTH represents the projection of columns of H ontothe subspace spanned by the columns of U, and henceis a rank-k approximation of H. Using an argumentsimilar to that used in Eq. (S1) it is possible to obtain

‖AATH‖2F =

k∑i=1

‖aTi H‖2 . (S5)

Therefore, using Eqs. (S1), (S4) and (S5), it is clear that

the low rank approximation AATH found by RSVDcaptures most of the energy in the optimal rank-k ap-proximation UUTH.

II. CALCULATING THE SVD FACTORSFROM THE RP-BASED FACTORIZATION

EFFICIENTLY

An efficient method of explicitly calculating the SVDfactors from the RP-based factorization is now presented.This is intended as a bridge between the two methods.Singular values can be obtained by performing an SVDon B, or by first calculating TB = BBT , of size ` × `.The eigenvectors UTB

of TB are identical to the left-singular vectors of B, and the eigenvalues ΣTB

are equalto the squares of the singular values of B. As TB is amuch more compressed matrix compared to B, it is farmore efficient to store it in memory and factorize itthereby revealing the singular values and left-singularvectors. These in turn can be further used to calculatethe singular values of H(`). The top-` right-singular vec-tors of H in RNs can be obtained using the left-singular

vectors of TB: U(`)H ≈ Q UTB

. Similarly, it can betrivially checked that ΣHVT

H = UTTB

B. Thus, all thepieces of the SVD factorization of H can be recoveredfrom RSVD factors, but at a small fraction of the compu-tational cost of the former. In doing so, the advantagesof RP-based factorization can be directly transferred tothe current SVD-based data analysis pipelines.

Page 6: arXiv:1801.04506v4 [gr-qc] 10 Jun 20194Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington

6

FIG. S1. Depiction of the geometrical intuition behind the preservation of the top singular subspaces in RSVD. Findingthe top-k singular vectors for a given H is akin to the finding the orthogonal directions u1, . . . uk such that the sum oflengths of the vectors uT

i H is maximized. Subfigure (a) shows these vectors in the original feature space. Subfigure (b)depicts the action of random projection of these vectors to a lower-dimensional space using the projection matrix Ω. Thelength-preserving properties of random projections guarantee that the distortion in their lengths lies within a factor of 1± ε.Subfigure (c) depicts the action of HΩ along its top-k left singular directions a1, . . . , ak, such that

∑i≤k ‖a

Ti (HΩ)‖2 is

maximum. Note that these vectors aTi (HΩ), can also be geometrically interpreted as the the random projection of the vectorsaTi H (as shown in subfigure (d) using Ω. Through this chain of subfigures, it follows that the low rank approximation AATHfound by RSVD captures most of the energy in the optimal rank-k approximation UUTH.

III. RANDOM PROJECTION BASEDCORRELATIONS

A naive implementation of matched-filter in time-domain can be very expensive, with a complexity ofO(N2

s ) per template. Here we show how random pro-jections can be applied to reduce this cost considerably:the whitened time-series vectors, in the form of the tem-plate and the data, in RNs can be first projected toa random k-dimensional (k Ns) subspace and thencross-correlated; the pairwise distance-preserving prop-erty of RP guarantees that the matched-filter output ρ′αin Rk will be approximately equal to ρα in RNs , i.e.,

〈ρ′α〉 =

⟨(H(2α−1) Ω− iH(2α) Ω

) (~SΩ)T⟩

= ρα ,

because 〈ΩΩT 〉 = I.It can also help searches to use the RP-based correla-

tion in conjunction with the RP-based QB factorizationof the template matrix described above. The key to thisfusion between the two RP-based methods lies in the factthat instead of matched-filtering the data ~S against ev-ery template in the bank, one can use the reduced set of` ≤ 2NT row vectors of B as surrogate templates for thispurpose. These correlations can be calculated in Rk by

projecting ~S and each row vector Bν ∈ RNs to the tar-get k-dimensional subspace. Such projections preservethe inner products between the data and Bν at every rel-

FIG. S2. The phase-maximized overlap time series of a nor-malized template (corresponding to equal component massesm1,2 = 6 M) correlated against a copy of itself time shiftedby 4 seconds. The template is taken to be 8 seconds long andsampled at 2048 Hz. The circular correlations are calculatedboth directly in a high dimensional space RNs and using RPin a lower dimensional Euclidean space Rk, as discussed inSec. III; here Ns = 16384 and k = Ns/2, Ns/8. The agree-ment between the two traces shows that RP can be used toefficiently calculate overlaps in a lower dimensional space.

Page 7: arXiv:1801.04506v4 [gr-qc] 10 Jun 20194Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington

7

ative time-shift within the ε bound, as guaranteed by theJL lemma. The complex SNR for each template can bereconstructed using the coefficient matrix CNT×`, whoseelements are Cα ≡

(Q(2α−1)ν − iQ(2α)ν

)as shown in the

main text. The phase-maximized SNRs of the templatesare given by the modulus of resulting complex SNRs.

The construction proceeds as follows: Suppose dataare sampled at a rate fs and that the duration in whichone decides to search for the signal’s time of arrival is τ ,which is taken to be longer than the longest template inthe bank. Then the number of points over which one isdiscretely searching for t0 is fsτ . We next construct apartial circulant matrix K(Bν) for every row of B, suchthat its dimensions are (fsτ)× (fsτ +Ns). Its nth rowKn(Bν) is a copy of Bν that is time-shifted by an amount∆tn = n/fs, where n ∈ [0, (fsτ)]. The remaining fsτ

elements in each row are set to zero. The data vector ~S,with (fsτ +Ns) time-points, and the circulant matricescan both be randomly projected to the subspace Rkusing Ω. Their subsequent multiplication is used toconstruct the cross-correlation:

ρ′α(∆tn) =∑ν

Cαν (Kn(Bν) Ω) (~SΩ)T , (S6)

where α is the index over the templates in the bankν = 1, . . . , ` is the index on the rows of B, and Ω hasdimensions of (fsτ + Ns) × `. A circulant matrix canbe diagonalized using FFT-like algorithms to enable effi-cient processing of matrix-vector products involving suchmatrices [36]. Figure S2 compares the phase-maximizedoverlap computed using this method for two choices ofk with that computed directly, i.e., without employingrandom projections.

FIG. S3. The distribution of averaged SNR loss summariz-ing the results of the Monte-Carlo injection study presentedin Sec. IV. The mean of the distribution (solid red verticalline) is close to the theoretically expected mean loss (dottedblack vertical line). The similarity of the results of this studywith a similar study presented in Ref. [28] establishes the va-lidity of efficient RP-based QB factorization of the templatebanks presented in this work.

IV. SNR RECONSTRUCTION USING QBDECOMPOSITION OF THE TEMPLATE BANK

We have shown above that very large template bankscan be efficiently QB decomposed using the RSVD algo-rithm thereby representing the template matrix by itsrank-` approximation. One can reconstruct the SNRtime series for each template in this bank to a high de-gree of accuracy by projecting the data on the top ` basisvectors, akin to the truncated SVD paradigm. The aver-age SNR loss can be estimated from the singular valuescorresponding to the discarded basis vectors.

We now present results from a Monte-Carlo study toexplicitly demonstrate that the fractional SNR loss (av-eraged over the bank) due to the rank-` approximationof the template matrix closely follows the theoreticallyestimated value, as evaluated using the expression forδρ/ρ in Sec. IV of main text, thereby validating theaccuracy of the RSVD factorization.

We consider a template bank H containing NT =581 templates covering the component mass space: 5 ≤m1,2/M ≤ 15 using non-spinning TaylorT4 waveforms.For this study, each waveform was taken to be 8 secondslong, sampled at 2048 Hz, thereby setting Ns = 16384.We use the aLIGO Zero Detuned High Power (ZDHP)noise power spectral density [39]. Signals were simulatedfor 500 CBC sources, with component masses randomlychosen from the aforementioned mass range. These wereseparately added to colored Gaussian noise with aLIGOZDHP power spectral density. The amplitudes of theinjected signals were adjusted for a target SNR of 8. Themass parameters of most of these signals were differentfrom those of the templates in H.

The template matrix was first QB decomposed to arank ` = 200 approximation using the RSVD algorithmthat corresponded to an averaged SNR loss 〈δρ/ρ〉 =∑2NT

`+1 σ2i∑2NT

1 σ2i

= 2× 10−4. This threshold was decided based

on the spectrum of the singular values of H, which wasobserved to fall sharply – similar to the examples shownin Fig. 2. of the main text.

For each simulated signal injection, the SNR for everytemplate was reconstructed using the 200 basis vectorsand compared with the SNR calculated from a directcircular correlation of these templates against the noisydata containing the injection. Thereafter the averagedSNR loss was evaluated. The distribution of this quan-tity over the set of all injections is shown in Fig. S3. Asshown there, the mean of the distribution agrees wellwith the target set at 2×10−4. Note that the correlationswere computed in RNs .

A similar study was presented by Cannon et al. [28]using truncated SVD factorization of the template ma-trix. The similarity of the results establishes the validityof efficient QB decomposition of large template banksafter random projection.

[1] B. P. Abbott et al. (Virgo, LIGO Scientific), Phys. Rev.Lett. 116, 061102 (2016), arXiv:1602.03837 [gr-qc].

[2] B. P. Abbott et al. (Virgo, LIGO Scientific), Phys. Rev.

Page 8: arXiv:1801.04506v4 [gr-qc] 10 Jun 20194Inter-University Centre for Astronomy and Astrophysics, Post Bag 4, Ganeshkhind, Pune 411 007, India 5 Department of Physics & Astronomy, Washington

8

Lett. 116, 241103 (2016), arXiv:1606.04855 [gr-qc].[3] B. P. Abbott et al. (Virgo, LIGO Scientific), Phys. Rev.

X6, 041015 (2016), arXiv:1606.04856 [gr-qc].[4] B. P. Abbott et al. (VIRGO, LIGO Scientific), Phys. Rev.

Lett. 118, 221101 (2017), arXiv:1706.01812 [gr-qc].[5] B. P. Abbott et al. (Virgo, LIGO Scientific), Phys. Rev.

Lett. 119, 141101 (2017), arXiv:1709.09660 [gr-qc].[6] B. P. Abbott et al. (Virgo, LIGO Scientific), Phys. Rev.

Lett. 119, 161101 (2017), arXiv:1710.05832 [gr-qc].[7] J. Aasi et al. (LIGO Scientific), Class. Quant. Grav. 32,

074001 (2015), arXiv:1411.4547 [gr-qc].[8] F. Acernese et al. (VIRGO), Class. Quant. Grav. 32,

024001 (2015), arXiv:1408.3978 [gr-qc].[9] B. P. Abbott et al. (Virgo, Fermi-GBM, INTEGRAL,

LIGO Scientific), Astrophys. J. 848, L13 (2017),arXiv:1710.05834 [astro-ph.HE].

[10] B. P. Abbott et al., Astrophys. J. 848, L 12 (2017),arXiv:1710.05833.

[11] T. Akutsu et al. (KAGRA) (2017) arXiv:1710.04823 [gr-qc].

[12] C. S. Unnikrishnan, Int. J. Mod. Phys. D22, 1341010(2013), arXiv:1510.06059 [physics.ins-det].

[13] J. Rana, A. Singhal, B. Gadre, V. Bhalerao, and S. Bose,Astrophys. J. 838, 108 (2017), arXiv:1603.01689 [astro-ph.IM].

[14] V. Srivastava, V. Bhalerao, A. P. Ravi, A. Ghosh, andS. Bose, Astrophys. J. 838, 46 (2017), arXiv:1610.07154[astro-ph.HE].

[15] S. Ghosh and S. Bose, (2013), arXiv:1308.6081 [astro-ph.HE].

[16] C. W. Helstrom, Elements of Signal Detection and Es-timation (Prentice-Hall, Inc., Upper Saddle River, NJ,USA, 1995).

[17] B. J. Owen, Phys. Rev. D53, 6749 (1996), arXiv:gr-qc/9511032 [gr-qc].

[18] B. S. Sathyaprakash and S. V. Dhurandhar, Phys. Rev.D44, 3819 (1991).

[19] R. Prix, Gravitational wave data analysis. Proceed-ings: 11th Workshop, GWDAW-11, Potsdam, Germany,Dec 18-21, 2006, Class. Quant. Grav. 24, S481 (2007),arXiv:0707.0428 [gr-qc].

[20] I. W. Harry, B. Allen, and B. S. Sathyaprakash, Phys.Rev. D80, 104014 (2009), arXiv:0908.2090 [gr-qc].

[21] S. Roy, A. S. Sengupta, and N. Thakor, Phys. Rev.D95, 104045 (2017), arXiv:1702.06771 [gr-qc].

[22] T. Cokelaer, Phys. Rev. D76, 102004 (2007),arXiv:0706.4437 [gr-qc].

[23] B. Abbott et al. (LIGO Scientific), Phys. Rev. D78,042002 (2008), arXiv:0712.2050 [gr-qc].

[24] I. W. Harry, A. H. Nitz, D. A. Brown, A. P. Lundgren,E. Ochsner, and D. Keppel, Phys. Rev. D89, 024010(2014), arXiv:1307.3562 [gr-qc].

[25] S. Babak, Class. Quant. Grav. 25, 195011 (2008),arXiv:0801.4070 [gr-qc].

[26] G. M. Manca and M. Vallisneri, Phys. Rev. D81, 024004(2010), arXiv:0909.0563 [gr-qc].

[27] S. Privitera, S. R. P. Mohapatra, P. Ajith, K. Can-non, N. Fotopoulos, M. A. Frei, C. Hanna, A. J. We-instein, and J. T. Whelan, Phys. Rev. D89, 024003(2014), arXiv:1310.5633 [gr-qc].

[28] K. Cannon, A. Chapman, C. Hanna, D. Keppel, A. C.Searle, and A. J. Weinstein, Phys. Rev. D82, 044025(2010), arXiv:1005.0012 [gr-qc].

[29] K. Cannon, C. Hanna, and D. Keppel, Phys. Rev. D84,084003 (2011), arXiv:1101.4939 [gr-qc].

[30] K. Cannon et al., Astrophys. J. 748, 136 (2012),arXiv:1107.2665 [astro-ph.IM].

[31] W. B. Johnson and J. Lindenstrauss, in Conference inmodern analysis and probability (New Haven, Conn.,1982), Contemp. Math., Vol. 26 (Amer. Math. Soc.,Providence, RI, 1984) pp. 189–206.

[32] E. Bingham and H. Mannila, in Proceedings of the sev-enth ACM SIGKDD international conference on Knowl-edge discovery and data mining (ACM, 2001) pp. 245–250.

[33] C. Messick et al., Phys. Rev. D95, 042001 (2017),arXiv:1604.04324 [astro-ph.IM].

[34] S. Dasgupta and A. Gupta, Random Struct. Algorithms22, 60 (2003).

[35] S. Dasgupta, in Proceedings of the Sixteenth confer-ence on Uncertainty in artificial intelligence (Mor-gan Kaufmann Publishers Inc., 2000) pp. 143–151,arXiv:1301.3849 [cs.LG].

[36] G. H. Golub and C. F. Van Loan, Matrix Computations(3rd Ed.) (Johns Hopkins University Press, Baltimore,MD, USA, 1996).

[37] N. Halko, P.-G. Martinsson, and J. A. Tropp, SIAMreview 53, 217 (2011), arXiv:0909.4061 [math.NA].

[38] W. H. Press, S. A. Teukolsky, W. T. Vetterling, andB. P. Flannery, Numerical Recipes 3rd Edition: The Artof Scientific Computing, 3rd ed. (Cambridge UniversityPress, New York, NY, USA, 2007).

[39] Advanced LIGO anticipated sensitivity curves, Tech.Rep. LIGO-T0900288-v3 (LIGO Scientific Collaboration,https://dcc.ligo.org/LIGO-T0900288/public, 2010).