381 Publications

A boundary integral equation approach to computing eigenvalues of the stokes operator

Travis Askham, M. Rachh

The eigenvalues and eigenfunctions of the Stokes operator have been the subject of intense analytical investigation and have applications in the study and simulation of the Navier–Stokes equations. As the Stokes operator is second order and has the divergence-free constraint, computing these eigenvalues and the corresponding eigenfunctions is a challenging task, particularly in complex geometries and at high frequencies. The boundary integral equation (BIE) framework provides robust and scalable eigenvalue computations due to (a) the reduction in the dimension of the problem to be discretized and (b) the absence of high-frequency “pollution” when using Green’s function to represent propagating waves. In this paper, we detail the theoretical justification for a BIE approach to the Stokes eigenvalue problem on simply- and multiply-connected planar domains, which entails a treatment of the uniqueness theory for oscillatory Stokes equations on exterior domains. Then, using well-established techniques for discretizing BIEs, we present numerical results which confirm the analytical claims of the paper and demonstrate the efficiency of the overall approach.

Show Abstract

High-order discretization of a stable time-domain integral equation for 3D acoustic scattering

A. Barnett, L. Greengard, Thomas Hagstrom

We develop a high-order, explicit method for acoustic scattering in three space dimensions based on a combined-field time-domain integral equation. The spatial discretization, of Nyström type, uses Gaussian quadrature on panels combined with a special treatment of the weakly singular kernels arising in near-neighbor interactions. In time, a new class of convolution splines is used in a predictor-corrector algorithm. Experiments on a torus and a perturbed torus are used to explore the stability and accuracy of the proposed scheme. This involved around one thousand solver runs, at up to 8th order and up to around 20,000 spatial unknowns, demonstrating 5-9 digits of accuracy. In addition we show that parameters in the combined field formulation, chosen on the basis of analysis for the sphere and other convex scatterers, work well in these cases.

Show Abstract

Measurement of Void Bias Using Separate Universe Simulations

Kwan Chuen Chan, Y. Li, Matteo Biagetti, Nico Hamaus

Cosmic voids are biased tracers of the large-scale structure of the universe. Separate universe simulations (SUS) enable accurate measurements of this biasing relation by implementing the peak-background split (PBS). In this work, we apply the SUS technique to measure the void bias parameters. We confirm that the PBS argument works well for underdense tracers. The response of the void size distribution depends on the void radius. For voids larger (smaller) than the size at the peak of the distribution, the void abundance responds negatively (positively) to a long wavelength mode. The linear bias from the SUS is in good agreement with the cross power spectrum measurement on large scales. Using the SUS, we have detected the quadratic void bias for the first time in simulations. We find that $ b_2 $ is negative when the magnitude of $ b_1 $ is small, and that it becomes positive and increases rapidly when $ |b_1| $ increases. We compare the results from voids identified in the halo density field with those from the dark matter distribution, and find that the results are qualitatively similar, but the biases generally shift to the larger voids sizes.

Show Abstract

Reducing bias and variance for CTF estimation in single particle cryo-EM

Ayelet Heimowitz, J. Andén, Amit Singer

When using an electron microscope for imaging of particles embedded in vitreous ice, the recorded image, or micrograph, is a significantly degraded version of the tomographic projection of the sample. Apart from noise, the image is affected by the optical configuration of the microscope. This transformation is typically modeled as a convolution with a point spread function. The Fourier transform of this function, known as the contrast transfer function (CTF), is oscillatory, attenuating and amplifying different frequency bands, and sometimes flipping their signs. High-resolution reconstruction requires this CTF to be accounted for, but as its form depends on experimental parameters, it must first be estimated from the micrograph. We present a new method for CTF estimation based on multitaper techniques that reduce bias and variance in the estimate. We also use known properties of the CTF and the background power spectrum to further reduce the variance through background subtraction and steerable basis projection. We show that the resulting power spectrum estimates better capture the zero-crossings of the CTF and yield accurate CTF estimates on several experimental micrographs.

Show Abstract
January 29, 2020

Factorization of the translation kernel for fast rigid image alignment

An important component of many image alignment methods is the calculation of inner products (correlations) between an image of $n\times n$ pixels and another image translated by some shift and rotated by some angle. For robust alignment of an image pair, the number of considered shifts and angles is typically high, thus the inner product calculation becomes a bottleneck. Existing methods, based on fast Fourier transforms (FFTs), compute all such inner products with computational complexity $\mathcal{O}(n^3 \log n)$ per image pair, which is reduced to $\mathcal{O}(N n^2)$ if only $N$ distinct shifts are needed. We propose to use a factorization of the translation kernel (FTK), an optimal interpolation method which represents images in a Fourier--Bessel basis and uses a rank-$H$ approximation of the translation kernel via an operator singular value decomposition (SVD). Its complexity is $\mathcal{O}(Hn(n + N))$ per image pair. We prove that $H = \mathcal{O}((W + \log(1/\epsilon))^2)$, where $2W$ is the magnitude of the maximum desired shift in pixels and $\epsilon$ is the desired accuracy. For fixed $W$ this leads to an acceleration when $N$ is large, such as when sub-pixel shift grids are considered. Finally, we present numerical results in an electron cryomicroscopy application showing speedup factors of $3$-$10$ with respect to the state of the art.

Show Abstract

Cryo-EM reconstruction of continuous heterogeneity by Laplacian spectral volumes

Amit Moscovich, Amit Halevi, J. Andén, Amit Singer

Single-particle electron cryomicroscopy is an essential tool for high-resolution 3D reconstruction of proteins and other biological macromolecules. An important challenge in cryo-EM is the reconstruction of non-rigid molecules with parts that move and deform. Traditional reconstruction methods fail in these cases, resulting in smeared reconstructions of the moving parts. This poses a major obstacle for structural biologists, who need high-resolution reconstructions of entire macromolecules, moving parts included. To address this challenge, we present a new method for the reconstruction of macromolecules exhibiting continuous heterogeneity. The proposed method uses projection images from multiple viewing directions to construct a graph Laplacian through which the manifold of three-dimensional conformations is analyzed. The 3D molecular structures are then expanded in a basis of Laplacian eigenvectors, using a novel generalized tomographic reconstruction algorithm to compute the expansion coefficients. These coefficients, which we name spectral volumes, provide a high-resolution visualization of the molecular dynamics. We provide a theoretical analysis and evaluate the method empirically on several simulated data sets.

Show Abstract

Perspective maximum likelihood-type estimation via proximal decomposition

Patrick L Combettes, C. Müller

We introduce a flexible optimization model for maximum likelihood-type estimation (M-estimation) that encompasses and generalizes a large class of existing statistical models, including Huber’s concomitant M-estimator, Owen’s Huber/Berhu concomitant estimator, the scaled lasso, support vector machine regression, and penalized estimation with structured sparsity. The model, termed perspective M-estimation, leverages the observation that convex M-estimators with concomitant scale as well as various regularizers are instances of perspective functions, a construction that extends a convex function to a jointly convex one in terms of an additional scale variable. These nonsmooth functions are shown to be amenable to proximal analysis, which leads to principled and provably convergent optimization algorithms via proximal splitting. We derive novel proximity operators for several perspective functions of interest via a geometrical approach based on duality. We then devise a new proximal splitting algorithm to solve the proposed M-estimation problem and establish the convergence of both the scale and regression iterates it produces to a solution. Numerical experiments on synthetic and real-world data illustrate the broad applicability of the proposed framework.

Show Abstract
January 8, 2020

Cosmological simulation in tides: power spectrum and halo shape responses, and shape assembly bias

Kazuyuki Akitsu, Y. Li, Teppei Okumura

The well-developed separate universe technique enables accurate calibration of the response of any observable to an isotropic long-wavelength density fluctuation. The large-scale environment also hosts tidal modes that perturb all observables anisotropically. As in the separate universe, both the long tidal and density modes can be absorbed by an effective anisotropic background, on which the interaction and evolution of the short modes change accordingly. We further develop the tidal simulation method, including proper corrections to the second order Lagrangian perturbation theory (2LPT) to generate initial conditions of the simulations. We measure the linear tidal responses of the matter power spectrum, at high redshift from our modified 2LPT, and at low redshift from the tidal simulations. Our results agree qualitatively with previous works, but exhibit quantitative differences in both cases. We also measure the linear tidal response of the halo shapes, or the shape bias, and find its universal relation with the linear halo bias, for which we provide a fitting formula. Furthermore, analogous to the assembly bias, we study the secondary dependence of the shape bias, and discover for the first time dependence on halo concentration and axis ratio. Our results provide useful insights for studies of the intrinsic alignment as source of either contamination or information. These effects need to be correctly taken into account when one uses intrinsic alignments of galaxy shapes as a precision cosmological tool.

Show Abstract
arXiv e-prints
2020

A fast solver for the narrow capture and narrow escape problems in the sphere

We present an efficient method to solve the narrow capture and narrow escape problems for the sphere. The narrow capture problem models the equilibrium behavior of a Brownian particle in the exterior of a sphere whose surface is reflective, except for a collection of small absorbing patches. The narrow escape problem is the dual problem: it models the behavior of a Brownian particle confined to the interior of a sphere whose surface is reflective, except for a collection of small patches through which it can escape. Mathematically, these give rise to mixed Dirichlet/Neumann boundary value problems of the Poisson equation. They are numerically challenging for two main reasons: (1) the solutions are non-smooth at Dirichlet-Neumann interfaces, and (2) they involve adaptive mesh refinement and the solution of large, ill-conditioned linear systems when the number of small patches is large. By using the Neumann Green's functions for the sphere, we recast each boundary value problem as a system of first-kind integral equations on the collection of patches. A block-diagonal preconditioner together with a multiple scattering formalism leads to a well-conditioned system of second-kind integral equations and a very efficient approach to discretization. This system is solved iteratively using GMRES. We develop a hierarchical, fast multipole method-like algorithm to accelerate each matrix-vector product. Our method is insensitive to the patch size, and the total cost scales with the number N of patches as O(N log N), after a precomputation whose cost depends only on the patch size and not on the number or arrangement of patches. We demonstrate the method with several numerical examples, and are able to achieve highly accurate solutions with 100,000 patches in one hour on a 60-core workstation.

Show Abstract

Kymatio: Scattering Transforms in Python

Mathieu Andreux, Tomás Angles, Georgios Exarchakis, Roberto Leonarduzzi, Gaspar Rochette, Louis Thiry, John Zarka, S. Mallat, J. Andén, Eugene Belilovsky, Joan Bruna, Vincent Lostanlen, Muawiz Chaudhary, Matthew J. Hirn, Edouard Oyallon, Sixin Zhang, Carmine Cella, M. Eickenberg

The wavelet scattering transform is an invariant and stable signal representation suitable for many signal processing and machine learning applications. We present the Kymatio software package, an easy-to-use, high-performance Python implementation of the scattering transform in 1D, 2D, and 3D that is compatible with modern deep learning frameworks, including PyTorch and TensorFlow/Keras. The transforms are implemented on both CPUs and GPUs, the latter offering a significant speedup over the former. The package also has a small memory footprint. Source code, documentation, and examples are available under a BSD license at https://www.kymat.iohttps://www.kymat.io.

Show Abstract
  • Previous Page
  • Viewing
  • Next Page
Advancing Research in Basic Science and MathematicsSubscribe to Flatiron Institute announcements and other foundation updates

privacy consent banner

Privacy preference

We use cookies to provide you with the best online experience. By clicking "Accept All," you help us understand how our site is used and enhance its performance. You can change your choice at any time here. To learn more, please visit our Privacy Policy.