math.FA
11 postsarXiv:2408.00933v2 Announce Type: replace-cross Abstract: The bad science matrix problem consists in finding, among all matrices $A \in \mathbb{R}^{n \times n}$ with rows having unit $\ell^2$ norm, one that maximizes $\beta(A) = \frac{1}{2^n} \sum_{x \in \{-1, 1\}^n} \|Ax\|_\infty$. Our main contribution is an explicit construction of an $n \times n$ matrix $A$ showing that $\beta(A) \geq \sqrt{\log_2(n+1)}$, which is only 18% smaller than the asymptotic rate. We prove that every entry of any optimal matrix is a square root of a rational number, and we find provably optimal matrices for $n \leq 4$.
arXiv:2501.10607v1 Announce Type: cross Abstract: Given $N$ geodesic caps on the normalized unit sphere in $\mathbb{R}^d$, and whose total surface area sums to one, what is the maximal surface area their union can cover? We show that when these caps have equal surface area, as both the dimension $d$ and the number of caps $N$ tend to infinity, the maximum proportion covered approaches $1 - e^{-1} \approx 0.632$. Furthermore, this maximum is achieved by a random partial sphere covering. Our result refines a classical estimate for the covering density of $\mathbb{R}^d$ by Erd\H{o}s, Few, and Rogers (Mathematika, 11(2):171--184, 1964).
arXiv:2501.11871v1 Announce Type: new Abstract: In $\mathbb{R}^3$, the primal and dual constructions yield completely different discrete Laplacians for tetrahedral meshes.In this article, we prove that the discrete Laplacian satisfies the Euler-Lagrange equation of the Dirichlet energy in terms of the associated discrete Laplacian corresponding to the dual construction. Specifically, for a three simplex immersed in $\mathbb{R}^3$, the associated discrete Laplacian on the tetrahedron can be expressed as the discrete Laplacian of the faces of the tetrahedron and the associated discrete mean curvature term given by the ambient space $\mathbb{R}^3$. Based on geometric foundations, we provide a mathematical proof showing that the dual construction gives a optimal Laplacian in $\mathbb{R}^3$ compared to the primal construction. Moreover, we show that the associated discrete mean curvature is more sensitive to the initial mesh than other state-of-the-art discrete mean curvatures when the angle changes instantaneously. Instead of improving the angular transient accuracy through mesh subdivision, we can improve the accuracy by providing a higher order approximation of the instantaneous change in angle to reduce the solution error.
arXiv:2501.11451v1 Announce Type: new Abstract: Tomographic investigations are a central tool in medical applications, allowing doctors to image the interior of patients. The corresponding measurement process is commonly modeled by the Radon transform. In practice, the solution of the tomographic problem requires discretization of the Radon transform and its adjoint (called the backprojection). There are various discretization schemes; often structured around three discretization parameters: spatial-, detector-, and angular resolutions. The most widespread approach uses the ray-driven Radon transform and the pixel-driven backprojection in a balanced resolution setting, i.e., the spatial resolution roughly equals the detector resolution. The use of these particular discretization approaches is based on anecdotal reports of their approximation performance, but there is little rigorous analysis of these methods' approximation errors. This paper presents a novel interpretation of ray-driven and pixel-driven methods as convolutional discretizations, illustrating that from an abstract perspective these methods are similar. Moreover, we announce statements concerning the convergence of the ray-driven Radon transform and the pixel-driven backprojection under balanced resolutions. Our considerations are supported by numerical experiments highlighting aspects of the discussed methods.
arXiv:2408.11612v2 Announce Type: replace Abstract: Computations in high-dimensional spaces can often be realized only approximately, using a certain number of projections onto lower dimensional subspaces or sampling from distributions. In this paper, we are interested in pairs of real-valued functions $(F,f)$ on $[0,\infty)$ that are related by the projection/slicing formula $F (\| x \|) = \mathbb E_{\xi} \big[ f \big(|\langle x,\xi \rangle| \big) \big]$ for $x\in\mathbb R^d$, where the expectation value is taken over uniformly distributed directions in $\mathbb R^d$. While it is known that $F$ can be obtained from $f$ by an Abel-like integral formula, we construct conversely $f$ from given $F$ using their Fourier transforms. First, we consider the relation between $F$ and $f$ for radial functions $F(\| \cdot\| )$ that are Fourier transforms of $L^1$ functions. Besides $d$- and one-dimensional Fourier transforms, it relies on a rotation operator, an averaging operator and a multiplication operator to manage the walk from $d$ to one dimension in the Fourier space. Then, we generalize the results to tempered distributions, where we are mainly interested in radial regular tempered distributions. Based on Bochner's theorem, this includes positive definite functions $F(\| \cdot\| )$ and, by the theory of fractional derivatives, also functions $F$ whose derivative of order $\lfloor d/2\rfloor$ is slowly increasing and continuous.
arXiv:2501.06539v1 Announce Type: new Abstract: We construct a Neural Network that approximates the matrix multiplication operator for any activation functions for which there exist a Neural Network which can approximate the scalar multiplication function. In particular, we use the Strassen algorithm for reducing the number of weights and layers needed for such Neural Network. This allows us to define another Neural Network for approximating the inverse matrix operator. Also, by relying on the Galerkin method, we apply those Neural Networks for resolving parametric elliptic PDEs for a whole set of parameters at the same time. Finally, we discuss the improvements with respect to prior results.
arXiv:2410.00355v2 Announce Type: replace-cross Abstract: Finding eigenvalue distributions for a number of sparse random matrix ensembles can be reduced to solving nonlinear integral equations of the Hammerstein type. While a systematic mathematical theory of such equations exists, it has not been previously applied to sparse matrix problems. We close this gap in the literature by showing how one can employ numerical solutions of Hammerstein equations to accurately recover the spectra of adjacency matrices and Laplacians of random graphs. While our treatment focuses on random graphs for concreteness, the methodology has broad applications to more general sparse random matrices.
arXiv:2501.03697v1 Announce Type: new Abstract: Identifying an appropriate function space for deep neural networks remains a key open question. While shallow neural networks are naturally associated with Reproducing Kernel Banach Spaces (RKBS), deep networks present unique challenges. In this work, we extend RKBS to chain RKBS (cRKBS), a new framework that composes kernels rather than functions, preserving the desirable properties of RKBS. We prove that any deep neural network function is a neural cRKBS function, and conversely, any neural cRKBS function defined on a finite dataset corresponds to a deep neural network. This approach provides a sparse solution to the empirical risk minimization problem, requiring no more than $N$ neurons per layer, where $N$ is the number of data points.
arXiv:2501.01929v1 Announce Type: cross Abstract: This paper extends the sample complexity theory for ill-posed inverse problems developed in a recent work by the authors [`Compressed sensing for inverse problems and the sample complexity of the sparse Radon transform', J. Eur. Math. Soc., to appear], which was originally focused on the sparse Radon transform. We demonstrate that the underlying abstract framework, based on infinite-dimensional compressed sensing and generalized sampling techniques, can effectively handle a variety of practical applications. Specifically, we analyze three case studies: (1) The reconstruction of a sparse signal from a finite number of pointwise blurred samples; (2) The recovery of the (sparse) source term of an elliptic partial differential equation from finite samples of the solution; and (3) A moderately ill-posed variation of the classical sensing problem of recovering a wavelet-sparse signal from finite Fourier samples, motivated by magnetic resonance imaging. For each application, we establish rigorous recovery guarantees by verifying the key theoretical requirements, including quasi-diagonalization and coherence bounds. Our analysis reveals that careful consideration of balancing properties and optimized sampling strategies can lead to improved reconstruction performance. The results provide a unified theoretical foundation for compressed sensing approaches to inverse problems while yielding practical insights for specific applications.
arXiv:2410.08361v2 Announce Type: replace-cross Abstract: In this paper, we study a Markov chain-based stochastic gradient algorithm in general Hilbert spaces, aiming to approximate the optimal solution of a quadratic loss function. We establish probabilistic upper bounds on its convergence. We further extend these results to an online regularized learning algorithm in reproducing kernel Hilbert spaces, where the samples are drawn along a Markov chain trajectory hence the samples are of the non i.i.d. type.
arXiv:2412.18468v1 Announce Type: cross Abstract: Matrix concentration inequalities and their recently discovered sharp counterparts provide powerful tools to bound the spectrum of random matrices whose entries are linear functions of independent random variables. However, in many applications in theoretical computer science and in other areas one encounters more general random matrix models, called matrix chaoses, whose entries are polynomials of independent random variables. Such models have often been studied on a case-by-case basis using ad-hoc methods that can yield suboptimal dimensional factors. In this paper we provide general matrix concentration inequalities for matrix chaoses, which enable the treatment of such models in a systematic manner. These inequalities are expressed in terms of flattenings of the coefficients of the matrix chaos. We further identify a special family of matrix chaoses of combinatorial type for which the flattening parameters can be computed mechanically by a simple rule. This allows us to provide a unified treatment of and improved bounds for matrix chaoses that arise in a variety of applications, including graph matrices, Khatri-Rao matrices, and matrices that arise in average case analysis of the sum-of-squares hierarchy.