In the Activation Edge-Multicover problem we are given a multigraph $G=(V,E)$ with activation costs $\{c_{e}^u,c_{e}^v\}$ for every edge $e=uv \in E$, and degree requirements $r=\{r_v:v \in V\}$. The goal is to find an edge subset $J \subseteq E$ of minimum activation cost $\sum_{v \in V}\max\{c_{uv}^v:uv \in J\}$,such that every $v \in V$ has at least $r_v$ neighbors in the graph $(V,J)$. Let $k= \max_{v \in V} r_v$ be the maximum requirement and let $\theta=\max_{e=uv \in E} \frac{\max\{c_e^u,c_e^v\}}{\min\{c_e^u,c_e^v\}}$ be the maximum quotient between the two costs of an edge. For $\theta=1$ the problem admits approximation ratio $O(\log k)$. For $k=1$ it generalizes the Set Cover problem (when $\theta=\infty$), and admits a tight approximation ratio $O(\log n)$. This implies approximation ratio $O(k \log n)$ for general $k$ and $\theta$, and no better approximation ratio was known. We obtain the first logarithmic approximation ratio $O(\log k +\log\min\{\theta,n\})$, that bridges between the two known ratios -- $O(\log k)$ for $\theta=1$ and $O(\log n)$ for $k=1$. This implies approximation ratio $O\left(\log k +\log\min\{\theta,n\}\right) +\beta \cdot (\theta+1)$ for the Activation $k$-Connected Subgraph problem, where $\beta$ is the best known approximation ratio for the ordinary min-cost version of the problem.
Given a sample of size $N$, it is often useful to select a subsample of smaller size $n<N$ to be used for statistical estimation or learning. Such a data selection step is useful to reduce the requirements of data labeling and the computational complexity of learning. We assume to be given $N$ unlabeled samples $\{{\boldsymbol x}_i\}_{i\le N}$, and to be given access to a `surrogate model' that can predict labels $y_i$ better than random guessing. Our goal is to select a subset of the samples, to be denoted by $\{{\boldsymbol x}_i\}_{i\in G}$, of size $|G|=n<N$. We then acquire labels for this set and we use them to train a model via regularized empirical risk minimization. By using a mixture of numerical experiments on real and synthetic data, and mathematical derivations under low- and high- dimensional asymptotics, we show that: $(i)$~Data selection can be very effective, in particular beating training on the full sample in some cases; $(ii)$~Certain popular choices in data selection methods (e.g. unbiased reweighted subsampling, or influence function-based subsampling) can be substantially suboptimal.
We consider the numerical solution of an abstract operator equation $Bu=f$ by using a least-squares approach. We assume that $B: X \to Y^*$ is an isomorphism, and that $A : Y \to Y^*$ implies a norm in $Y$, where $X$ and $Y$ are Hilbert spaces. The minimizer of the least-squares functional $\frac{1}{2} \, \| Bu-f \|_{A^{-1}}^2$, i.e., the solution of the operator equation, is then characterized by the gradient equation $Su=B^* A^{-1}f$ with an elliptic and self-adjoint operator $S:=B^* A^{-1} B : X \to X^*$. When introducing the adjoint $p = A^{-1}(f-Bu)$ we end up with a saddle point formulation to be solved numerically by using a mixed finite element method. Based on a discrete inf-sup stability condition we derive related a priori error estimates. While the adjoint $p$ is zero by construction, its approximation $p_h$ serves as a posteriori error indicator to drive an adaptive scheme when discretized appropriately. While this approach can be applied to rather general equations, here we consider second order linear partial differential equations, including the Poisson equation, the heat equation, and the wave equation, in order to demonstrate its potential, which allows to use almost arbitrary space-time finite element methods for the adaptive solution of time-dependent partial differential equations.
We analyze the conforming approximation of the time-harmonic Maxwell's equations using N\'ed\'elec (edge) finite elements. We prove that the approximation is asymptotically optimal, i.e., the approximation error in the energy norm is bounded by the best-approximation error times a constant that tends to one as the mesh is refined and/or the polynomial degree is increased. Moreover, under the same conditions on the mesh and/or the polynomial degree, we establish discrete inf-sup stability with a constant that corresponds to the continuous constant up to a factor of two at most. Our proofs apply under minimal regularity assumptions on the exact solution, so that general domains, material coefficients, and right-hand sides are allowed.
We consider approximating solutions to parameterized linear systems of the form $A(\mu_1,\mu_2) x(\mu_1,\mu_2) = b$, where $(\mu_1, \mu_2) \in \mathbb{R}^2$. Here the matrix $A(\mu_1,\mu_2) \in \mathbb{R}^{n \times n}$ is nonsingular, large, and sparse and depends nonlinearly on the parameters $\mu_1$ and $\mu_2$. Specifically, the system arises from a discretization of a partial differential equation and $x(\mu_1,\mu_2) \in \mathbb{R}^n$, $b \in \mathbb{R}^n$. This work combines companion linearization with the Krylov subspace method preconditioned bi-conjugate gradient (BiCG) and a decomposition of a tensor matrix of precomputed solutions, called snapshots. As a result, a reduced order model of $x(\mu_1,\mu_2)$ is constructed, and this model can be evaluated in a cheap way for many values of the parameters. The decomposition is performed efficiently using the sparse grid based higher-order proper generalized decomposition (HOPGD), and the snapshots are generated as one variable functions of $\mu_1$ or of $\mu_2$. Tensor decompositions performed on a set of snapshots can fail to reach a certain level of accuracy, and it is not possible to know a priori if the decomposition will be successful. This method offers a way to generate a new set of solutions on the same parameter space at little additional cost. An interpolation of the model is used to produce approximations on the entire parameter space, and this method can be used to solve a parameter estimation problem. Numerical examples of a parameterized Helmholtz equation show the competitiveness of our approach. The simulations are reproducible, and the software is available online.
We identify a family of $O(|E(G)|^2)$ nontrivial facets of the connected matching polytope of a graph $G$, that is, the convex hull of incidence vectors of matchings in $G$ whose covered vertices induce a connected subgraph. Accompanying software to further inspect the polytope of an input graph is available.
Let $(M,g)$ be a Riemannian manifold. If $\mu$ is a probability measure on $M$ given by a continuous density function, one would expect the Fr\'{e}chet means of data-samples $Q=(q_1,q_2,\dots, q_N)\in M^N$, with respect to $\mu$, to behave ``generically''; e.g. the probability that the Fr\'{e}chet mean set $\mbox{FM}(Q)$ has any elements that lie in a given, positive-codimension submanifold, should be zero for any $N\geq 1$. Even this simplest instance of genericity does not seem to have been proven in the literature, except in special cases. The main result of this paper is a general, and stronger, genericity property: given i.i.d. absolutely continuous $M$-valued random variables $X_1,\dots, X_N$, and a subset $A\subset M$ of volume-measure zero, $\mbox{Pr}\left\{\mbox{FM}(\{X_1,\dots,X_N\})\subset M\backslash A\right\}=1.$ We also establish a companion theorem for equivariant Fr\'{e}chet means, defined when $(M,g)$ arises as the quotient of a Riemannian manifold $(\widetilde{M},\tilde{g})$ by a free, isometric action of a finite group. The equivariant Fr\'{e}chet means lie in $\widetilde{M}$, but, as we show, project down to the ordinary Fr\'{e}chet sample means, and enjoy a similar genericity property. Both these theorems are proven as consequences of a purely geometric (and quite general) result that constitutes the core mathematics in this paper: If $A\subset M$ has volume zero in $M$ , then the set $\{Q\in M^N : \mbox{FM}(Q) \cap A\neq\emptyset\}$ has volume zero in $M^N$. We conclude the paper with an application to partial scaling-rotation means, a type of mean for symmetric positive-definite matrices.
We present here a new splitting method to solve Lyapunov equations of the type $AP + PA^T=-BB^T$ in a Kronecker product form. Although that resulting matrix is of order $n^2$, each iteration of the method demands only two operations with the matrix $A$: a multiplication of the form $(A-\sigma I) \hat{B}$ and a inversion of the form $(A-\sigma I)^{-1}\hat{B}$. We see that for some choice of a parameter the iteration matrix is such that all their eigenvalues are in absolute value less than 1, which means that it should converge without depending on the starting vector. Nevertheless we present a theorem that enables us how to get a good starting vector for the method.
The numerical solution of continuum damage mechanics (CDM) problems suffers from convergence-related challenges during the material softening stage, and consequently existing iterative solvers are subject to a trade-off between computational expense and solution accuracy. In this work, we present a novel unified arc-length (UAL) method, and we derive the formulation of the analytical tangent matrix and governing system of equations for both local and non-local gradient damage problems. Unlike existing versions of arc-length solvers that monolithically scale the external force vector, the proposed method treats the latter as an independent variable and determines the position of the system on the equilibrium path based on all the nodal variations of the external force vector. This approach renders the proposed solver substantially more efficient and robust than existing solvers used in CDM problems. We demonstrate the considerable advantages of the proposed algorithm through several benchmark 1D problems with sharp snap-backs and 2D examples under various boundary conditions and loading scenarios. The proposed UAL approach exhibits a superior ability of overcoming critical increments along the equilibrium path. Moreover, the proposed UAL method is 1-2 orders of magnitude faster than force-controlled arc-length and monolithic Newton-Raphson solvers.
This paper studies the extreme singular values of non-harmonic Fourier matrices. Such a matrix of size $m\times s$ can be written as $\Phi=[ e^{-2\pi i j x_k}]_{j=0,1,\dots,m-1, k=1,2,\dots,s}$ for some set $\mathcal{X}=\{x_k\}_{k=1}^s$. The main results provide explicit lower bounds for the smallest singular value of $\Phi$ under the assumption $m\geq 6s$ and without any restrictions on $\mathcal{X}$. They show that for an appropriate scale $\tau$ determined by a density criteria, interactions between elements in $\mathcal{X}$ at scales smaller than $\tau$ are most significant and depends on the multiscale structure of $\mathcal{X}$ at fine scales, while distances larger than $\tau$ are less important and only depend on the local sparsity of the far away points. Theoretical and numerical comparisons show that the main results significantly improve upon classical bounds and achieve the same rate that was previously discovered for more restrictive settings.
We prove tight bounds on the site percolation threshold for $k$-uniform hypergraphs of maximum degree $\Delta$ and for $k$-uniform hypergraphs of maximum degree $\Delta$ in which any pair of edges overlaps in at most $r$ vertices. The hypergraphs that achieve these bounds are hypertrees, but unlike in the case of graphs, there are many different $k$-uniform, $\Delta$-regular hypertrees. Determining the extremal tree for a given $k, \Delta, r$ involves an optimization problem, and our bounds arise from a convex relaxation of this problem. By combining our percolation bounds with the method of disagreement percolation we obtain improved bounds on the uniqueness threshold for the hard-core model on hypergraphs satisfying the same constraints. Our uniqueness conditions imply exponential weak spatial mixing, and go beyond the known bounds for rapid mixing of local Markov chains and existence of efficient approximate counting and sampling algorithms. Our results lead to natural conjectures regarding the aforementioned algorithmic tasks, based on the intuition that uniqueness thresholds for the extremal hypertrees for percolation determine computational thresholds.