We provide an algorithm for the minimum 2-vertex-connected spanning subgraph problem with approximation ratio $\frac{4}{3}$, improving upon the previous best factor $\frac{10}{7}$.
The Byzantine consensus problem involves $n$ processes, out of which t < n could be faulty and behave arbitrarily. Three properties characterize consensus: (1) termination, requiring correct (non-faulty) processes to eventually reach a decision, (2) agreement, preventing them from deciding different values, and (3) validity, precluding ``unreasonable'' decisions. But, what is a reasonable decision? Strong validity, a classical property, stipulates that, if all correct processes propose the same value, only that value can be decided. Weak validity, another established property, stipulates that, if all processes are correct and they propose the same value, that value must be decided. The space of possible validity properties is vast. However, their impact on consensus remains unclear. This paper addresses the question of which validity properties allow Byzantine consensus to be solvable with partial synchrony, and at what cost. First, we determine necessary and sufficient conditions for a validity property to make the consensus problem solvable; we say that such validity properties are solvable. Notably, we prove that, if n <= 3t, all solvable validity properties are trivial (there exists an always-admissible decision). Furthermore, we show that, with any non-trivial (and solvable) validity property, consensus requires Omega(t^2) messages. This extends the seminal Dolev-Reischuk bound, originally proven for strong validity, to all non-trivial validity properties. Lastly, we give a general Byzantine consensus algorithm, we call Universal, for any solvable (and non-trivial) validity property. Importantly, Universal incurs O(n^2) message complexity. Thus, together with our lower bound, Universal implies a fundamental result in partial synchrony: with t \in Omega(n), the message complexity of all (non-trivial) consensus variants is Theta(n^2).
We derive optimality conditions for the optimum sample allocation problem, formulated as the determination of the fixed strata sample sizes that minimize the total cost of the survey, under assumed level of the variance of the stratified estimator and one-sided upper bounds imposed on sample sizes in strata. In this context, we take that the variance function is of some generic form that involves the stratified $\pi$ estimator of the population total with stratified simple random sampling without replacement design as a special case. The optimality conditions mentioned above will be derived with the use of convex optimization theory and the Karush-Kuhn-Tucker conditions. Based on the established optimality conditions we give a formal proof of the existing procedure, termed here as LRNA, that solves the allocation problem considered. We formulate the LRNA in such a way that it also provides the solution to classical optimum allocation problem (i.e. minimization of the estimator's variance under fixed total cost) under one-sided lower bounds imposed on sample sizes in strata. From this standpoint, the LRNA can be considered as a counterparty to the popular recursive Neyman allocation procedure that is used to solve the classical problem of optimum sample allocation but with one-sided upper bounds. Ready-to-use R-implementation of the LRNA is available through our package stratallo, which is published on the Comprehensive R Archive Network (CRAN) package repository.
We study a class of stochastic semilinear damped wave equations driven by additive Wiener noise. Owing to the damping term, under appropriate conditions on the nonlinearity, the solution admits a unique invariant distribution. We apply semi-discrete and fully-discrete methods in order to approximate this invariant distribution, using a spectral Galerkin method and an exponential Euler integrator for spatial and temporal discretization respectively. We prove that the considered numerical schemes also admit unique invariant distributions, and we prove error estimates between the approximate and exact invariant distributions, with identification of the orders of convergence. To the best of our knowledge this is the first result in the literature concerning numerical approximation of invariant distributions for stochastic damped wave equations.
In this paper, we present a polynomial time 2-approximation algorithm for the {\em unrooted prize-collecting forest with $K$ components} (URPCF$_K$) problem, the goal of which is to find a forest with exactly $K$ connected components to minimize the weight of the forest plus the penalty incurred by the vertices not spanned by the forest. For its rooted version RPCF$_K$, a 2-approximation algorithm is known. For the unrooted version, transforming it into a rooted version by guessing roots runs in time exponentially depending on $K$, which is unacceptable when $K$ is not a constant. We conquer this challenge by designing a rootless growing plus rootless pruning algorithm. As an application, we make use of this algorithm to solve the {\em prize-collecting min-sensor sweep cover} problem, improving previous approximation ratio 8 to 5. Keywords: approximation algorithm, prize-collecting Steiner forest, sweep cover.
The stochastic block model is a canonical random graph model for clustering and community detection on network-structured data. Decades of extensive study on the problem have established many profound results, among which the phase transition at the Kesten-Stigum threshold is particularly interesting both from a mathematical and an applied standpoint. It states that no estimator based on the network topology can perform substantially better than chance on sparse graphs if the model parameter is below certain threshold. Nevertheless, if we slightly extend the horizon to the ubiquitous semi-supervised setting, such a fundamental limitation will disappear completely. We prove that with arbitrary fraction of the labels revealed, the detection problem is feasible throughout the parameter domain. Moreover, we introduce two efficient algorithms, one combinatorial and one based on optimization, to integrate label information with graph structures. Our work brings a new perspective to stochastic model of networks and semidefinite program research.
In this work, we propose an extension of telescopic derivative operators for the DGSEM with Gauss nodes, and we prove that this formulation is equivalent to its usual matrix counterpart. Among other possible applications, this allows extending the stabilization methods already developed for Gauss-Lobatto nodes to Gauss nodes, also ensuring properties such as entropy stability while retaining their improved accuracy.
Message-passing Graph Neural Networks (GNNs), which collect information from adjacent nodes achieve dismal performance on heterophilic graphs. Various schemes have been proposed to solve this problem, and propagating signed information on heterophilic edges has gained great attention. Recently, some works provided theoretical analysis that signed propagation always leads to performance improvement under a binary class scenario. However, we notice that prior analyses do not align well with multi-class benchmark datasets. This paper provides a new understanding of signed propagation for multi-class scenarios and points out two drawbacks in terms of message-passing and parameter update: (1) Message-passing: if two nodes belong to different classes but have a high similarity, signed propagation can decrease the separability. (2) Parameter update: the prediction uncertainty (e.g., conflict evidence) of signed neighbors increases during training, which can impede the stability of the algorithm. Based on the observation, we introduce two novel strategies for improving signed propagation under multi-class graphs. The proposed scheme combines calibration to secure robustness while reducing uncertainty. We show the efficacy of our theorem through extensive experiments on six benchmark graph datasets.
Spectral Clustering is one of the most traditional methods to solve segmentation problems. Based on Normalized Cuts, it aims at partitioning an image using an objective function defined by a graph. Despite their mathematical attractiveness, spectral approaches are traditionally neglected by the scientific community due to their practical issues and underperformance. In this paper, we adopt a sparse graph formulation based on the inclusion of extra nodes to a simple grid graph. While the grid encodes the pixel spatial disposition, the extra nodes account for the pixel color data. Applying the original Normalized Cuts algorithm to this graph leads to a simple and scalable method for spectral image segmentation, with an interpretable solution. Our experiments also demonstrate that our proposed methodology over performs traditional spectral algorithms for segmentation.
Ptychography, a prevalent imaging technique in fields such as biology and optics, poses substantial challenges in its reconstruction process, characterized by nonconvexity and large-scale requirements. This paper presents a novel approach by introducing a class of variational models that incorporate the weighted difference of anisotropic--isotropic total variation. This formulation enables the handling of measurements corrupted by Gaussian or Poisson noise, effectively addressing the nonconvex challenge. To tackle the large-scale nature of the problem, we propose an efficient stochastic alternating direction method of multipliers, which guarantees convergence under mild conditions. Numerical experiments validate the superiority of our approach by demonstrating its capability to successfully reconstruct complex-valued images, especially in recovering the phase components even in the presence of highly corrupted measurements.
Graph Convolutional Networks (GCNs) and their variants have experienced significant attention and have become the de facto methods for learning graph representations. GCNs derive inspiration primarily from recent deep learning approaches, and as a result, may inherit unnecessary complexity and redundant computation. In this paper, we reduce this excess complexity through successively removing nonlinearities and collapsing weight matrices between consecutive layers. We theoretically analyze the resulting linear model and show that it corresponds to a fixed low-pass filter followed by a linear classifier. Notably, our experimental evaluation demonstrates that these simplifications do not negatively impact accuracy in many downstream applications. Moreover, the resulting model scales to larger datasets, is naturally interpretable, and yields up to two orders of magnitude speedup over FastGCN.