A simple topological graph T = (V(T), E(T)) is a drawing of a graph in the plane where every two edges have at most one common point (an endpoint or a crossing) and no three edges pass through a single crossing. Topological graphs G and H are isomorphic if H can be obtained from G by a homeomorphism of the sphere, and weakly isomorphic if G and H have the same set of pairs of crossing edges. We generalize results of Pach and Toth and the author's previous results on counting different drawings of a graph under both notions of isomorphism. We prove that for every graph G with n vertices, m edges and no isolated vertices the number of weak isomorphism classes of simple topological graphs that realize G is at most 2^O(n^2 log(m/n)), and at most 2^O(mn^{1/2} log n) if m < n^{3/2}. As a consequence we obtain a new upper bound 2^O(n^{3/2} log n) on the number of intersection graphs of n pseudosegments. We improve the upper bound on the number of weak isomorphism classes of simple complete topological graphs with n vertices to 2^{n^2 alpha(n)^O(1)}, using an upper bound on the size of a set of permutations with bounded VC-dimension recently proved by Cibulka and the author. We show that the number of isomorphism classes of simple topological graphs that realize G is at most 2^{m^2+O(mn)} and at least 2^Omega(m^2) for graphs with m > (6+epsilon)n.
The study of homomorphisms of $(n,m)$-graphs, that is, adjacency preserving vertex mappings of graphs with $n$ types of arcs and $m$ types of edges was initiated by Ne\v{s}et\v{r}il and Raspaud [Journal of Combinatorial Theory, Series B 2000]. Later, some attempts were made to generalize the switch operation that is popularly used in the study of signed graphs, and study its effect on the above mentioned homomorphism. In this article, we too provide a generalization of the switch operation on $(n,m)$-graphs, which to the best of our knowledge, encapsulates all the previously known generalizations as special cases. We approach to study the homomorphism with respect to the switch operation axiomatically. We prove some fundamental results, which, we believe, will be essential tools in the further study of this topic. We also prove the existence of a categorical product for $(n,m)$-graphs with respect to a particular class of generalized switch. We also provide a way to calculate the product explicitly, and prove general properties of the product. Also, in the process of proving the fundamental results, we have provided yet another solution to an open problem posed by Klostermeyer and MacGillivray [Discrete Mathematics 2004].
We present a novel framework based on semi-bounded spatial operators for analyzing and discretizing initial boundary value problems on moving and deforming domains. This development extends an existing framework for well-posed problems and energy stable discretizations from stationary domains to the general case including arbitrary mesh motion. In particular, we show that an energy estimate derived in the physical coordinate system is equivalent to a semi-bounded property with respect to a stationary reference domain. The continuous analysis leading up to this result is based on a skew-symmetric splitting of the material time derivative, and thus relies on the property of integration-by-parts. Following this, a mimetic energy stable arbitrary Lagrangian-Eulerian framework for semi-discretization is formulated, based on approximating the material time derivative in a way consistent with discrete summation-by-parts. Thanks to the semi-bounded property, a method-of-lines approach using standard explicit or implicit time integration schemes can be applied to march the system forward in time. The same type of stability arguments applies as for the corresponding stationary domain problem, without regards to additional properties such as discrete geometric conservation. As an additional bonus we demonstrate that discrete geometric conservation, in the sense of exact free-stream preservation, can still be achieved in an automatic way with the new framework. However, we stress that this is not necessary for stability.
We examine the use of the Euler-Maclaurin formula and new derived uniform asymptotic expansions for the numerical evaluation of the Lerch transcendent $\Phi(z, s, a)$ for $z, s, a \in \mathbb{C}$ to arbitrary precision. A detailed analysis of these expansions is accompanied by rigorous error bounds. A complete scheme of computation for large and small values of the parameters and argument is described along with algorithmic details to achieve high performance. The described algorithm has been extensively tested in different regimes of the parameters and compared with current state-of-the-art codes. An open-source implementation of $\Phi(z, s, a)$ based on the algorithms described in this paper is available.
Gaussian processes (GPs) are an attractive class of machine learning models because of their simplicity and flexibility as building blocks of more complex Bayesian models. Meanwhile, graph neural networks (GNNs) emerged recently as a promising class of models for graph-structured data in semi-supervised learning and beyond. Their competitive performance is often attributed to a proper capturing of the graph inductive bias. In this work, we introduce this inductive bias into GPs to improve their predictive performance for graph-structured data. We show that a prominent example of GNNs, the graph convolutional network, is equivalent to some GP when its layers are infinitely wide; and we analyze the kernel universality and the limiting behavior in depth. We further present a programmable procedure to compose covariance kernels inspired by this equivalence and derive example kernels corresponding to several interesting members of the GNN family. We also propose a computationally efficient approximation of the covariance matrix for scalable posterior inference with large-scale data. We demonstrate that these graph-based kernels lead to competitive classification and regression performance, as well as advantages in computation time, compared with the respective GNNs.
We study the convergence rate of discretized Riemannian Hamiltonian Monte Carlo on sampling from distributions in the form of $e^{-f(x)}$ on a convex body $\mathcal{M}\subset\mathbb{R}^{n}$. We show that for distributions in the form of $e^{-\alpha^{\top}x}$ on a polytope with $m$ constraints, the convergence rate of a family of commonly-used integrators is independent of $\left\Vert \alpha\right\Vert _{2}$ and the geometry of the polytope. In particular, the implicit midpoint method (IMM) and the generalized Leapfrog method (LM) have a mixing time of $\widetilde{O}\left(mn^{3}\right)$ to achieve $\epsilon$ total variation distance to the target distribution. These guarantees are based on a general bound on the convergence rate for densities of the form $e^{-f(x)}$ in terms of parameters of the manifold and the integrator. Our theoretical guarantee complements the empirical results of [KLSV22], which shows that RHMC with IMM can sample ill-conditioned, non-smooth and constrained distributions in very high dimension efficiently in practice.
For a set system $(V,{\mathcal C}\subseteq 2^V)$, we call a subset $C\in{\mathcal C}$ a component. A nonempty subset $Y\subseteq C$ is a minimal removable set (MRS) of $C$ if $C\setminus Y\in{\mathcal C}$ and no proper nonempty subset $Z\subsetneq Y$ satisfies $C\setminus Z\in{\mathcal C}$. In this paper, we consider the problem of enumerating all components in a set system such that, for every two components $C,C'\in{\mathcal C}$ with $C'\subsetneq C$, every MRS $X$ of $C$ satisfies either $X\subseteq C'$ or $X\cap C'=\emptyset$. We provide a partition-based algorithm for this problem, which yields the first linear delay algorithms to enumerate all 2-edge-connected induced subgraphs, and to enumerate all 2-vertex-connected induced subgraphs.
Graph Neural Networks (GNNs) have been studied from the lens of expressive power and generalization. However, their optimization properties are less well understood. We take the first step towards analyzing GNN training by studying the gradient dynamics of GNNs. First, we analyze linearized GNNs and prove that despite the non-convexity of training, convergence to a global minimum at a linear rate is guaranteed under mild assumptions that we validate on real-world graphs. Second, we study what may affect the GNNs' training speed. Our results show that the training of GNNs is implicitly accelerated by skip connections, more depth, and/or a good label distribution. Empirical results confirm that our theoretical results for linearized GNNs align with the training behavior of nonlinear GNNs. Our results provide the first theoretical support for the success of GNNs with skip connections in terms of optimization, and suggest that deep GNNs with skip connections would be promising in practice.
This paper focuses on the expected difference in borrower's repayment when there is a change in the lender's credit decisions. Classical estimators overlook the confounding effects and hence the estimation error can be magnificent. As such, we propose another approach to construct the estimators such that the error can be greatly reduced. The proposed estimators are shown to be unbiased, consistent, and robust through a combination of theoretical analysis and numerical testing. Moreover, we compare the power of estimating the causal quantities between the classical estimators and the proposed estimators. The comparison is tested across a wide range of models, including linear regression models, tree-based models, and neural network-based models, under different simulated datasets that exhibit different levels of causality, different degrees of nonlinearity, and different distributional properties. Most importantly, we apply our approaches to a large observational dataset provided by a global technology firm that operates in both the e-commerce and the lending business. We find that the relative reduction of estimation error is strikingly substantial if the causal effects are accounted for correctly.
With the rapid increase of large-scale, real-world datasets, it becomes critical to address the problem of long-tailed data distribution (i.e., a few classes account for most of the data, while most classes are under-represented). Existing solutions typically adopt class re-balancing strategies such as re-sampling and re-weighting based on the number of observations for each class. In this work, we argue that as the number of samples increases, the additional benefit of a newly added data point will diminish. We introduce a novel theoretical framework to measure data overlap by associating with each sample a small neighboring region rather than a single point. The effective number of samples is defined as the volume of samples and can be calculated by a simple formula $(1-\beta^{n})/(1-\beta)$, where $n$ is the number of samples and $\beta \in [0,1)$ is a hyperparameter. We design a re-weighting scheme that uses the effective number of samples for each class to re-balance the loss, thereby yielding a class-balanced loss. Comprehensive experiments are conducted on artificially induced long-tailed CIFAR datasets and large-scale datasets including ImageNet and iNaturalist. Our results show that when trained with the proposed class-balanced loss, the network is able to achieve significant performance gains on long-tailed datasets.
We examine the problem of question answering over knowledge graphs, focusing on simple questions that can be answered by the lookup of a single fact. Adopting a straightforward decomposition of the problem into entity detection, entity linking, relation prediction, and evidence combination, we explore simple yet strong baselines. On the popular SimpleQuestions dataset, we find that basic LSTMs and GRUs plus a few heuristics yield accuracies that approach the state of the art, and techniques that do not use neural networks also perform reasonably well. These results show that gains from sophisticated deep learning techniques proposed in the literature are quite modest and that some previous models exhibit unnecessary complexity.