Let $G=(V(G),E(G))$ be a finite simple undirected graph with vertex set $V(G)$, edge set $E(G)$ and vertex subset $S\subseteq V(G)$. $S$ is termed \emph{open-dominating} if every vertex of $G$ has at least one neighbor in $S$, and \emph{open-independent, open-locating-dominating} (an $OLD_{oind}$-set for short) if no two vertices in $G$ have the same set of neighbors in $S$, and each vertex in $S$ is open-dominated exactly once by $S$. The problem of deciding whether or not $G$ has an $OLD_{oind}$-set has important applications, has been studied elsewhere and is known to be $\mathcal{NP}$-complete. Reinforcing this result, it appears that the problem is notoriously difficult as we show that its complexity remains the same even for just planar bipartite graphs. Also, we present characterizations of both $P_4$-tidy graphs and the complementary prisms of cographs, that have an $OLD_{oind}$-set.
We consider a novel graph-structured change point problem. We observe a random vector with piecewise constant mean and whose independent, sub-Gaussian coordinates correspond to the $n$ nodes of a fixed graph. We are interested in recovering the partition of the nodes associated to the constancy regions of the mean vector. Although graph-valued signals of this type have been previously studied in the literature for the different tasks of testing for the presence of an anomalous cluster and of estimating the mean vector, no localisation results are known outside the classical case of chain graphs. When the partition $\mathcal{S}$ consists of only two elements, we characterise the difficulty of the localisation problem in terms of: the maximal noise variance $\sigma^2$, the size $\Delta$ of the smaller element of the partition, the magnitude $\kappa$ of the difference in the signal values and the sum of the effective resistance edge weights $|\partial_r(\mathcal{S})|$ of the corresponding cut. We demonstrate an information theoretical lower bound implying that, in the low signal-to-noise ratio regime $\kappa^2 \Delta \sigma^{-2} |\partial_r(\mathcal{S})|^{-1} \lesssim 1$, no consistent estimator of the true partition exists. On the other hand, when $\kappa^2 \Delta \sigma^{-2} |\partial_r(\mathcal{S})|^{-1} \gtrsim \zeta_n \log\{r(|E|)\}$, with $r(|E|)$ being the sum of effective resistance weighted edges and $\zeta_n$ being any diverging sequence in $n$, we show that a polynomial-time, approximate $\ell_0$-penalised least squared estimator delivers a localisation error of order $ \kappa^{-2} \sigma^2 |\partial_r(\mathcal{S})| \log\{r(|E|)\}$. Aside from the $\log\{r(|E|)\}$ term, this rate is minimax optimal. Finally, we provide upper bounds on the localisation error for more general partitions of unknown sizes.
We consider a matching problem in a bipartite graph $G=(A\cup B,E)$ where each node in $A$ is an agent having preferences in partial order over her neighbors, while nodes in $B$ are objects with no preferences. The size of our matching is more important than node preferences; thus, we are interested in maximum matchings only. Any pair of maximum matchings in $G$ (equivalently, perfect matchings or assignments) can be compared by holding a head-to-head election between them where agents are voters. The goal is to compute an assignment $M$ such that there is no better or "more popular" assignment. This is the popular assignment problem and it generalizes the well-studied popular matching problem. Popular assignments need not always exist. We show a polynomial-time algorithm that decides if the given instance admits one or not, and computes one, if so. In instances with no popular assignment, we consider the problem of finding an almost popular assignment, i.e., an assignment with minimum unpopularity margin. We show an $O^*(|E|^k)$ time algorithm for deciding if there exists an assignment with unpopularity margin at most $k$. We show that this algorithm is essentially optimal by proving that the problem is $\mathsf{W}_l[1]$-hard with parameter $k$. We also consider the minimum-cost popular assignment problem when there are edge costs, and show its $\mathsf{NP}$-hardness even when all edge costs are in $\{0,1\}$ and agents have strict preferences. By contrast, we propose a polynomial-time algorithm to the problem of deciding if there exists a popular assignment with a given set of forced/forbidden edges (this tractability holds even for partially ordered preferences). Our algorithms are combinatorial and based on LP duality. They search for an appropriate witness or dual certificate, and when a certificate cannot be found, we prove that the desired assignment does not exist in $G$.
Strategic term rewriting and attribute grammars are two powerful programming techniques widely used in language engineering. The former, relies on strategies to apply term rewrite rules in defining language transformations, while the latter is suitable to express context-dependent language processing algorithms. Each of these techniques, however, is usually implemented by its own powerful and large language processor system. As a result, it makes such systems harder to extend and to combine. In this paper, we present the embedding of both strategic tree rewriting and attribute grammars in a zipper-based, purely functional setting. Zippers provide a simple, but generic tree-walk mechanism that is the building block technique we use to express the purely-functional embedding of both techniques. The embedding of the two techniques in the same setting has several advantages: First, we easily combine/zip attribute grammars and strategies, thus providing language engineers the best of the two worlds. Second, the combined embedding is easier to maintain and extend since it is written in a concise and uniform setting. This results in a very small library which is able to express advanced (static) analysis and transformation tasks. We show the expressive power of our library in optimizing Haskell let expressions, expressing several Haskell refactorings and solving several language processing tasks of the LDTA Tool Challenge.
We study the complexity of approximating the partition function $Z_{\mathrm{Ising}}(G; \beta)$ of the Ising model in terms of the relation between the edge interaction $\beta$ and a parameter $\Delta$ which is an upper bound on the maximum degree of the input graph $G$. Following recent trends in both statistical physics and algorithmic research, we allow the edge interaction $\beta$ to be any complex number. Many recent partition function results focus on complex parameters, both because of physical relevance and because of the key role of the complex case in delineating the tractability/intractability phase transition of the approximation problem. In this work we establish both new tractability results and new intractability results. Our tractability results show that $Z_{\mathrm{Ising}}(-; \beta)$ has an FPTAS when $\lvert \beta - 1 \rvert / \lvert \beta + 1 \rvert < \tan(\pi / (4 \Delta - 4))$. The core of the proof is showing that there are no inputs~$G$ that make the partition function $0$ when $\beta$ is in this range. Our result significantly extends the known zero-free region of the Ising model (and hence the known approximation results). Our intractability results show that it is $\mathrm{\#P}$-hard to multiplicatively approximate the norm and to additively approximate the argument of $Z_{\mathrm{Ising}}(-; \beta)$ when $\beta \in \mathbb{C}$ is an algebraic number such that $\beta \not \in \mathbb{R} \cup \{i, -i\}$ and $\lvert \beta - 1\rvert / \lvert \beta + 1 \rvert > 1 / \sqrt{\Delta - 1}$. These are the first results to show intractability of approximating $Z_{\mathrm{Ising}}(-, \beta)$ on bounded degree graphs with complex $\beta$. Moreover, we demonstrate situations in which zeros of the partition function imply hardness of approximation in the Ising model.
We introduce a framework for generating, organizing, and reasoning with computational knowledge. It is motivated by the observation that most problems in Computational Sciences and Engineering (CSE) can be described as that of completing (from data) a computational graph representing dependencies between functions and variables. Functions and variables may be known, unknown, or random. Data comes in the form of observations of distinct values of a finite number of subsets of the variables of the graph. The underlying problem combines a regression problem (approximating unknown functions) with a matrix completion problem (recovering unobserved variables in the data). Replacing unknown functions by Gaussian Processes (GPs) and conditioning on observed data provides a simple but efficient approach to completing such graphs. Since the proposed framework is highly expressive, it has a vast potential application scope. Since the completion process can be automatized, as one solves $\sqrt{\sqrt{2}+\sqrt{3}}$ on a pocket calculator without thinking about it, one could, with the proposed framework, solve a complex CSE problem by drawing a diagram. Compared to traditional kriging, the proposed framework can be used to recover unknown functions with much scarcer data by exploiting interdependencies between multiple functions and variables. The Computational Graph Completion (CGC) problem addressed by the proposed framework could therefore also be interpreted as a generalization of that of solving linear systems of equations to that of approximating unknown variables and functions with noisy, incomplete, and nonlinear dependencies. Numerous examples illustrate the flexibility, scope, efficacy, and robustness of the CGC framework and show how it can be used as a pathway to identifying simple solutions to classical CSE problems (digital twin modeling, dimension reduction, mode decomposition, etc.).
The random order streaming model has been very fruitful for graph streams, allowing for polylogarithmic or even constant space estimators for fundamental graph problems such as matching size estimation, counting the number of connected components and more. However, the assumption that there are no correlations between the order of arrival of edges in the stream is quite strong. In this paper we introduce (hidden) batch random order streams, where edges are grouped in "batches" (which are unknown to the algorithm) that arrive in a random order, as a natural framework for modelling hidden correlations in the arrival order of edges, and present algorithms and lower bounds for this model. On the algorithmic side, we show how known techniques for connected component counting in constant space due to Peng and Sohler [SODA `18] easily translate from random order streams to our model with only a small loss in parameters. Our algorithm obtains an additive $\varepsilon n$ approximation to the number of connected components in the input graph using space $(1/\varepsilon)^{O(1/\varepsilon)}$ by building a representative sample of vertices in the graph that belong to $O(1/\varepsilon)$-size components to estimate the count. On the lower bound side, we show that $(1/\varepsilon)^{\Omega(1/\varepsilon)}$ space is necessary for finding a connected component of size $O(1/\varepsilon)$ even in graphs where most vertices reside in such components -- this makes progress towards an open problem of Peng and Sohler [SODA `18] and constitutes our main technical contribution. The lower bound uses Fourier analytic techniques inspired by the Boolean Hidden Matching problem. Our main innovation here is the first framework for applying such a Fourier analytic approach to a communication game with a polynomial number of players.
Let $G$ be a multigraph with $n$ vertices and $e>4n$ edges, drawn in the plane such that any two parallel edges form a simple closed curve with at least one vertex in its interior and at least one vertex in its exterior. Pach and T\'oth (A Crossing Lemma for Multigraphs, SoCG 2018) extended the Crossing Lemma of Ajtai et al. (Crossing-free subgraphs, North-Holland Mathematics Studies, 1982) and Leighton (Complexity issues in VLSI, Foundations of computing series, 1983) by showing that if no two adjacent edges cross and every pair of nonadjacent edges cross at most once, then the number of edge crossings in $G$ is at least $\alpha e^3/n^2$, for a suitable constant $\alpha>0$. The situation turns out to be quite different if nonparallel edges are allowed to cross any number of times. It is proved that in this case the number of crossings in $G$ is at least $\alpha e^{2.5}/n^{1.5}$. The order of magnitude of this bound cannot be improved.
Median graphs are connected graphs in which for all three vertices there is a unique vertex that belongs to shortest paths between each pair of these three vertices. In this paper we provide several novel characterizations of planar median graphs. More specifically, we characterize when a planar graph $G$ is a median graph in terms of forbidden subgraphs and the structure of isometric cycles in $G$, and also in terms of subgraphs of $G$ that are contained inside and outside of 4-cycles with respect to an arbitrary planar embedding of $G$. These results lead us to a new characterization of planar median graphs in terms of cubesquare-graphs that is, graphs that can be obtained by starting with cubes and square graphs, and iteratively replacing 4-cycle boundaries (relative to some embedding) by cubes or square-graphs. As a corollary we also show that a graph is planar median if and only if it can be obtained from cubes and square-graphs by a sequence of ``square-boundary'' amalgamations. These considerations also lead to an $\mathcal{O}(n\log n)$-time recognition algorithm to compute a decomposition of a planar median graph with $n$ vertices into cubes and square-graphs.
In the Online Machine Covering problem jobs, defined by their sizes, arrive one by one and have to be assigned to $m$ parallel and identical machines, with the goal of maximizing the load of the least-loaded machine. In this work, we study the Machine Covering problem in the recently popular random-order model. Here no extra resources are present, but instead the adversary is weakened in that it can only decide upon the input set while jobs are revealed uniformly at random. It is particularly relevant to Machine Covering where lower bounds are usually associated to highly structured input sequences. We first analyze Graham's Greedy-strategy in this context and establish that its competitive ratio decreases slightly to $\Theta\left(\frac{m}{\log(m)}\right)$ which is asymptotically tight. Then, as our main result, we present an improved $\tilde{O}(\sqrt[4]{m})$-competitive algorithm for the problem. This result is achieved by exploiting the extra information coming from the random order of the jobs, using sampling techniques to devise an improved mechanism to distinguish jobs that are relatively large from small ones. We complement this result with a first lower bound showing that no algorithm can have a competitive ratio of $O\left(\frac{\log(m)}{\log\log(m)}\right)$ in the random-order model. This lower bound is achieved by studying a novel variant of the Secretary problem, which could be of independent interest.
Let $\kappa(s,t)$ denote the maximum number of internally disjoint paths in an undirected graph $G$. We consider designing a data structure that includes a list of cuts, and answers in $O(1)$ time the following query: given $s,t \in V$, determine whether $\kappa(s,t) \leq k$, and if so, return a pointer to an $st$-cut of size $\leq k$ in the list. A trivial data structure includes a list of $n(n-1)/2$ cuts and requires $\Theta(kn^2)$ space. We show that $O(kn)$ cuts suffice, thus reducing the space to $O(k^2 n+n^2)$. In the case when $G$ is $k$-connected, we show that $O(n)$ cuts suffice, and that these cuts can be partitioned into $O(k)$ laminar families; this reduces the space to $O(kn)$. The latter result slightly improves and substantially simplifies a recent result of Pettie and Yin [ICALP 2021].