We propose $\mathbb{VD}$-$\mathbb{GR}$ - a novel visual dialog model that combines pre-trained language models (LMs) with graph neural networks (GNNs). Prior works mainly focused on one class of models at the expense of the other, thus missing out on the opportunity of combining their respective benefits. At the core of $\mathbb{VD}$-$\mathbb{GR}$ is a novel integration mechanism that alternates between spatial-temporal multi-modal GNNs and BERT layers, and that covers three distinct contributions: First, we use multi-modal GNNs to process the features of each modality (image, question, and dialog history) and exploit their local structures before performing BERT global attention. Second, we propose hub-nodes that link to all other nodes within one modality graph, allowing the model to propagate information from one GNN (modality) to the other in a cascaded manner. Third, we augment the BERT hidden states with fine-grained multi-modal GNN features before passing them to the next $\mathbb{VD}$-$\mathbb{GR}$ layer. Evaluations on VisDial v1.0, VisDial v0.9, VisDialConv, and VisPro show that $\mathbb{VD}$-$\mathbb{GR}$ achieves new state-of-the-art results across all four datasets.
Formalized $1$-category theory forms a core component of various libraries of mathematical proofs. However, more sophisticated results in fields from algebraic topology to theoretical physics, where objects have "higher structure," rely on infinite-dimensional categories in place of $1$-dimensional categories, and $\infty$-category theory has thusfar proved unamenable to computer formalization. Using a new proof assistant called Rzk, which is designed to support Riehl--Shulman's simplicial extension of homotopy type theory for synthetic $\infty$-category theory, we provide the first formalizations of results from $\infty$-category theory. This includes in particular a formalization of the Yoneda lemma, often regarded as the fundamental theorem of category theory, a theorem which roughly states that an object of a given category is determined by its relationship to all of the other objects of the category. A key feature of our framework is that, thanks to the synthetic theory, many constructions are automatically natural or functorial. We plan to use Rzk to formalize further results from $\infty$-category theory, such as the theory of limits and colimits and adjunctions.
Shape is a powerful tool to understand point sets. A formal notion of shape is given by $\alpha$-shapes, which generalize the convex hull and provide adjustable level of detail. Many real-world point sets have an inherent temporal property as natural processes often happen over time, like lightning strikes during thunderstorms or moving animal swarms. To explore such point sets, where each point is associated with one timestamp, interactive applications may utilize $\alpha$-shapes and allow the user to specify different time windows and $\alpha$-values. We show how to compute the temporal $\alpha$-shape $\alpha_T$, a minimal description of all $\alpha$-shapes over all time windows, in output-sensitive linear time. We also give complexity bounds on $|\alpha_T|$. We use $\alpha_T$ to interactively visualize $\alpha$-shapes of user-specified time windows without having to constantly compute requested $\alpha$-shapes. Experimental results suggest that our approach outperforms an existing approach by a factor of at least $\sim$52 and that the description we compute has reasonable size in practice. The basis for our algorithm is an existing algorithm which computes all Delaunay triangles over all time windows using $\mathcal{O}(1)$ time per triangle. Our approach generalizes to higher dimensions with the same runtime for fixed $d$.
In this paper, we study the algebraic structure of $(\sigma,\delta)$-polycyclic codes as submodules in the quotient module $S/Sf$, where $S=R[x,\sigma,\delta]$ is the Ore extension, $f\in S$, and $R$ is a finite but not necessarily commutative ring. We establish that the Euclidean duals of $(\sigma,\delta)$-polycyclic codes are $(\sigma,\delta)$-sequential codes. By using $(\sigma,\delta)$-Pseudo Linear Transformation (PLT), we define the annihilator dual of $(\sigma,\delta)$-polycyclic codes. Then, we demonstrate that the annihilator duals of $(\sigma,\delta)$-polycyclic codes maintain their $(\sigma,\delta)$-polycyclic nature. Furthermore, we classify when two $(\sigma,\delta)$-polycyclic codes are Hamming isometrical equivalent. By employing Wedderburn polynomials, we introduce simple-root $(\sigma,\delta)$-polycyclic codes. Subsequently, we define the $(\sigma, \delta)$-Mattson-Solomon transform for this class of codes and we address the problem of decomposing these codes by using the properties of Wedderburn polynomials.
We introduce a joint posterior $p$-value, an extension of the posterior predictive $p$-value for multiple test statistics, designed to address limitations of existing Bayesian $p$-values in the setting of continuous model expansion. In particular, we show that the posterior predictive $p$-value, as well as its sampled variant, become more conservative as the parameter dimension grows, and we demonstrate the ability of the joint $p$-value to overcome this problem in cases where we can select test statistics that are negatively associated under the posterior. We validate these conclusions with a pair of simulation examples in which the joint $p$-value achieves substantial gains to power with only a modest increase in computational cost.
A 2018 conjecture of Brewster, McGuinness, Moore, and Noel asserts that for $k \ge 3$, if a graph has chromatic number greater than $k$, then it contains at least as many cycles of length $0 \bmod k$ as the complete graph on $k+1$ vertices. Our main result confirms this in the $k=3$ case by showing every $4$-critical graph contains at least $4$ cycles of length $0 \bmod 3$, and that $K_4$ is the unique such graph achieving the minimum. We make progress on the general conjecture as well, showing that $(k+1)$-critical graphs with minimum degree $k$ have at least as many cycles of length $0\bmod r$ as $K_{k+1}$, provided $k+1 \ne 0 \bmod r$. We also show that $K_{k+1}$ uniquely minimizes the number of cycles of length $1\bmod k$ among all $(k+1)$-critical graphs, strengthening a recent result of Moore and West and extending it to the $k=3$ case.
Let $-A$ be the generator of a bounded $C_0$-semigroup $(e^{-tA})_{t \geq 0}$ on a Hilbert space. First we study the long-time asymptotic behavior of the Cayley transform $V_{\omega}(A) := (A-\omega I) (A+\omega I)^{-1}$ with $\omega >0$. We give a decay estimate for $\|V_{\omega}(A)^nA^{-1}\|$ when $(e^{-tA})_{t \geq 0}$ is polynomially stable. Considering the case where the parameter $\omega$ varies, we estimate $\|\prod_{k=1}^n (V_{\omega_k}(A))A^{-1}\|$ for exponentially stable $C_0$-semigroups $(e^{-tA})_{t \geq 0}$. Next we show that if the generator $-A$ of the bounded $C_0$-semigroup has a bounded inverse, then $\sup_{t \geq 0} \|e^{-tA^{-1}} A^{-\alpha} \| < \infty$ for all $\alpha >0$. We also present an estimate for the rate of decay of $\|e^{-tA^{-1}} A^{-1} \|$, assuming that $(e^{-tA})_{t \geq 0}$ is polynomially stable. To obtain these results, we use operator norm estimates offered by a functional calculus called the $\mathcal{B}$-calculus.
0-dimensional persistent homology is known, from a computational point of view, as the easy case. Indeed, given a list of $n$ edges in non-decreasing order of filtration value, one only needs a union-find data structure to keep track of the connected components and we get the persistence diagram in time $O(n\alpha(n))$. The running time is thus usually dominated by sorting the edges in $\Theta(n\log(n))$. A little-known fact is that, in the particularly simple case of studying the sublevel sets of a piecewise-linear function on $\mathbb{R}$ or $\mathbb{S}^1$, persistence can actually be computed in linear time. This note presents a simple algorithm that achieves this complexity and an extension to image persistence. An implementation is available in Gudhi.
In-context learning (ICL) is a new learning paradigm that has gained popularity along with the development of large language models. In this work, we adapt a recently proposed hardness metric, pointwise $\mathcal{V}$-usable information (PVI), to an in-context version (in-context PVI). Compared to the original PVI, in-context PVI is more efficient in that it requires only a few exemplars and does not require fine-tuning. We conducted a comprehensive empirical analysis to evaluate the reliability of in-context PVI. Our findings indicate that in-context PVI estimates exhibit similar characteristics to the original PVI. Specific to the in-context setting, we show that in-context PVI estimates remain consistent across different exemplar selections and numbers of shots. The variance of in-context PVI estimates across different exemplar selections is insignificant, which suggests that in-context PVI are stable. Furthermore, we demonstrate how in-context PVI can be employed to identify challenging instances. Our work highlights the potential of in-context PVI and provides new insights into the capabilities of ICL.
We say that a (multi)graph $G = (V,E)$ has geometric thickness $t$ if there exists a straight-line drawing $\varphi : V \rightarrow \mathbb{R}^2$ and a $t$-coloring of its edges where no two edges sharing a point in their relative interior have the same color. The Geometric Thickness problem asks whether a given multigraph has geometric thickness at most $t$. This problem was shown to be NP-hard for $t=2$ [Durocher, Gethner, and Mondal, CG 2016]. In this paper, we settle the computational complexity of Geometric Thickness by showing that it is $\exists \mathbb{R}$-complete already for thickness $57$. Moreover, our reduction shows that the problem is $\exists \mathbb{R}$-complete for $8280$-planar graphs, where a graph is $k$-planar if it admits a topological drawing with at most $k$ crossings per edge. In the course of our paper, we answer previous questions on the geometric thickness and on other related problems, in particular, that simultaneous graph embeddings of $58$ edge-disjoint graphs and pseudo-segment stretchability with chromatic number $57$ are $\exists \mathbb{R}$-complete.
We introduce a numerical scheme that approximates solutions to linear PDE's by minimizing a residual in the $W^{-1,p'}(\Omega)$ norm with exponents $p> 2$. The resulting problem is solved by regularized Kacanov iterations, allowing to compute the solution to the non-linear minimization problem even for large exponents $p\gg 2$. Such large exponents remedy instabilities of finite element methods for problems like convection-dominated diffusion.