A recent line of work has shown the unconditional advantage of constant-depth quantum computation, or $\mathsf{QNC^0}$, over $\mathsf{NC^0}$, $\mathsf{AC^0}$, and related models of classical computation. Problems exhibiting this advantage include search and sampling tasks related to the parity function, and it is natural to ask whether $\mathsf{QNC^0}$ can be used to help compute parity itself. We study $\mathsf{AC^0\circ QNC^0}$ -- a hybrid circuit model where $\mathsf{AC^0}$ operates on measurement outcomes of a $\mathsf{QNC^0}$ circuit, and conjecture $\mathsf{AC^0\circ QNC^0}$ cannot even achieve $\Omega(1)$ correlation with parity. As evidence for this conjecture, we prove: $\bullet$ When the $\mathsf{QNC^0}$ circuit is ancilla-free, this model achieves only negligible correlation with parity. $\bullet$ For the general (non-ancilla-free) case, we show via a connection to nonlocal games that the conjecture holds for any class of postprocessing functions that has approximate degree $o(n)$ and is closed under restrictions, even when the $\mathsf{QNC^0}$ circuit is given arbitrary quantum advice. By known results this confirms the conjecture for linear-size $\mathsf{AC^0}$ circuits. $\bullet$ Towards the a switching lemma for $\mathsf{AC^0\circ QNC^0}$, we study the effect of quantum preprocessing on the decision tree complexity of Boolean functions. We find that from this perspective, nonlocal channels are no better than randomness: a Boolean function $f$ precomposed with an $n$-party nonlocal channel is together equal to a randomized decision tree with worst-case depth at most $\mathrm{DT}_\mathrm{depth}[f]$. Our results suggest that while $\mathsf{QNC^0}$ is surprisingly powerful for search and sampling, that power is "locked away" in the global correlations of its output, inaccessible to simple classical computation for solving decision problems.
In recent years, few-shot action recognition has attracted increasing attention. It generally adopts the paradigm of meta-learning. In this field, overcoming the overlapping distribution of classes and outliers is still a challenging problem based on limited samples. We believe the combination of Multi-modal and Multi-view can improve this issue depending on information complementarity. Therefore, we propose a method of Multi-view Distillation based on Multi-modal Fusion. Firstly, a Probability Prompt Selector for the query is constructed to generate probability prompt embedding based on the comparison score between the prompt embeddings of the support and the visual embedding of the query. Secondly, we establish a Multi-view. In each view, we fuse the prompt embedding as consistent information with visual and the global or local temporal context to overcome the overlapping distribution of classes and outliers. Thirdly, we perform the distance fusion for the Multi-view and the mutual distillation of matching ability from one to another, enabling the model to be more robust to the distribution bias. Our code is available at the URL: \url{//github.com/cofly2014/MDMF}.
Generating random and pseudorandom numbers with a deterministic system is a long-standing challenge in theoretical research and engineering applications. Several pseudorandom number generators based on the inversive congruential method have been designed as attractive alternatives to those based on the classical linear congruential method. This paper discloses the least period of sequences generated by iterating an inversive pseudorandom number generator over the ring $\mathbb{Z}_e$ by transforming it into a two-order linear congruential recurrence relation. Depending on whether the sequence is periodic or ultimately periodic, all states in the domain can be attributed to two types of objects: some cycles of different lengths and one unilateral connected digraph whose structure remains unchanged concerning parameter $e$. The graph structure of the generator over the ring $\mathbb{Z}_e$ is precisely disclosed with rigorous theoretical analysis and verified experimentally. The adopted analysis methodology can be extended to study the graph structure of other nonlinear maps.
Subshifts are colorings of $\mathbb{Z}^d$ defined by families of forbidden patterns. Given a subshift and a finite pattern, its extender set is the set of admissible completions of this pattern. It has been conjectured that the behavior of extender sets, and in particular their growth called extender entropy (arXiv:1711.07515), could provide a way to separate the classes of sofic and effective subshifts. We prove here that both classes have the same possible extender entropies: exactly the $\Pi_3$ real numbers of $[0,+\infty)$. We also consider computational properties of extender entropies for subshifts with some language or dynamical properties: computable language, minimal and some mixing properties.
This work presents an optimization-based scalable quantum neural network framework for approximating $n$-qubit unitaries through generic parametric representation of unitaries, which are obtained as product of exponential of basis elements of a new basis that we propose as an alternative to Pauli string basis. We call this basis as the Standard Recursive Block Basis, which is constructed using a recursive method, and its elements are permutation-similar to block Hermitian unitary matrices.
We address here spanning tree problems on a graph with binary edge weights. For a general weighted graph the minimum spanning tree is solved in super-linear running time, even when the edges of the graph are pre-sorted. A related problem, of finding a spanning tree with a pre-specified sum of weights, is NP-hard. In contrast, for a graph with binary weights associated with the edges, it is shown that the minimum spanning tree and finding a spanning tree with a given total sum, are solvable in linear time with simple algorithms.
In this paper, we develop an accurate pseudospectral method to approximate numerically the Riesz-Feller operator $D_\gamma^\alpha$ on $\mathbb R$, where $\alpha\in(0,2)$, and $|\gamma|\le\min\{\alpha, 2 - \alpha\}$. This operator can be written as a linear combination of the Weyl-Marchaud derivatives $\mathcal{D}^{\alpha}$ and $\overline{\mathcal{D}^\alpha}$, when $\alpha\in(0,1)$, and of $\partial_x\mathcal{D}^{\alpha-1}$ and $\partial_x\overline{\mathcal{D}^{\alpha-1}}$, when $\alpha\in(1,2)$. Given the so-called Higgins functions $\lambda_k(x) = ((ix-1)/(ix+1))^k$, where $k\in\mathbb Z$, we compute explicitly, using complex variable techniques, $\mathcal{D}^{\alpha}[\lambda_k](x)$, $\overline{\mathcal{D}^\alpha}[\lambda_k](x)$, $\partial_x\mathcal{D}^{\alpha-1}[\lambda_k](x)$, $\partial_x\overline{\mathcal{D}^{\alpha-1}}[\lambda_k](x)$ and $D_\gamma^\alpha[\lambda_k](x)$, in terms of the Gaussian hypergeometric function ${}_2F_1$, and relate these results to previous ones for the fractional Laplacian. This enables us to approximate $\mathcal{D}^{\alpha}[u](x)$, $\overline{\mathcal{D}^\alpha}[u](x)$, $\partial_x\mathcal{D}^{\alpha-1}[u](x)$, $\partial_x\overline{\mathcal{D}^{\alpha-1}}[u](x)$ and $D_\gamma^\alpha[u](x)$, for bounded continuous functions $u(x)$. Finally, we simulate a nonlinear Riesz-Feller fractional diffusion equation, characterized by having front propagating solutions whose speed grows exponentially in time.
The hazard function represents one of the main quantities of interest in the analysis of survival data. We propose a general approach for parametrically modelling the dynamics of the hazard function using systems of autonomous ordinary differential equations (ODEs). This modelling approach can be used to provide qualitative and quantitative analyses of the evolution of the hazard function over time. Our proposal capitalises on the extensive literature of ODEs which, in particular, allow for establishing basic rules or laws on the dynamics of the hazard function via the use of autonomous ODEs. We show how to implement the proposed modelling framework in cases where there is an analytic solution to the system of ODEs or where an ODE solver is required to obtain a numerical solution. We focus on the use of a Bayesian modelling approach, but the proposed methodology can also be coupled with maximum likelihood estimation. A simulation study is presented to illustrate the performance of these models and the interplay of sample size and censoring. Two case studies using real data are presented to illustrate the use of the proposed approach and to highlight the interpretability of the corresponding models. We conclude with a discussion on potential extensions of our work and strategies to include covariates into our framework.
We consider nonlinear solvers for the incompressible, steady (or at a fixed time step for unsteady) Navier-Stokes equations in the setting where partial measurement data of the solution is available. The measurement data is incorporated/assimilated into the solution through a nudging term addition to the the Picard iteration that penalized the difference between the coarse mesh interpolants of the true solution and solver solution, analogous to how continuous data assimilation (CDA) is implemented for time dependent PDEs. This was considered in the paper [Li et al. {\it CMAME} 2023], and we extend the methodology by improving the analysis to be in the $L^2$ norm instead of a weighted $H^1$ norm where the weight depended on the coarse mesh width, and to the case of noisy measurement data. For noisy measurement data, we prove that the CDA-Picard method is stable and convergent, up to the size of the noise. Numerical tests illustrate the results, and show that a very good strategy when using noisy data is to use CDA-Picard to generate an initial guess for the classical Newton iteration.
We propose a Block Majorization Minimization method with Extrapolation (BMMe) for solving a class of multi-convex optimization problems. The extrapolation parameters of BMMe are updated using a novel adaptive update rule. By showing that block majorization minimization can be reformulated as a block mirror descent method, with the Bregman divergence adaptively updated at each iteration, we establish subsequential convergence for BMMe. We use this method to design efficient algorithms to tackle nonnegative matrix factorization problems with the $\beta$-divergences ($\beta$-NMF) for $\beta\in [1,2]$. These algorithms, which are multiplicative updates with extrapolation, benefit from our novel results that offer convergence guarantees. We also empirically illustrate the significant acceleration of BMMe for $\beta$-NMF through extensive experiments.
The advent of large-scale inference has spurred reexamination of conventional statistical thinking. In a Gaussian model for $n$ many $z$-scores with at most $k < \frac{n}{2}$ nonnulls, Efron suggests estimating the location and scale parameters of the null distribution. Placing no assumptions on the nonnull effects, the statistical task can be viewed as a robust estimation problem. However, the best known robust estimators fail to be consistent in the regime $k \asymp n$ which is especially relevant in large-scale inference. The failure of estimators which are minimax rate-optimal with respect to other formulations of robustness (e.g. Huber's contamination model) might suggest the impossibility of consistent estimation in this regime and, consequently, a major weakness of Efron's suggestion. A sound evaluation of Efron's model thus requires a complete understanding of consistency. We sharply characterize the regime of $k$ for which consistent estimation is possible and further establish the minimax estimation rates. It is shown consistent estimation of the location parameter is possible if and only if $\frac{n}{2} - k = \omega(\sqrt{n})$, and consistent estimation of the scale parameter is possible in the entire regime $k < \frac{n}{2}$. Faster rates than those in Huber's contamination model are achievable by exploiting the Gaussian character of the data. The minimax upper bound is obtained by considering estimators based on the empirical characteristic function. The minimax lower bound involves constructing two marginal distributions whose characteristic functions match on a wide interval containing zero. The construction notably differs from those in the literature by sharply capturing a scaling of $n-2k$ in the minimax estimation rate of the location.