亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Discretization of the uniform norm of functions from a given finite dimensional subspace of continuous functions is studied. We pay special attention to the case of trigonometric polynomials with frequencies from an arbitrary finite set with fixed cardinality. We give two different proofs of the fact that for any $N$-dimensional subspace of the space of continuous functions it is sufficient to use $e^{CN}$ sample points for an accurate upper bound for the uniform norm. Previous known results show that one cannot improve on the exponential growth of the number of sampling points for a good discretization theorem in the uniform norm. Also, we prove a general result, which connects the upper bound on the number of sampling points in the discretization theorem for the uniform norm with the best $m$-term bilinear approximation of the Dirichlet kernel associated with the given subspace. We illustrate application of our technique on the example of trigonometric polynomials.

相關內容

We study algorithmic applications of a natural discretization for the hard-sphere model and the Widom-Rowlinson model in a region $\mathbb{V}\subset\mathbb{R}^d$. These models are used in statistical physics to describe mixtures of one or multiple particle types subjected to hard-core interactions. For each type, particles follow a Poisson point process with a type specific activity parameter (fugacity). The Gibbs distribution is characterized by the mixture of these point processes conditioned that no two particles are closer than a type-dependent distance threshold. A key part in better understanding the Gibbs distribution is its normalizing constant, called partition function. We give sufficient conditions that the partition function of a discrete hard-core model on a geometric graph based on a point set $X \subset \mathbb{V}$ closely approximates those of such continuous models. Previously, this was only shown for the hard-sphere model on cubic regions $\mathbb{V}=[0, \ell)^d$ when $X$ is exponential in the volume of the region $\nu(\mathbb{V})$, limiting algorithmic applications. In the same setting, our refined analysis only requires a quadratic number of points, which we argue to be tight. We use our improved discretization results to approximate the partition functions of the hard-sphere model and the Widom-Rowlinson efficiently in $\nu(\mathbb{V})$. For the hard-sphere model, we obtain the first quasi-polynomial deterministic approximation algorithm for the entire fugacity regime for which, so far, only randomized approximations are known. Furthermore, we simplify a recently introduced fully polynomial randomized approximation algorithm. Similarly, we obtain the best known deterministic and randomized approximation bounds for the Widom-Rowlinson model. Moreover, we obtain approximate sampling algorithms for the respective spin systems within the same fugacity regimes.

We build stochastic models for analyzing Jaya and semi-steady-state Jaya algorithms. The analysis shows that for semi-steady-state Jaya (a) the maximum expected value of the number of worst-index updates per generation is a paltry 1.7 regardless of the population size; (b) regardless of the population size, the expectation of the number of best-index updates per generation decreases monotonically with generations; (c) exact upper bounds as well as asymptotics of the expected best-update counts can be obtained for specific distributions; the upper bound is 0.5 for normal and logistic distributions, $\ln 2$ for the uniform distribution, and $e^{-\gamma} \ln 2$ for the exponential distribution, where $\gamma$ is the Euler-Mascheroni constant; the asymptotic is $e^{-\gamma} \ln 2$ for logistic and exponential distributions and $\ln 2$ for the uniform distribution (the asymptotic cannot be obtained analytically for the normal distribution). The models lead to the derivation of computational complexities of Jaya and semi-steady-state Jaya. The theoretical analysis is supported with empirical results on a benchmark suite. The insights provided by our stochastic models should help design new, improved population-based search/optimization heuristics.

This paper proposes a regularization of the Monge-Amp\`ere equation in planar convex domains through uniformly elliptic Hamilton-Jacobi-Bellman equations. The regularized problem possesses a unique strong solution $u_\varepsilon$ and is accessible to the discretization with finite elements. This work establishes locally uniform convergence of $u_\varepsilon$ to the convex Alexandrov solution $u$ to the Monge-Amp\`ere equation as the regularization parameter $\varepsilon$ approaches $0$. A mixed finite element method for the approximation of $u_\varepsilon$ is proposed, and the regularized finite element scheme is shown to be locally uniformly convergent. Numerical experiments provide empirical evidence for the efficient approximation of singular solutions $u$.

A singularly perturbed parabolic problem of convection-diffusion type with a discontinuous initial condition is examined. An analytic function is identified which matches the discontinuity in the initial condition and also satisfies the homogenous parabolic differential equation associated with the problem. The difference between this analytical function and the solution of the parabolic problem is approximated numerically, using an upwind finite difference operator combined with an appropriate layer-adapted mesh. The numerical method is shown to be parameter-uniform. Numerical results are presented to illustrate the theoretical error bounds established in the paper.

We consider the problem of sparse normal means estimation in a distributed setting with communication constraints. We assume there are $M$ machines, each holding $d$-dimensional observations of a $K$-sparse vector $\mu$ corrupted by additive Gaussian noise. The $M$ machines are connected in a star topology to a fusion center, whose goal is to estimate the vector $\mu$ with a low communication budget. Previous works have shown that to achieve the centralized minimax rate for the $\ell_2$ risk, the total communication must be high - at least linear in the dimension $d$. This phenomenon occurs, however, at very weak signals. We show that at signal-to-noise ratios (SNRs) that are sufficiently high - but not enough for recovery by any individual machine - the support of $\mu$ can be correctly recovered with significantly less communication. Specifically, we present two algorithms for distributed estimation of a sparse mean vector corrupted by either Gaussian or sub-Gaussian noise. We then prove that above certain SNR thresholds, with high probability, these algorithms recover the correct support with total communication that is sublinear in the dimension $d$. Furthermore, the communication decreases exponentially as a function of signal strength. If in addition $KM\ll \tfrac{d}{\log d}$, then with an additional round of sublinear communication, our algorithms achieve the centralized rate for the $\ell_2$ risk. Finally, we present simulations that illustrate the performance of our algorithms in different parameter regimes.

We study norm-based uniform convergence bounds for neural networks, aiming at a tight understanding of how these are affected by the architecture and type of norm constraint, for the simple class of scalar-valued one-hidden-layer networks, and inputs bounded in Euclidean norm. We begin by proving that in general, controlling the spectral norm of the hidden layer weight matrix is insufficient to get uniform convergence guarantees (independent of the network width), while a stronger Frobenius norm control is sufficient, extending and improving on previous work. Motivated by the proof constructions, we identify and analyze two important settings where a mere spectral norm control turns out to be sufficient: First, when the network's activation functions are sufficiently smooth (with the result extending to deeper networks); and second, for certain types of convolutional networks. In the latter setting, we study how the sample complexity is additionally affected by parameters such as the amount of overlap between patches and the overall number of patches.

In this paper, we study the convergence properties of off-policy policy improvement algorithms with state-action density ratio correction under function approximation setting, where the objective function is formulated as a max-max-min optimization problem. We characterize the bias of the learning objective and present two strategies with finite-time convergence guarantees. In our first strategy, we present algorithm P-SREDA with convergence rate $O(\epsilon^{-3})$, whose dependency on $\epsilon$ is optimal. In our second strategy, we propose a new off-policy actor-critic style algorithm named O-SPIM. We prove that O-SPIM converges to a stationary point with total complexity $O(\epsilon^{-4})$, which matches the convergence rate of some recent actor-critic algorithms in the on-policy setting.

In micro-fluidics not only does capillarity dominate but also thermal fluctuations become important. On the level of the lubrication approximation, this leads to a quasi-linear fourth-order parabolic equation for the film height $h$ driven by space-time white noise. The gradient flow structure of its deterministic counterpart, the thin-film equation, which encodes the balance between driving capillary and limiting viscous forces, provides the guidance for the thermodynamically consistent introduction of fluctuations. We follow this route on the level of a spatial discretization of the gradient flow structure. Starting from an energetically conformal finite-element (FE) discretization, we point out that the numerical mobility function introduced by Gr\"un and Rumpf can be interpreted as a discretization of the metric tensor in the sense of a mixed FE method with lumping. While this discretization was devised in order to preserve the so-called entropy estimate, we use this to show that the resulting high-dimensional stochastic differential equation (SDE) preserves pathwise and pointwise strict positivity, at least in case of the physically relevant mobility function arising from the no-slip boundary condition. As a consequence, this discretization gives rise to a consistent invariant measure, namely a discretization of the Brownian excursion (up to the volume constraint), and thus features an entropic repulsion. The price to pay over more naive discretizations is that when writing the SDE in It\^o's form, which is the basis for the Euler-Mayurama time discretization, a correction term appears. To conclude, we perform various numerical experiments to compare the behavior of our discretization to that of the more naive finite difference discretization of the equation.

We study the query complexity of finding a Tarski fixed point over the $k$-dimensional grid $\{1,\ldots,n\}^k$. Improving on the previous best upper bound of $\smash{O(\log^{2\lceil k/3\rceil} n)}$ [FPS20], we give a new algorithm with query complexity $\smash{O(\log^{\lceil (k+1)/2\rceil} n)}$. This is based on a novel decomposition theorem about a weaker variant of the Tarski fixed point problem, where the input consists of a monotone function $f:[n]^k\rightarrow [n]^k$ and a monotone sign function $b:[n]^k\rightarrow \{-1,0,1\}$ and the goal is to find an $x\in [n]^k$ that satisfies $either$ $f(x)\preceq x$ and $b(x)\le 0$ $or$ $f(x)\succeq x$ and $b(x)\ge 0$.

To learn intrinsic low-dimensional structures from high-dimensional data that most discriminate between classes, we propose the principle of Maximal Coding Rate Reduction ($\text{MCR}^2$), an information-theoretic measure that maximizes the coding rate difference between the whole dataset and the sum of each individual class. We clarify its relationships with most existing frameworks such as cross-entropy, information bottleneck, information gain, contractive and contrastive learning, and provide theoretical guarantees for learning diverse and discriminative features. The coding rate can be accurately computed from finite samples of degenerate subspace-like distributions and can learn intrinsic representations in supervised, self-supervised, and unsupervised settings in a unified manner. Empirically, the representations learned using this principle alone are significantly more robust to label corruptions in classification than those using cross-entropy, and can lead to state-of-the-art results in clustering mixed data from self-learned invariant features.

北京阿比特科技有限公司