亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

The doubly minimized Petz Renyi mutual information of order $\alpha$ is defined as the minimization of the Petz divergence of order $\alpha$ of a fixed bipartite quantum state relative to any product state. In this work, we establish several properties of this type of Renyi mutual information, including its additivity for $\alpha\in [1/2,2]$. As an application, we show that the direct exponent of certain binary quantum state discrimination problems is determined by the doubly minimized Petz Renyi mutual information of order $\alpha\in (1/2,1)$. This provides an operational interpretation of this type of Renyi mutual information, and generalizes a previous result for classical probability distributions to the quantum setting.

相關內容

互(hu)信(xin)息(xi)(Mutual Information)是信(xin)息(xi)論里(li)一(yi)(yi)種(zhong)有用的(de)(de)信(xin)息(xi)度量(liang)(liang)(liang),它可以看成是一(yi)(yi)個(ge)隨(sui)(sui)機(ji)變(bian)量(liang)(liang)(liang)中包含(han)的(de)(de)關于另(ling)一(yi)(yi)個(ge)隨(sui)(sui)機(ji)變(bian)量(liang)(liang)(liang)的(de)(de)信(xin)息(xi)量(liang)(liang)(liang),或者(zhe)說是一(yi)(yi)個(ge)隨(sui)(sui)機(ji)變(bian)量(liang)(liang)(liang)由于已知另(ling)一(yi)(yi)個(ge)隨(sui)(sui)機(ji)變(bian)量(liang)(liang)(liang)而減少(shao)的(de)(de)不(bu)肯(ken)定(ding)性.

For decades, Simultaneous Ascending Auction (SAA) has been the most widely used mechanism for spectrum auctions, and it has recently gained popularity for allocating 5G licenses in many countries. Despite its relatively simple rules, SAA introduces a complex strategic game with an unknown optimal bidding strategy. Given the high stakes involved, with billions of euros sometimes on the line, developing an efficient bidding strategy is of utmost importance. In this work, we extend our previous method, a Simultaneous Move Monte-Carlo Tree Search (SM-MCTS) based algorithm named $SMS^{\alpha}$ to incomplete information framework. For this purpose, we compare three determinization approaches which allow us to rely on complete information SM-MCTS. This algorithm addresses, in incomplete framework, the four key strategic issues of SAA: the exposure problem, the own price effect, budget constraints, and the eligibility management problem. Through extensive numerical experiments on instances of realistic size with an uncertain framework, we show that $SMS^{\alpha}$ largely outperforms state-of-the-art algorithms by achieving higher expected utility while taking less risks, no matter which determinization method is chosen.

We consider the discretization of the $1d$-integral Dirichlet fractional Laplacian by $hp$-finite elements. We present quadrature schemes to set up the stiffness matrix and load vector that preserve the exponential convergence of $hp$-FEM on geometric meshes. The schemes are based on Gauss-Jacobi and Gauss-Legendre rules. We show that taking a number of quadrature points slightly exceeding the polynomial degree is enough to preserve root exponential convergence. The total number of algebraic operations to set up the system is $\mathcal{O}(N^{5/2})$, where $N$ is the problem size. Numerical example illustrate the analysis. We also extend our analysis to the fractional Laplacian in higher dimensions for $hp$-finite element spaces based on shape regular meshes.

The efficacy of interpolating via Variably Scaled Kernels (VSKs) is known to be dependent on the definition of a proper scaling function, but no numerical recipes to construct it are available. Previous works suggest that such a function should mimic the target one, but no theoretical evidence is provided. This paper fills both the gaps: it proves that a scaling function reflecting the target one may lead to enhanced approximation accuracy, and it provides a user-independent tool for learning the scaling function by means of Discontinuous Neural Networks ({\delta}NN), i.e., NNs able to deal with possible discontinuities. Numerical evidence supports our claims, as it shows that the key features of the target function can be clearly recovered in the learned scaling function.

We present a novel and flexible framework for localized tuning of Hamiltonian Monte Carlo (HMC) samplers by Gibbs sampling the algorithm's tuning parameters conditionally based on the position and momentum at each step. For adaptively sampling path lengths, the framework encompasses randomized HMC, multinomial HMC, the No-U-Turn Sampler (NUTS), and the Apogee-to-Apogee Path Sampler as special cases. The Gibbs self-tuning (GIST) framework is illustrated with an alternative to NUTS for locally adapting path lengths, evaluated with an exact Hamiltonian for an ill-conditioned normal and with the leapfrog algorithm for a test suite of diverse models.

The discretization of fluid-poromechanics systems is typically highly demanding in terms of computational effort. This is particularly true for models of multiphysics flows in the brain, due to the geometrical complexity of the cerebral anatomy - requiring a very fine computational mesh for finite element discretization - and to the high number of variables involved. Indeed, this kind of problems can be modeled by a coupled system encompassing the Stokes equations for the cerebrospinal fluid in the brain ventricles and Multiple-network Poro-Elasticity (MPE) equations describing the brain tissue, the interstitial fluid, and the blood vascular networks at different space scales. The present work aims to rigorously derive a posteriori error estimates for the coupled Stokes-MPE problem, as a first step towards the design of adaptive refinement strategies or reduced order models to decrease the computational demand of the problem. Through numerical experiments, we verify the reliability and optimal efficiency of the proposed a posteriori estimator and identify the role of the different solution variables in its composition.

\v{C}ech Persistence diagrams (PDs) are topological descriptors routinely used to capture the geometry of complex datasets. They are commonly compared using the Wasserstein distances $OT_{p}$; however, the extent to which PDs are stable with respect to these metrics remains poorly understood. We partially close this gap by focusing on the case where datasets are sampled on an $m$-dimensional submanifold of $\mathbb{R}^{d}$. Under this manifold hypothesis, we show that convergence with respect to the $OT_{p}$ metric happens exactly when $p\gt m$. We also provide improvements upon the bottleneck stability theorem in this case and prove new laws of large numbers for the total $\alpha$-persistence of PDs. Finally, we show how these theoretical findings shed new light on the behavior of the feature maps on the space of PDs that are used in ML-oriented applications of Topological Data Analysis.

The use of trajectory data with abundant spatial-temporal information is pivotal in Intelligent Transport Systems (ITS) and various traffic system tasks. Location-Based Services (LBS) capitalize on this trajectory data to offer users personalized services tailored to their location information. However, this trajectory data contains sensitive information about users' movement patterns and habits, necessitating confidentiality and protection from unknown collectors. To address this challenge, privacy-preserving methods like K-anonymity and Differential Privacy have been proposed to safeguard private information in the dataset. Despite their effectiveness, these methods can impact the original features by introducing perturbations or generating unrealistic trajectory data, leading to suboptimal performance in downstream tasks. To overcome these limitations, we propose a Federated Variational AutoEncoder (FedVAE) approach, which effectively generates a new trajectory dataset while preserving the confidentiality of private information and retaining the structure of the original features. In addition, FedVAE leverages Variational AutoEncoder (VAE) to maintain the original feature space and generate new trajectory data, and incorporates Federated Learning (FL) during the training stage, ensuring that users' data remains locally stored to protect their personal information. The results demonstrate its superior performance compared to other existing methods, affirming FedVAE as a promising solution for enhancing data privacy and utility in location-based applications.

Since the work of Polyanskiy, Poor and Verd\'u on the finite blocklength performance of capacity-achieving codes for discrete memoryless channels, many papers have attempted to find further results for more practically relevant channels. However, it seems that the complexity of computing capacity-achieving codes has not been investigated until now. We study this question for the simplest non-trivial Gaussian channels, i.e., the additive colored Gaussian noise channel. To assess the computational complexity, we consider the classes $\mathrm{FP}_1$ and $\#\mathrm{P}_1$. $\mathrm{FP}_1$ includes functions computable by a deterministic Turing machine in polynomial time, whereas $\#\mathrm{P}_1$ encompasses functions that count the number of solutions verifiable in polynomial time. It is widely assumed that $\mathrm{FP}_1\neq\#\mathrm{P}_1$. It is of interest to determine the conditions under which, for a given $M \in \mathbb{N}$, where $M$ describes the precision of the deviation of $C(P,N)$, for a certain blocklength $n_M$ and a decoding error $\epsilon > 0$ with $\epsilon\in\mathbb{Q}$, the following holds: $R_{n_M}(\epsilon)>C(P,N)-\frac{1}{2^M}$. It is shown that there is a polynomial-time computable $N_*$ such that for sufficiently large $P_*\in\mathbb{Q}$, the sequences $\{R_{n_M}(\epsilon)\}_{{n_M}\in\mathbb{N}}$, where each $R_{n_M}(\epsilon)$ satisfies the previous condition, cannot be computed in polynomial time if $\mathrm{FP}_1\neq\#\mathrm{P}_1$. Hence, the complexity of computing the sequence $\{R_{n_M}(\epsilon)\}_{n_M\in\mathbb{N}}$ grows faster than any polynomial as $M$ increases. Consequently, it is shown that either the sequence of achievable rates $\{R_{n_M}(\epsilon)\}_{n_M\in\mathbb{N}}$ as a function of the blocklength, or the sequence of blocklengths $\{n_M\}_{M\in\mathbb{N}}$ corresponding to the achievable rates, is not a polynomial-time computable sequence.

The present article aims to design and analyze efficient first-order strong schemes for a generalized A\"{i}t-Sahalia type model arising in mathematical finance and evolving in a positive domain $(0, \infty)$, which possesses a diffusion term with superlinear growth and a highly nonlinear drift that blows up at the origin. Such a complicated structure of the model unavoidably causes essential difficulties in the construction and convergence analysis of time discretizations. By incorporating implicitness in the term $\alpha_{-1} x^{-1}$ and a corrective mapping $\Phi_h$ in the recursion, we develop a novel class of explicit and unconditionally positivity-preserving (i.e., for any step-size $h>0$) Milstein-type schemes for the underlying model. In both non-critical and general critical cases, we introduce a novel approach to analyze mean-square error bounds of the novel schemes, without relying on a priori high-order moment bounds of the numerical approximations. The expected order-one mean-square convergence is attained for the proposed scheme. The above theoretical guarantee can be used to justify the optimal complexity of the Multilevel Monte Carlo method. Numerical experiments are finally provided to verify the theoretical findings.

In this paper, we compute stiffness matrix of the nonlocal Laplacian discretized by the piecewise linear finite element on nonuniform meshes, and implement the FEM in the Fourier transformed domain. We derive useful integral expressions of the entries that allow us to explicitly or semi-analytically evaluate the entries for various interaction kernels. Moreover, the limiting cases of the nonlocal stiffness matrix when the interactional radius $\delta\rightarrow0$ or $\delta\rightarrow\infty$ automatically lead to integer and fractional FEM stiffness matrices, respectively, and the FEM discretisation is intrinsically compatible. We conduct ample numerical experiments to study and predict some of its properties and test on different types of nonlocal problems. To the best of our knowledge, such a semi-analytic approach has not been explored in literature even in the one-dimensional case.

北京阿比特科技有限公司