亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

This paper is concerned with the phase estimation algorithm in quantum computing algorithms, especially the scenarios where (1) the input vector is not an eigenvector; (2) the unitary operator is not exactly implemented; (3) random approximations are used for the unitary operator, e.g., the QDRIFT method. We characterize the probability of computing the phase values in terms of the consistency error, including the residual error, Trotter splitting error, or statistical mean-square error.

相關內容

Quantum Annealing (QA) is a computational framework where a quantum system's continuous evolution is used to find the global minimum of an objective function over an unstructured search space. It can be seen as a general metaheuristic for optimization problems, including NP-hard ones if we allow an exponentially large running time. While QA is widely studied from a heuristic point of view, little is known about theoretical guarantees on the quality of the solutions obtained in polynomial time. In this paper we use a technique borrowed from theoretical physics, the Lieb-Robinson (LR) bound, and develop new tools proving that short, constant time quantum annealing guarantees constant factor approximations ratios for some optimization problems when restricted to bounded degree graphs. Informally, on bounded degree graphs the LR bound allows us to retrieve a (relaxed) locality argument, through which the approximation ratio can be deduced by studying subgraphs of bounded radius. We illustrate our tools on problems MaxCut and Maximum Independent Set for cubic graphs, providing explicit approximation ratios and the runtimes needed to obtain them. Our results are of similar flavor to the well-known ones obtained in the different but related QAOA (quantum optimization algorithms) framework. Eventually, we discuss theoretical and experimental arguments for further improvements.

The discovery of structure from time series data is a key problem in fields of study working with complex systems. Most identifiability results and learning algorithms assume the underlying dynamics to be discrete in time. Comparatively few, in contrast, explicitly define dependencies in infinitesimal intervals of time, independently of the scale of observation and of the regularity of sampling. In this paper, we consider score-based structure learning for the study of dynamical systems. We prove that for vector fields parameterized in a large class of neural networks, least squares optimization with adaptive regularization schemes consistently recovers directed graphs of local independencies in systems of stochastic differential equations. Using this insight, we propose a score-based learning algorithm based on penalized Neural Ordinary Differential Equations (modelling the mean process) that we show to be applicable to the general setting of irregularly-sampled multivariate time series and to outperform the state of the art across a range of dynamical systems.

We present an implicit-explicit finite volume scheme for isentropic two phase flow in all Mach number regimes. The underlying model belongs to the class of symmetric hyperbolic thermodynamically compatible models. The key element of the scheme consists of a linearisation of pressure and enthalpy terms at a reference state. The resulting stiff linear parts are integrated implicitly, whereas the non-linear higher order and transport terms are treated explicitly. Due to the flux splitting, the scheme is stable under a CFL condition which determined by the resolution of the slow material waves and allows large time steps even in the presence of fast acoustic waves. Further the singular Mach number limits of the model are studied and the asymptotic preserving property of the scheme is proven. In numerical simulations the consistency with single phase flow, accuracy and the approximation of material waves in different Mach number regimes are assessed.

Maximal parabolic $L^p$-regularity of linear parabolic equations on an evolving surface is shown by pulling back the problem to the initial surface and studying the maximal $L^p$-regularity on a fixed surface. By freezing the coefficients in the parabolic equations at a fixed time and utilizing a perturbation argument around the freezed time, it is shown that backward difference time discretizations of linear parabolic equations on an evolving surface along characteristic trajectories can preserve maximal $L^p$-regularity in the discrete setting. The result is applied to prove the stability and convergence of time discretizations of nonlinear parabolic equations on an evolving surface, with linearly implicit backward differentiation formulae characteristic trajectories of the surface, for general locally Lipschitz nonlinearities. The discrete maximal $L^p$-regularity is used to prove the boundedness and stability of numerical solutions in the $L^\infty(0,T;W^{1,\infty})$ norm, which is used to bound the nonlinear terms in the stability analysis. Optimal-order error estimates of time discretizations in the $L^\infty(0,T;W^{1,\infty})$ norm is obtained by combining the stability analysis with the consistency estimates.

We present substantially generalized and improved quantum algorithms over prior work for inhomogeneous linear and nonlinear ordinary differential equations (ODE). In Berry et al., (2017), a quantum algorithm for a certain class of linear ODEs is given, where the matrix involved needs to be diagonalizable. The quantum algorithm for linear ODEs presented here extends to many classes of non-diagonalizable matrices. The algorithm here can also be exponentially faster for certain classes of diagonalizable matrices. Our linear ODE algorithm is then applied to nonlinear differential equations using Carleman linearization (an approach taken recently by us in Liu et al., (2021)). The improvement over that result is two-fold. First, we obtain an exponentially better dependence on error. This kind of logarithmic dependence on error has also been achieved by Xue et al., (2021), but only for homogeneous nonlinear equations. Second, the present algorithm can handle any sparse, invertible matrix (that models dissipation) if it has a negative log-norm (including non-diagonalizable matrices), whereas Liu et al., (2021) and Xue et al., (2021) additionally require normality.

We propose a new wavelet-based method for density estimation when the data are size-biased. More specifically, we consider a power of the density of interest, where this power exceeds 1/2. Warped wavelet bases are employed, where warping is attained by some continuous cumulative distribution function. This can be seen as a general framework in which the conventional orthonormal wavelet estimation is the case where warping distribution is the standard uniform c.d.f. We show that both linear and nonlinear wavelet estimators are consistent, with optimal and/or near-optimal rates. Monte Carlo simulations are performed to compare four special settings which are easy to interpret in practice. An application with a real dataset on fatal traffic accidents involving alcohol illustrates the method. We observe that warped bases provide more flexible and superior estimates for both simulated and real data. Moreover, we find that estimating the power of a density (for instance, its square root) further improves the results.

The distributed convex optimization problem over the multi-agent system is considered in this paper, and it is assumed that each agent possesses its own cost function and communicates with its neighbours over a sequence of time-varying directed graphs. However, due to some reasons there exist communication delays while agents receive information from other agents, and we are going to seek the optimal value of the sum of agents' loss functions in this case. We desire to handle this problem with the push-sum distributed dual averaging (PS-DDA) algorithm. It is proved that this algorithm converges and the error decays at a rate $\mathcal{O}\left(T^{-0.5}\right)$ with proper step size, where $T$ is iteration span. The main result presented in this paper also illustrates the convergence of the proposed algorithm is related to the maximum value of the communication delay on one edge. We finally apply the theoretical results to numerical simulations to show the PS-DDA algorithm's performance.

In this paper, we focus on the construction of a hybrid scheme for the approximation of non-Maxwellian kinetic models with uncertainties. In the context of multiagent systems, the introduction of a kernel at the kinetic level is useful to avoid unphysical interactions. The methods here proposed, combine a direct simulation Monte Carlo (DSMC) in the phase space together with stochastic Galerkin (sG) methods in the random space. The developed schemes preserve the main physical properties of the solution together with accuracy in the random space. The consistency of the methods is tested with respect to surrogate Fokker-Planck models that can be obtained in the quasi-invariant regime of parameters. Several applications of the schemes to non-Maxwellian models of multiagent systems are reported.

In this work, we consider the distributed optimization of non-smooth convex functions using a network of computing units. We investigate this problem under two regularity assumptions: (1) the Lipschitz continuity of the global objective function, and (2) the Lipschitz continuity of local individual functions. Under the local regularity assumption, we provide the first optimal first-order decentralized algorithm called multi-step primal-dual (MSPD) and its corresponding optimal convergence rate. A notable aspect of this result is that, for non-smooth functions, while the dominant term of the error is in $O(1/\sqrt{t})$, the structure of the communication network only impacts a second-order term in $O(1/t)$, where $t$ is time. In other words, the error due to limits in communication resources decreases at a fast rate even in the case of non-strongly-convex objective functions. Under the global regularity assumption, we provide a simple yet efficient algorithm called distributed randomized smoothing (DRS) based on a local smoothing of the objective function, and show that DRS is within a $d^{1/4}$ multiplicative factor of the optimal convergence rate, where $d$ is the underlying dimension.

In this paper, we study the optimal convergence rate for distributed convex optimization problems in networks. We model the communication restrictions imposed by the network as a set of affine constraints and provide optimal complexity bounds for four different setups, namely: the function $F(\xb) \triangleq \sum_{i=1}^{m}f_i(\xb)$ is strongly convex and smooth, either strongly convex or smooth or just convex. Our results show that Nesterov's accelerated gradient descent on the dual problem can be executed in a distributed manner and obtains the same optimal rates as in the centralized version of the problem (up to constant or logarithmic factors) with an additional cost related to the spectral gap of the interaction matrix. Finally, we discuss some extensions to the proposed setup such as proximal friendly functions, time-varying graphs, improvement of the condition numbers.

北京阿比特科技有限公司