亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

We present a novel stabilized isogeometric formulation for the Stokes problem, where the geometry of interest is obtained via overlapping NURBS (non-uniform rational B-spline) patches, i.e., one patch on top of another in an arbitrary but predefined hierarchical order. All the visible regions constitute the computational domain, whereas independent patches are coupled through visible interfaces using Nitsche's formulation. Such a geometric representation inevitably involves trimming, which may yield trimmed elements of extremely small measures (referred to as bad elements) and thus lead to the instability issue. Motivated by the minimal stabilization method that rigorously guarantees stability for trimmed geometries [1], in this work we generalize it to the Stokes problem on overlapping patches. Central to our method is the distinct treatments for the pressure and velocity spaces: Stabilization for velocity is carried out for the flux terms on interfaces, whereas pressure is stabilized in all the bad elements. We provide a priori error estimates with a comprehensive theoretical study. Through a suite of numerical tests, we first show that optimal convergence rates are achieved, which consistently agrees with our theoretical findings. Second, we show that the accuracy of pressure is significantly improved by several orders using the proposed stabilization method, compared to the results without stabilization. Finally, we also demonstrate the flexibility and efficiency of the proposed method in capturing local features in the solution field.

相關內容

Karger (STOC 1995) gave the first FPTAS for the network (un)reliability problem, setting in motion research over the next three decades that obtained increasingly faster running times, eventually leading to a $\tilde{O}(n^2)$-time algorithm (Karger, STOC 2020). This represented a natural culmination of this line of work because the algorithmic techniques used can enumerate $\Theta(n^2)$ (near)-minimum cuts. In this paper, we go beyond this quadratic barrier and obtain a faster FPTAS for the network unreliability problem. Our algorithm runs in $m^{1+o(1)} + \tilde{O}(n^{1.5})$ time. Our main contribution is a new estimator for network unreliability in very reliable graphs. These graphs are usually the bottleneck for network unreliability since the disconnection event is elusive. Our estimator is obtained by defining an appropriate importance sampling subroutine on a dual spanning tree packing of the graph. To complement this estimator for very reliable graphs, we use recursive contraction for moderately reliable graphs. We show that an interleaving of sparsification and contraction can be used to obtain a better parametrization of the recursive contraction algorithm that yields a faster running time matching the one obtained for the very reliable case.

In a seminal paper, Kannan and Lov\'asz (1988) considered a quantity $\mu_{KL}(\Lambda,K)$ which denotes the best volume-based lower bound on the covering radius $\mu(\Lambda,K)$ of a convex body $K$ with respect to a lattice $\Lambda$. Kannan and Lov\'asz proved that $\mu(\Lambda,K) \leq n \cdot \mu_{KL}(\Lambda,K)$ and the Subspace Flatness Conjecture by Dadush (2012) claims a $O(\log(2n))$ factor suffices, which would match the lower bound from the work of Kannan and Lov\'asz. We settle this conjecture up to a constant in the exponent by proving that $\mu(\Lambda,K) \leq O(\log^{3}(2n)) \cdot \mu_{KL} (\Lambda,K)$. Our proof is based on the Reverse Minkowski Theorem due to Regev and Stephens-Davidowitz (2017). Following the work of Dadush (2012, 2019), we obtain a $(\log(2n))^{O(n)}$-time randomized algorithm to solve integer programs in $n$ variables. Another implication of our main result is a near-optimal flatness constant of $O(n \log^{3}(2n))$.

The traditional method of computing singular value decomposition (SVD) of a data matrix is based on a least squares principle, thus, is very sensitive to the presence of outliers. Hence the resulting inferences across different applications using the classical SVD are extremely degraded in the presence of data contamination (e.g., video surveillance background modelling tasks, etc.). A robust singular value decomposition method using the minimum density power divergence estimator (rSVDdpd) has been found to provide a satisfactory solution to this problem and works well in applications. For example, it provides a neat solution to the background modelling problem of video surveillance data in the presence of camera tampering. In this paper, we investigate the theoretical properties of the rSVDdpd estimator such as convergence, equivariance and consistency under reasonable assumptions. Since the dimension of the parameters, i.e., the number of singular values and the dimension of singular vectors can grow linearly with the size of the data, the usual M-estimation theory has to be suitably modified with concentration bounds to establish the asymptotic properties. We believe that we have been able to accomplish this satisfactorily in the present work. We also demonstrate the efficiency of rSVDdpd through extensive simulations.

In this study we consider domains that are composed of an infinite sequence of self-similar rings and corresponding finite element spaces over those domains. The rings are parameterized using piecewise polynomial or tensor-product B-spline mappings of degree $q$ over quadrilateral meshes. We then consider finite element discretizations which, over each ring, are mapped, piecewise polynomial functions of degree $p$. Such domains that are composed of self-similar rings may be created through a subdivision scheme or from a scaled boundary parameterization. We study approximation properties over such recursively parameterized domains. The main finding is that, for generic isoparametric discretizations (i.e., where $p=q$), the approximation properties always depend only on the degree of polynomials that can be reproduced exactly in the physical domain and not on the degree $p$ of the mapped elements. Especially, in general, $L^\infty$-errors converge at most with the rate $h^2$, where $h$ is the mesh size, independent of the degree $p=q$. This has implications for subdivision based isogeometric analysis, which we will discuss in this paper.

In this paper, we look at the pressure checkerboard problem that arises in an Eulerian meshless method that solves the incompressible Navier-Stokes equations using the generalized finite difference method (GFDM). Although, the checkerboard problem has been dealt with extensively in mesh-based methods, the literature in connection with meshless methods is comparatively scarce. In this paper, we explore the occurrence of the checkerboard problem in a meshless method. A few unsuccessful attempts to resolve the checkerboard problem are reported. The successful fix for the problem entails an algorithm that adapts the point cloud by adding points in the regions of pressure oscillations. The algorithm uses an error indicator that detects the presence of the checkerboard oscillations in the solution. The algorithm minimizes the computational effort since it ensures the use of additional points only in regions of concern, as directed by the error indicator, in contrast to an approach of using a highly refined set of points throughout the domain. It also requires no a priori estimates of the regions where the oscillations occur and integrates conveniently in the framework of the meshless method since no re-meshing strategies are involved. The results are compared with literature and a good match is observed.

This paper presents a convergence analysis of a Krylov subspace spectral (KSS) method applied to a 1-D wave equation in an inhomogeneous medium. It will be shown that for sufficiently regular initial data, this KSS method yields unconditional stability, spectral accuracy in space, and second-order accuracy in time, in the case of constant wave speed and a bandlimited reaction term coefficient. Numerical experiments that corroborate the established theory are included, along with an investigation of generalizations, such as to higher space dimensions and nonlinear PDEs, that features performance comparisons with other Krylov subspace-based time-stepping methods. This paper also includes the first stability analysis of a KSS method that does not assume a bandlimited reaction term coefficient.

We generalize signature Gr\"obner bases, previously studied in the free algebra over a field or polynomial rings over a ring, to ideals in the mixed algebra $R[x_1,...,x_k]\langle y_1,\dots,y_n \rangle$ where $R$ is a principal ideal domain. We give an algorithm for computing them, combining elements from the theory of commutative and noncommutative (signature) Gr\"obner bases, and prove its correctness. Applications include extensions of the free algebra with commutative variables, e.g., for homogenization purposes or for performing ideal theoretic operations such as intersections, and computations over $\mathbb{Z}$ as universal proofs over fields of arbitrary characteristic. By extending the signature cover criterion to our setting, our algorithm also lifts some technical restrictions from previous noncommutative signature-based algorithms, now allowing, e.g., elimination orderings. We provide a prototype implementation for the case when $R$ is a field, and show that our algorithm for the mixed algebra is more efficient than classical approaches using existing algorithms.

This contribution presents a model order reduction framework for real-time efficient solution of trimmed, multi-patch isogeometric Kirchhoff-Love shells. In several scenarios, such as design and shape optimization, multiple simulations need to be performed for a given set of physical or geometrical parameters. This step can be computationally expensive in particular for real world, practical applications. We are interested in geometrical parameters and take advantage of the flexibility of splines in representing complex geometries. In this case, the operators are geometry-dependent and generally depend on the parameters in a non-affine way. Moreover, the solutions obtained from trimmed domains may vary highly with respect to different values of the parameters. Therefore, we employ a local reduced basis method based on clustering techniques and the Discrete Empirical Interpolation Method to construct affine approximations and efficient reduced order models. In addition, we discuss the application of the reduction strategy to parametric shape optimization. Finally, we demonstrate the performance of the proposed framework to parameterized Kirchhoff-Love shells through benchmark tests on trimmed, multi-patch meshes including a complex geometry. The proposed approach is accurate and achieves a significant reduction of the online computational cost in comparison to the standard reduced basis method.

We describe ACE0, a lightweight platform for evaluating the suitability and viability of AI methods for behaviour discovery in multiagent simulations. Specifically, ACE0 was designed to explore AI methods for multi-agent simulations used in operations research studies related to new technologies such as autonomous aircraft. Simulation environments used in production are often high-fidelity, complex, require significant domain knowledge and as a result have high R&D costs. Minimal and lightweight simulation environments can help researchers and engineers evaluate the viability of new AI technologies for behaviour discovery in a more agile and potentially cost effective manner. In this paper we describe the motivation for the development of ACE0.We provide a technical overview of the system architecture, describe a case study of behaviour discovery in the aerospace domain, and provide a qualitative evaluation of the system. The evaluation includes a brief description of collaborative research projects with academic partners, exploring different AI behaviour discovery methods.

Substantial progress has been made recently on developing provably accurate and efficient algorithms for low-rank matrix factorization via nonconvex optimization. While conventional wisdom often takes a dim view of nonconvex optimization algorithms due to their susceptibility to spurious local minima, simple iterative methods such as gradient descent have been remarkably successful in practice. The theoretical footings, however, had been largely lacking until recently. In this tutorial-style overview, we highlight the important role of statistical models in enabling efficient nonconvex optimization with performance guarantees. We review two contrasting approaches: (1) two-stage algorithms, which consist of a tailored initialization step followed by successive refinement; and (2) global landscape analysis and initialization-free algorithms. Several canonical matrix factorization problems are discussed, including but not limited to matrix sensing, phase retrieval, matrix completion, blind deconvolution, robust principal component analysis, phase synchronization, and joint alignment. Special care is taken to illustrate the key technical insights underlying their analyses. This article serves as a testament that the integrated consideration of optimization and statistics leads to fruitful research findings.

北京阿比特科技有限公司