亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

We derive unconditionally stable and convergent variable-step BDF2 scheme for solving the MBE model with slope selection. The discrete orthogonal convolution kernels of the variable-step BDF2 method is commonly utilized recently for solving the phase field models. In this paper, we further prove some new inequalities, concerning the vector forms, for the kernels especially dealing with the nonlinear terms in the slope selection model. The convergence rate of the fully discrete scheme is proved to be two both in time and space in $L^2$ norm under the setting of the variable time steps. Energy dissipation law is proved rigorously with a modified energy by adding a small term to the discrete version of the original free energy functional. Two numerical examples including an adaptive time-stepping strategy are given to verify the convergence rate and the energy dissipation law.

相關內容

In this paper we propose a new approach to realizability interpretations for nonstandard arithmetic. We deal with nonstandard analysis in the context of (semi)intuitionistic realizability, focusing on the Lightstone-Robinson construction of a model for nonstandard analysis through an ultrapower. In particular, we consider an extension of the $\lambda$-calculus with a memory cell, that contains an integer (the state), in order to indicate in which slice of the ultrapower $\cal{M}^{\mathbb{N}}$ the computation is being done. We pay attention to the nonstandard principles (and their computational content) obtainable in this setting. In particular, we give non-trivial realizers to Idealization and a non-standard version of the LLPO principle. We then discuss how to quotient this product to mimic the Lightstone-Robinson construction.

The purpose of this paper is to analyze a mixed method for linear elasticity eigenvalue problem, which approximates numerically the stress, displacement, and rotation, by piecewise $(k+1)$, $k$ and $(k+1)$-th degree polynomial functions ($k\geq 1$), respectively. The numerical eigenfunction of stress is symmetric. By the discrete $H^1$-stability of numerical displacement, we prove an $O(h^{k+2})$ approximation to the $L^{2}$-orthogonal projection of the eigenspace of exact displacement for the eigenvalue problem, with proper regularity assumption. Thus via postprocessing, we obtain a better approximation to the eigenspace of exact displacement for the eigenproblem than conventional methods. We also prove that numerical approximation to the eigenfunction of stress is locking free with respect to Poisson ratio. We introduce a hybridization to reduce the mixed method to a condensed eigenproblem and prove an $O(h^2)$ initial approximation (independent of the inverse of the elasticity operator) of the eigenvalue for the nonlinear eigenproblem by using the discrete $H^1$-stability of numerical displacement, while only an $O(h)$ approximation can be obtained if we use the traditional inf-sup condition. Finally, we report some numerical experiments.

In many research fields in artificial intelligence, it has been shown that deep neural networks are useful to estimate unknown functions on high dimensional input spaces. However, their generalization performance is not yet completely clarified from the theoretical point of view because they are nonidentifiable and singular learning machines. Moreover, a ReLU function is not differentiable, to which algebraic or analytic methods in singular learning theory cannot be applied. In this paper, we study a deep ReLU neural network in overparametrized cases and prove that the Bayesian free energy, which is equal to the minus log marginal likelihoodor the Bayesian stochastic complexity, is bounded even if the number of layers are larger than necessary to estimate an unknown data-generating function. Since the Bayesian generalization error is equal to the increase of the free energy as a function of a sample size, our result also shows that the Bayesian generalization error does not increase even if a deep ReLU neural network is designed to be sufficiently large or in an opeverparametrized state.

Continuum kinetic theories provide an important tool for the analysis and simulation of particle suspensions. When those particles are anisotropic, the addition of a particle orientation vector to the kinetic description yields a $2d-1$ dimensional theory which becomes intractable to simulate, especially in three dimensions or near states where the particles are highly aligned. Coarse-grained theories that track only moments of the particle distribution functions provide a more efficient simulation framework, but require closure assumptions. For the particular case where the particles are apolar, the Bingham closure has been found to agree well with the underlying kinetic theory; yet the closure is non-trivial to compute, requiring the solution of an often nearly-singular nonlinear equation at every spatial discretization point at every timestep. In this paper, we present a robust, accurate, and efficient numerical scheme for evaluating the Bingham closure, with a controllable error/efficiency tradeoff. To demonstrate the utility of the method, we carry out high-resolution simulations of a coarse-grained continuum model for a suspension of active particles in parameter regimes inaccessible to kinetic theories. Analysis of these simulations reveals that inaccurately computing the closure can act to effectively limit spatial resolution in the coarse-grained fields. Pushing these simulations to the high spatial resolutions enabled by our method reveals a coupling between vorticity and topological defects in the suspension director field, as well as signatures of energy transfer between scales in this active fluid model.

Large neural networks can improve the accuracy and generalization on tasks across many domains. However, this trend cannot continue indefinitely due to limited hardware memory. As a result, researchers have devised a number of memory optimization methods (MOMs) to alleviate the memory bottleneck, such as gradient checkpointing, quantization, and swapping. In this work, we study memory optimization methods and show that, although these strategies indeed lower peak memory usage, they can actually decrease training throughput by up to 9.3x. To provide practical guidelines for practitioners, we propose a simple but effective performance model PAPAYA to quantitatively explain the memory and training time trade-off. PAPAYA can be used to determine when to apply the various memory optimization methods in training different models. We outline the circumstances in which memory optimization techniques are more advantageous based on derived implications from PAPAYA. We assess the accuracy of PAPAYA and the derived implications on a variety of machine models, showing that it achieves over 0.97 R score on predicting the peak memory/throughput, and accurately predicts the effectiveness of MOMs across five evaluated models on vision and NLP tasks.

A novel methodology is proposed for clustering multivariate time series data using energy distance defined in Sz\'ekely and Rizzo (2013). Specifically, a dissimilarity matrix is formed using the energy distance statistic to measure separation between the finite dimensional distributions for the component time series. Once the pairwise dissimilarity matrix is calculated, a hierarchical clustering method is then applied to obtain the dendrogram. This procedure is completely nonparametric as the dissimilarities between stationary distributions are directly calculated without making any model assumptions. In order to justify this procedure, asymptotic properties of the energy distance estimates are derived for general stationary and ergodic time series. The method is illustrated in a simulation study for various component time series that are either linear or nonlinear. Finally the methodology is applied to two examples; one involves GDP of selected countries and the other is population size of various states in the U.S.A. in the years 1900 -1999.

Recently, neural networks have been widely applied for solving partial differential equations (PDEs). Although such methods have been proven remarkably successful on practical engineering problems, they have not been shown, theoretically or empirically, to converge to the underlying PDE solution with arbitrarily high accuracy. The primary difficulty lies in solving the highly non-convex optimization problems resulting from the neural network discretization, which are difficult to treat both theoretically and practically. It is our goal in this work to take a step toward remedying this. For this purpose, we develop a novel greedy training algorithm for shallow neural networks. Our method is applicable to both the variational formulation of the PDE and also to the residual minimization formulation pioneered by physics informed neural networks (PINNs). We analyze the method and obtain a priori error bounds when solving PDEs from the function class defined by shallow networks, which rigorously establishes the convergence of the method as the network size increases. Finally, we test the algorithm on several benchmark examples, including high dimensional PDEs, to confirm the theoretical convergence rate. Although the method is expensive relative to traditional approaches such as finite element methods, we view this work as a proof of concept for neural network-based methods, which shows that numerical methods based upon neural networks can be shown to rigorously converge.

We show that the representation cost of fully connected neural networks with homogeneous nonlinearities - which describes the implicit bias in function space of networks with $L_2$-regularization or with losses such as the cross-entropy - converges as the depth of the network goes to infinity to a notion of rank over nonlinear functions. We then inquire under which conditions the global minima of the loss recover the `true' rank of the data: we show that for too large depths the global minimum will be approximately rank 1 (underestimating the rank); we then argue that there is a range of depths which grows with the number of datapoints where the true rank is recovered. Finally, we discuss the effect of the rank of a classifier on the topology of the resulting class boundaries and show that autoencoders with optimal nonlinear rank are naturally denoising.

The memory hierarchy has a high impact on the performance and power consumption in the system. Moreover, current embedded systems, included in mobile devices, are specifically designed to run multimedia applications, which are memory intensive. This increases the pressure on the memory subsystem and affects the performance and energy consumption. In this regard, the thermal problems, performance degradation and high energy consumption, can cause irreversible damage to the devices. We address the optimization of the whole memory subsystem with three approaches integrated as a single methodology. Firstly, the thermal impact of register file is analyzed and optimized. Secondly, the cache memory is addressed by optimizing cache configuration according to running applications and improving both performance and power consumption. Finally, we simplify the design and evaluation process of general-purpose and customized dynamic memory manager, in the main memory. To this aim, we apply different evolutionary algorithms in combination with memory simulators and profiling tools. This way, we are able to evaluate the quality of each candidate solution and take advantage of the exploration of solutions given by the optimization algorithm.We also provide an experimental experience where our proposal is assessed using well-known benchmark applications.

Graph Neural Networks (GNNs) have been studied from the lens of expressive power and generalization. However, their optimization properties are less well understood. We take the first step towards analyzing GNN training by studying the gradient dynamics of GNNs. First, we analyze linearized GNNs and prove that despite the non-convexity of training, convergence to a global minimum at a linear rate is guaranteed under mild assumptions that we validate on real-world graphs. Second, we study what may affect the GNNs' training speed. Our results show that the training of GNNs is implicitly accelerated by skip connections, more depth, and/or a good label distribution. Empirical results confirm that our theoretical results for linearized GNNs align with the training behavior of nonlinear GNNs. Our results provide the first theoretical support for the success of GNNs with skip connections in terms of optimization, and suggest that deep GNNs with skip connections would be promising in practice.

北京阿比特科技有限公司