亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

In this paper we develop a neural network for the numerical simulation of time-dependent linear transport equations with diffusive scaling and uncertainties. The goal of the network is to resolve the computational challenges of curse-of-dimensionality and multiple scales of the problem. We first show that a standard Physics-Informed Neural Network (PINNs) fails to capture the multiscale nature of the problem, hence justifies the need to use Asymptotic-Preserving Neural Networks (APNNs). We show that not all classical AP formulations are fit for the neural network approach. We construct a micro-macro decomposition based neutral network, and also build in a mass conservation mechanism into the loss function, in order to capture the dynamic and multiscale nature of the solutions. Numerical examples are used to demonstrate the effectiveness of this APNNs.

相關內容

神(shen)經(jing)網(wang)(wang)絡(luo)(Neural Networks)是世界上三(san)個最古老的(de)(de)(de)(de)(de)神(shen)經(jing)建模(mo)學(xue)(xue)(xue)(xue)會(hui)的(de)(de)(de)(de)(de)檔案(an)期刊:國(guo)際神(shen)經(jing)網(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)會(hui)(INNS)、歐洲(zhou)神(shen)經(jing)網(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)會(hui)(ENNS)和日本神(shen)經(jing)網(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)會(hui)(JNNS)。神(shen)經(jing)網(wang)(wang)絡(luo)提供了一(yi)(yi)(yi)個論壇,以(yi)(yi)發展和培(pei)育一(yi)(yi)(yi)個國(guo)際社會(hui)的(de)(de)(de)(de)(de)學(xue)(xue)(xue)(xue)者和實踐者感興趣的(de)(de)(de)(de)(de)所有方(fang)面(mian)的(de)(de)(de)(de)(de)神(shen)經(jing)網(wang)(wang)絡(luo)和相關方(fang)法的(de)(de)(de)(de)(de)計(ji)(ji)算智能。神(shen)經(jing)網(wang)(wang)絡(luo)歡迎高質(zhi)量(liang)論文的(de)(de)(de)(de)(de)提交(jiao),有助(zhu)于(yu)(yu)全面(mian)的(de)(de)(de)(de)(de)神(shen)經(jing)網(wang)(wang)絡(luo)研究,從行為(wei)和大腦建模(mo),學(xue)(xue)(xue)(xue)習(xi)(xi)算法,通過數(shu)學(xue)(xue)(xue)(xue)和計(ji)(ji)算分析,系(xi)統的(de)(de)(de)(de)(de)工程和技(ji)術(shu)應(ying)(ying)用,大量(liang)使用神(shen)經(jing)網(wang)(wang)絡(luo)的(de)(de)(de)(de)(de)概念和技(ji)術(shu)。這一(yi)(yi)(yi)獨特而廣泛的(de)(de)(de)(de)(de)范圍促進(jin)了生(sheng)物(wu)和技(ji)術(shu)研究之(zhi)間的(de)(de)(de)(de)(de)思想交(jiao)流,并有助(zhu)于(yu)(yu)促進(jin)對生(sheng)物(wu)啟發的(de)(de)(de)(de)(de)計(ji)(ji)算智能感興趣的(de)(de)(de)(de)(de)跨學(xue)(xue)(xue)(xue)科(ke)(ke)社區的(de)(de)(de)(de)(de)發展。因此,神(shen)經(jing)網(wang)(wang)絡(luo)編(bian)委(wei)會(hui)代表(biao)的(de)(de)(de)(de)(de)專家領域包括(kuo)心理學(xue)(xue)(xue)(xue),神(shen)經(jing)生(sheng)物(wu)學(xue)(xue)(xue)(xue),計(ji)(ji)算機(ji)科(ke)(ke)學(xue)(xue)(xue)(xue),工程,數(shu)學(xue)(xue)(xue)(xue),物(wu)理。該(gai)雜志(zhi)發表(biao)文章(zhang)、信(xin)件和評(ping)論以(yi)(yi)及給編(bian)輯的(de)(de)(de)(de)(de)信(xin)件、社論、時事、軟件調查和專利信(xin)息。文章(zhang)發表(biao)在五(wu)個部分之(zhi)一(yi)(yi)(yi):認知科(ke)(ke)學(xue)(xue)(xue)(xue),神(shen)經(jing)科(ke)(ke)學(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)習(xi)(xi)系(xi)統,數(shu)學(xue)(xue)(xue)(xue)和計(ji)(ji)算分析、工程和應(ying)(ying)用。 官網(wang)(wang)地址:

The application of machine learning (ML) techniques, especially neural networks, has seen tremendous success at processing images and language. This is because we often lack formal models to understand visual and audio input, so here neural networks can unfold their abilities as they can model solely from data. In the field of physics we typically have models that describe natural processes reasonably well on a formal level. Nonetheless, in recent years, ML has also proven useful in these realms, be it by speeding up numerical simulations or by improving accuracy. One important and so far unsolved problem in classical physics is understanding turbulent fluid motion. In this work we construct a strongly simplified representation of turbulence by using the Gledzer-Ohkitani-Yamada (GOY) shell model. With this system we intend to investigate the potential of ML-supported and physics-constrained small-scale turbulence modelling. Instead of standard supervised learning we propose an approach that aims to reconstruct statistical properties of turbulence such as the self-similar inertial-range scaling, where we could achieve encouraging experimental results. Furthermore we discuss pitfalls when combining machine learning with differential equations.

High-resolution spectroscopic surveys of the Milky Way have entered the Big Data regime and have opened avenues for solving outstanding questions in Galactic archaeology. However, exploiting their full potential is limited by complex systematics, whose characterization has not received much attention in modern spectroscopic analyses. In this work, we present a novel method to disentangle the component of spectral data space intrinsic to the stars from that due to systematics. Using functional principal component analysis on a sample of $18,933$ giant spectra from APOGEE, we find that the intrinsic structure above the level of observational uncertainties requires ${\approx}$10 functional principal components (FPCs). Our FPCs can reduce the dimensionality of spectra, remove systematics, and impute masked wavelengths, thereby enabling accurate studies of stellar populations. To demonstrate the applicability of our FPCs, we use them to infer stellar parameters and abundances of 28 giants in the open cluster M67. We employ Sequential Neural Likelihood, a simulation-based Bayesian inference method that learns likelihood functions using neural density estimators, to incorporate non-Gaussian effects in spectral likelihoods. By hierarchically combining the inferred abundances, we limit the spread of the following elements in M67: $\mathrm{Fe} \lesssim 0.02$ dex; $\mathrm{C} \lesssim 0.03$ dex; $\mathrm{O}, \mathrm{Mg}, \mathrm{Si}, \mathrm{Ni} \lesssim 0.04$ dex; $\mathrm{Ca} \lesssim 0.05$ dex; $\mathrm{N}, \mathrm{Al} \lesssim 0.07$ dex (at 68% confidence). Our constraints suggest a lack of self-pollution by core-collapse supernovae in M67, which has promising implications for the future of chemical tagging to understand the star formation history and dynamical evolution of the Milky Way.

Controlling antenna tilts in cellular networks is imperative to reach an efficient trade-off between network coverage and capacity. In this paper, we devise algorithms learning optimal tilt control policies from existing data (in the so-called passive learning setting) or from data actively generated by the algorithms (the active learning setting). We formalize the design of such algorithms as a Best Policy Identification (BPI) problem in Contextual Linear Multi-Arm Bandits (CL-MAB). An arm represents an antenna tilt update; the context captures current network conditions; the reward corresponds to an improvement of performance, mixing coverage and capacity; and the objective is to identify, with a given level of confidence, an approximately optimal policy (a function mapping the context to an arm with maximal reward). For CL-MAB in both active and passive learning settings, we derive information-theoretical lower bounds on the number of samples required by any algorithm returning an approximately optimal policy with a given level of certainty, and devise algorithms achieving these fundamental limits. We apply our algorithms to the Remote Electrical Tilt (RET) optimization problem in cellular networks, and show that they can produce optimal tilt update policy using much fewer data samples than naive or existing rule-based learning algorithms.

In this paper, a nonlinear system of fractional ordinary differential equations with multiple scales in time is investigated. We are interested in the effective long-term computation of the solution. The main challenge is how to obtain the solution of the coupled problem at a lower computational cost. We analysize a multiscale method for the nonlinear system where the fast system has a periodic applied force and the slow equation contains fractional derivatives as a simplication of the atherosclerosis with a plaque growth. A local periodic equation is derived to approximate the original system and the error estimates are given. Then a finite difference method is designed to approximate the original and the approximate problems. We construct four examples, including three with exact solutions and one following the original problem setting, to test the accuracy and computational efficiency of the proposed method. It is observed that, the computational time is very much reduced and the multiscale method performs very well in comparison to fully resolved simulation for the case of small time scale separation. The larger the time scale separation is, the more effective the multiscale method is.

Memory replay may be key to learning in biological brains, which manage to learn new tasks continually without catastrophically interfering with previous knowledge. On the other hand, artificial neural networks suffer from catastrophic forgetting and tend to only perform well on tasks that they were recently trained on. In this work we explore the application of latent space based memory replay for classification using artificial neural networks. We are able to preserve good performance in previous tasks by storing only a small percentage of the original data in a compressed latent space version.

Due to the fundamental limit to reducing power consumption of running deep learning models on von-Neumann architecture, research on neuromorphic computing systems based on low-power spiking neural networks using analog neurons is in the spotlight. In order to integrate a large number of neurons, neurons need to be designed to occupy a small area, but as technology scales down, analog neurons are difficult to scale, and they suffer from reduced voltage headroom/dynamic range and circuit nonlinearities. In light of this, this paper first models the nonlinear behavior of existing current-mirror-based voltage-domain neurons designed in a 28nm process, and show SNN inference accuracy can be severely degraded by the effect of neuron's nonlinearity. Then, to mitigate this problem, we propose a novel neuron, which processes incoming spikes in the time domain and greatly improves the linearity, thereby improving the inference accuracy compared to the existing voltage-domain neuron. Tested on the MNIST dataset, the inference error rate of the proposed neuron differs by less than 0.1% from that of the ideal neuron.

Turbulence simulation with classical numerical solvers requires very high-resolution grids to accurately resolve dynamics. Here we train learned simulators at low spatial and temporal resolutions to capture turbulent dynamics generated at high resolution. We show that our proposed model can simulate turbulent dynamics more accurately than classical numerical solvers at the same low resolutions across various scientifically relevant metrics. Our model is trained end-to-end from data and is capable of learning a range of challenging chaotic and turbulent dynamics at low resolution, including trajectories generated by the state-of-the-art Athena++ engine. We show that our simpler, general-purpose architecture outperforms various more specialized, turbulence-specific architectures from the learned turbulence simulation literature. In general, we see that learned simulators yield unstable trajectories; however, we show that tuning training noise and temporal downsampling solves this problem. We also find that while generalization beyond the training distribution is a challenge for learned models, training noise, convolutional architectures, and added loss constraints can help. Broadly, we conclude that our learned simulator outperforms traditional solvers run on coarser grids, and emphasize that simple design choices can offer stability and robust generalization.

The radiative transfer equation is a fundamental equation in transport theory and applications, which is a 5-dimensional PDE in the stationary one-velocity case, leading to great difficulties in numerical simulation. To tackle this bottleneck, we first use the discrete ordinate technique to discretize the scattering term, an integral with respect to the angular variables, resulting in a semi-discrete hyperbolic system. Then, we make the spatial discretization by means of the discontinuous Galerkin (DG) method combined with the sparse grid method. The final linear system is solved by the block Gauss-Seidal iteration method. The computational complexity and error analysis are developed in detail, which show the new method is more efficient than the original discrete ordinate DG method. A series of numerical results are performed to validate the convergence behavior and effectiveness of the proposed method.

Alternating Direction Method of Multipliers (ADMM) is a widely used tool for machine learning in distributed settings, where a machine learning model is trained over distributed data sources through an interactive process of local computation and message passing. Such an iterative process could cause privacy concerns of data owners. The goal of this paper is to provide differential privacy for ADMM-based distributed machine learning. Prior approaches on differentially private ADMM exhibit low utility under high privacy guarantee and often assume the objective functions of the learning problems to be smooth and strongly convex. To address these concerns, we propose a novel differentially private ADMM-based distributed learning algorithm called DP-ADMM, which combines an approximate augmented Lagrangian function with time-varying Gaussian noise addition in the iterative process to achieve higher utility for general objective functions under the same differential privacy guarantee. We also apply the moments accountant method to bound the end-to-end privacy loss. The theoretical analysis shows that DP-ADMM can be applied to a wider class of distributed learning problems, is provably convergent, and offers an explicit utility-privacy tradeoff. To our knowledge, this is the first paper to provide explicit convergence and utility properties for differentially private ADMM-based distributed learning algorithms. The evaluation results demonstrate that our approach can achieve good convergence and model accuracy under high end-to-end differential privacy guarantee.

We introduce a new family of deep neural network models. Instead of specifying a discrete sequence of hidden layers, we parameterize the derivative of the hidden state using a neural network. The output of the network is computed using a black-box differential equation solver. These continuous-depth models have constant memory cost, adapt their evaluation strategy to each input, and can explicitly trade numerical precision for speed. We demonstrate these properties in continuous-depth residual networks and continuous-time latent variable models. We also construct continuous normalizing flows, a generative model that can train by maximum likelihood, without partitioning or ordering the data dimensions. For training, we show how to scalably backpropagate through any ODE solver, without access to its internal operations. This allows end-to-end training of ODEs within larger models.

北京阿比特科技有限公司