亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Continuous-depth neural networks, such as Neural ODEs, have refashioned the understanding of residual neural networks in terms of non-linear vector-valued optimal control problems. The common solution is to use the adjoint sensitivity method to replicate a forward-backward pass optimisation problem. We propose a new approach which explicates the network's `depth' as a fundamental variable, thus reducing the problem to a system of forward-facing initial value problems. This new method is based on the principle of `Invariant Imbedding' for which we prove a general solution, applicable to all non-linear, vector-valued optimal control problems with both running and terminal loss. Our new architectures provide a tangible tool for inspecting the theoretical--and to a great extent unexplained--properties of network depth. They also constitute a resource of discrete implementations of Neural ODEs comparable to classes of imbedded residual neural networks. Through a series of experiments, we show the competitive performance of the proposed architectures for supervised learning and time series prediction.

相關內容

神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(Neural Networks)是(shi)世界上(shang)三個最古(gu)老(lao)的(de)(de)(de)(de)(de)神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)建模(mo)學(xue)(xue)會(hui)的(de)(de)(de)(de)(de)檔案期刊:國際神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)學(xue)(xue)會(hui)(INNS)、歐(ou)洲神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)學(xue)(xue)會(hui)(ENNS)和(he)(he)(he)(he)(he)日本神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)學(xue)(xue)會(hui)(JNNS)。神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)提供了(le)一個論(lun)(lun)(lun)(lun)壇,以發(fa)展和(he)(he)(he)(he)(he)培育(yu)一個國際社(she)(she)會(hui)的(de)(de)(de)(de)(de)學(xue)(xue)者(zhe)和(he)(he)(he)(he)(he)實踐者(zhe)感(gan)興(xing)趣的(de)(de)(de)(de)(de)所(suo)有(you)方面(mian)(mian)的(de)(de)(de)(de)(de)神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)和(he)(he)(he)(he)(he)相關方法的(de)(de)(de)(de)(de)計(ji)(ji)算(suan)(suan)智(zhi)能。神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)歡迎高質(zhi)量(liang)論(lun)(lun)(lun)(lun)文(wen)的(de)(de)(de)(de)(de)提交(jiao),有(you)助于(yu)全面(mian)(mian)的(de)(de)(de)(de)(de)神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)研究(jiu),從(cong)行為和(he)(he)(he)(he)(he)大腦建模(mo),學(xue)(xue)習(xi)(xi)算(suan)(suan)法,通過(guo)數學(xue)(xue)和(he)(he)(he)(he)(he)計(ji)(ji)算(suan)(suan)分(fen)析,系統的(de)(de)(de)(de)(de)工(gong)程(cheng)和(he)(he)(he)(he)(he)技術(shu)應(ying)用,大量(liang)使用神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)的(de)(de)(de)(de)(de)概念和(he)(he)(he)(he)(he)技術(shu)。這(zhe)一獨特而廣泛的(de)(de)(de)(de)(de)范圍促進(jin)了(le)生(sheng)物和(he)(he)(he)(he)(he)技術(shu)研究(jiu)之(zhi)(zhi)間的(de)(de)(de)(de)(de)思想交(jiao)流,并有(you)助于(yu)促進(jin)對生(sheng)物啟(qi)發(fa)的(de)(de)(de)(de)(de)計(ji)(ji)算(suan)(suan)智(zhi)能感(gan)興(xing)趣的(de)(de)(de)(de)(de)跨學(xue)(xue)科(ke)社(she)(she)區的(de)(de)(de)(de)(de)發(fa)展。因此(ci),神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)編委會(hui)代表的(de)(de)(de)(de)(de)專家領域包括(kuo)心理(li)學(xue)(xue),神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)生(sheng)物學(xue)(xue),計(ji)(ji)算(suan)(suan)機科(ke)學(xue)(xue),工(gong)程(cheng),數學(xue)(xue),物理(li)。該雜志發(fa)表文(wen)章(zhang)、信件(jian)和(he)(he)(he)(he)(he)評論(lun)(lun)(lun)(lun)以及給編輯的(de)(de)(de)(de)(de)信件(jian)、社(she)(she)論(lun)(lun)(lun)(lun)、時事、軟件(jian)調查和(he)(he)(he)(he)(he)專利(li)信息(xi)。文(wen)章(zhang)發(fa)表在(zai)五個部(bu)分(fen)之(zhi)(zhi)一:認知(zhi)科(ke)學(xue)(xue),神(shen)經(jing)(jing)(jing)(jing)(jing)(jing)科(ke)學(xue)(xue),學(xue)(xue)習(xi)(xi)系統,數學(xue)(xue)和(he)(he)(he)(he)(he)計(ji)(ji)算(suan)(suan)分(fen)析、工(gong)程(cheng)和(he)(he)(he)(he)(he)應(ying)用。 官網(wang)(wang)地(di)址:

Graph Convolutional Networks (GCNs) are one of the most popular architectures that are used to solve classification problems accompanied by graphical information. We present a rigorous theoretical understanding of the effects of graph convolutions in multi-layer networks. We study these effects through the node classification problem of a non-linearly separable Gaussian mixture model coupled with a stochastic block model. First, we show that a single graph convolution expands the regime of the distance between the means where multi-layer networks can classify the data by a factor of at least $1/\sqrt[4]{\mathbb{E}{\rm deg}}$, where $\mathbb{E}{\rm deg}$ denotes the expected degree of a node. Second, we show that with a slightly stronger graph density, two graph convolutions improve this factor to at least $1/\sqrt[4]{n}$, where $n$ is the number of nodes in the graph. Finally, we provide both theoretical and empirical insights into the performance of graph convolutions placed in different combinations among the layers of a network, concluding that the performance is mutually similar for all combinations of the placement. We present extensive experiments on both synthetic and real-world data that illustrate our results.

Feature propagation in Deep Neural Networks (DNNs) can be associated to nonlinear discrete dynamical systems. The novelty, in this paper, lies in letting the discretization parameter (time step-size) vary from layer to layer, which needs to be learned, in an optimization framework. The proposed framework can be applied to any of the existing networks such as ResNet, DenseNet or Fractional-DNN. This framework is shown to help overcome the vanishing and exploding gradient issues. Stability of some of the existing continuous DNNs such as Fractional-DNN is also studied. The proposed approach is applied to an ill-posed 3D-Maxwell's equation.

The adaptive processing of structured data is a long-standing research topic in machine learning that investigates how to automatically learn a mapping from a structured input to outputs of various nature. Recently, there has been an increasing interest in the adaptive processing of graphs, which led to the development of different neural network-based methodologies. In this thesis, we take a different route and develop a Bayesian Deep Learning framework for graph learning. The dissertation begins with a review of the principles over which most of the methods in the field are built, followed by a study on graph classification reproducibility issues. We then proceed to bridge the basic ideas of deep learning for graphs with the Bayesian world, by building our deep architectures in an incremental fashion. This framework allows us to consider graphs with discrete and continuous edge features, producing unsupervised embeddings rich enough to reach the state of the art on several classification tasks. Our approach is also amenable to a Bayesian nonparametric extension that automatizes the choice of almost all model's hyper-parameters. Two real-world applications demonstrate the efficacy of deep learning for graphs. The first concerns the prediction of information-theoretic quantities for molecular simulations with supervised neural models. After that, we exploit our Bayesian models to solve a malware-classification task while being robust to intra-procedural code obfuscation techniques. We conclude the dissertation with an attempt to blend the best of the neural and Bayesian worlds together. The resulting hybrid model is able to predict multimodal distributions conditioned on input graphs, with the consequent ability to model stochasticity and uncertainty better than most works. Overall, we aim to provide a Bayesian perspective into the articulated research field of deep learning for graphs.

Residual networks (ResNets) have displayed impressive results in pattern recognition and, recently, have garnered considerable theoretical interest due to a perceived link with neural ordinary differential equations (neural ODEs). This link relies on the convergence of network weights to a smooth function as the number of layers increases. We investigate the properties of weights trained by stochastic gradient descent and their scaling with network depth through detailed numerical experiments. We observe the existence of scaling regimes markedly different from those assumed in neural ODE literature. Depending on certain features of the network architecture, such as the smoothness of the activation function, one may obtain an alternative ODE limit, a stochastic differential equation or neither of these. These findings cast doubts on the validity of the neural ODE model as an adequate asymptotic description of deep ResNets and point to an alternative class of differential equations as a better description of the deep network limit.

Dynamic neural network is an emerging research topic in deep learning. Compared to static models which have fixed computational graphs and parameters at the inference stage, dynamic networks can adapt their structures or parameters to different inputs, leading to notable advantages in terms of accuracy, computational efficiency, adaptiveness, etc. In this survey, we comprehensively review this rapidly developing area by dividing dynamic networks into three main categories: 1) instance-wise dynamic models that process each instance with data-dependent architectures or parameters; 2) spatial-wise dynamic networks that conduct adaptive computation with respect to different spatial locations of image data and 3) temporal-wise dynamic models that perform adaptive inference along the temporal dimension for sequential data such as videos and texts. The important research problems of dynamic networks, e.g., architecture design, decision making scheme, optimization technique and applications, are reviewed systematically. Finally, we discuss the open problems in this field together with interesting future research directions.

Deep learning methods for graphs achieve remarkable performance on many node-level and graph-level prediction tasks. However, despite the proliferation of the methods and their success, prevailing Graph Neural Networks (GNNs) neglect subgraphs, rendering subgraph prediction tasks challenging to tackle in many impactful applications. Further, subgraph prediction tasks present several unique challenges, because subgraphs can have non-trivial internal topology, but also carry a notion of position and external connectivity information relative to the underlying graph in which they exist. Here, we introduce SUB-GNN, a subgraph neural network to learn disentangled subgraph representations. In particular, we propose a novel subgraph routing mechanism that propagates neural messages between the subgraph's components and randomly sampled anchor patches from the underlying graph, yielding highly accurate subgraph representations. SUB-GNN specifies three channels, each designed to capture a distinct aspect of subgraph structure, and we provide empirical evidence that the channels encode their intended properties. We design a series of new synthetic and real-world subgraph datasets. Empirical results for subgraph classification on eight datasets show that SUB-GNN achieves considerable performance gains, outperforming strong baseline methods, including node-level and graph-level GNNs, by 12.4% over the strongest baseline. SUB-GNN performs exceptionally well on challenging biomedical datasets when subgraphs have complex topology and even comprise multiple disconnected components.

Modern neural network training relies heavily on data augmentation for improved generalization. After the initial success of label-preserving augmentations, there has been a recent surge of interest in label-perturbing approaches, which combine features and labels across training samples to smooth the learned decision surface. In this paper, we propose a new augmentation method that leverages the first and second moments extracted and re-injected by feature normalization. We replace the moments of the learned features of one training image by those of another, and also interpolate the target labels. As our approach is fast, operates entirely in feature space, and mixes different signals than prior methods, one can effectively combine it with existing augmentation methods. We demonstrate its efficacy across benchmark data sets in computer vision, speech, and natural language processing, where it consistently improves the generalization performance of highly competitive baseline networks.

Graph neural networks (GNNs) are a popular class of machine learning models whose major advantage is their ability to incorporate a sparse and discrete dependency structure between data points. Unfortunately, GNNs can only be used when such a graph-structure is available. In practice, however, real-world graphs are often noisy and incomplete or might not be available at all. With this work, we propose to jointly learn the graph structure and the parameters of graph convolutional networks (GCNs) by approximately solving a bilevel program that learns a discrete probability distribution on the edges of the graph. This allows one to apply GCNs not only in scenarios where the given graph is incomplete or corrupted but also in those where a graph is not available. We conduct a series of experiments that analyze the behavior of the proposed method and demonstrate that it outperforms related methods by a significant margin.

Graphs, which describe pairwise relations between objects, are essential representations of many real-world data such as social networks. In recent years, graph neural networks, which extend the neural network models to graph data, have attracted increasing attention. Graph neural networks have been applied to advance many different graph related tasks such as reasoning dynamics of the physical system, graph classification, and node classification. Most of the existing graph neural network models have been designed for static graphs, while many real-world graphs are inherently dynamic. For example, social networks are naturally evolving as new users joining and new relations being created. Current graph neural network models cannot utilize the dynamic information in dynamic graphs. However, the dynamic information has been proven to enhance the performance of many graph analytical tasks such as community detection and link prediction. Hence, it is necessary to design dedicated graph neural networks for dynamic graphs. In this paper, we propose DGNN, a new {\bf D}ynamic {\bf G}raph {\bf N}eural {\bf N}etwork model, which can model the dynamic information as the graph evolving. In particular, the proposed framework can keep updating node information by capturing the sequential information of edges, the time intervals between edges and information propagation coherently. Experimental results on various dynamic graphs demonstrate the effectiveness of the proposed framework.

We propose a Bayesian convolutional neural network built upon Bayes by Backprop and elaborate how this known method can serve as the fundamental construct of our novel, reliable variational inference method for convolutional neural networks. First, we show how Bayes by Backprop can be applied to convolutional layers where weights in filters have probability distributions instead of point-estimates; and second, how our proposed framework leads with various network architectures to performances comparable to convolutional neural networks with point-estimates weights. In the past, Bayes by Backprop has been successfully utilised in feedforward and recurrent neural networks, but not in convolutional ones. This work symbolises the extension of the group of Bayesian neural networks which encompasses all three aforementioned types of network architectures now.

北京阿比特科技有限公司