亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

In recent years, new regularization methods based on (deep) neural networks have shown very promising empirical performance for the numerical solution of ill-posed problems, such as in medical imaging and imaging science. Due to the nonlinearity of neural networks, these methods often lack satisfactory theoretical justification. In this work, we rigorously discuss the convergence of a successful unsupervised approach that utilizes untrained convolutional neural networks to represent solutions to linear ill-posed problems. Untrained neural networks have particular appeal for many applications because they do not require paired training data. The regularization property of the approach relies solely on the architecture of the neural network instead. Due to the vast over-parameterization of the employed neural network, suitable early stopping is essential for the success of the method. We establish that the classical discrepancy principle is an adequate method for early stopping of two-layer untrained convolutional neural networks learned by gradient descent, and furthermore, it yields an approximation with minimax optimal convergence rates. Numerical results are also presented to illustrate the theoretical findings.

相關內容

神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(Neural Networks)是(shi)世界上三個(ge)最古老的(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)建模學(xue)(xue)(xue)(xue)(xue)會的(de)(de)(de)檔案期刊:國(guo)際(ji)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)會(INNS)、歐(ou)洲神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)會(ENNS)和(he)(he)(he)日(ri)本神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)會(JNNS)。神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)提(ti)供了一(yi)(yi)個(ge)論壇,以發展(zhan)和(he)(he)(he)培育一(yi)(yi)個(ge)國(guo)際(ji)社會的(de)(de)(de)學(xue)(xue)(xue)(xue)(xue)者和(he)(he)(he)實(shi)踐者感(gan)(gan)興(xing)趣的(de)(de)(de)所有方面(mian)的(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)和(he)(he)(he)相關方法(fa)的(de)(de)(de)計(ji)算智能(neng)。神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)歡迎高質量論文的(de)(de)(de)提(ti)交,有助于全面(mian)的(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)研究(jiu),從行為和(he)(he)(he)大(da)腦(nao)建模,學(xue)(xue)(xue)(xue)(xue)習(xi)算法(fa),通過數(shu)學(xue)(xue)(xue)(xue)(xue)和(he)(he)(he)計(ji)算分析,系(xi)(xi)統的(de)(de)(de)工(gong)程和(he)(he)(he)技術(shu)應(ying)用(yong),大(da)量使用(yong)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)的(de)(de)(de)概(gai)念和(he)(he)(he)技術(shu)。這一(yi)(yi)獨特而廣泛的(de)(de)(de)范圍促(cu)(cu)進(jin)了生(sheng)物(wu)和(he)(he)(he)技術(shu)研究(jiu)之(zhi)間的(de)(de)(de)思想交流,并有助于促(cu)(cu)進(jin)對生(sheng)物(wu)啟發的(de)(de)(de)計(ji)算智能(neng)感(gan)(gan)興(xing)趣的(de)(de)(de)跨學(xue)(xue)(xue)(xue)(xue)科社區(qu)的(de)(de)(de)發展(zhan)。因此,神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)編委會代(dai)表(biao)的(de)(de)(de)專(zhuan)(zhuan)家(jia)領域包括心理(li)學(xue)(xue)(xue)(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)(jing)生(sheng)物(wu)學(xue)(xue)(xue)(xue)(xue),計(ji)算機科學(xue)(xue)(xue)(xue)(xue),工(gong)程,數(shu)學(xue)(xue)(xue)(xue)(xue),物(wu)理(li)。該雜志發表(biao)文章、信(xin)(xin)(xin)件和(he)(he)(he)評論以及給編輯的(de)(de)(de)信(xin)(xin)(xin)件、社論、時事、軟件調查和(he)(he)(he)專(zhuan)(zhuan)利信(xin)(xin)(xin)息。文章發表(biao)在五個(ge)部分之(zhi)一(yi)(yi):認知科學(xue)(xue)(xue)(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)(jing)科學(xue)(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)(xue)習(xi)系(xi)(xi)統,數(shu)學(xue)(xue)(xue)(xue)(xue)和(he)(he)(he)計(ji)算分析、工(gong)程和(he)(he)(he)應(ying)用(yong)。 官網(wang)地址(zhi):

In recent work, the authors have developed a generic methodology for calibrating the noise in fluid dynamics stochastic partial differential equations where the stochasticity was introduced to parametrize subgrid-scale processes. The stochastic parameterization of sub-grid scale processes is required in the estimation of uncertainty in weather and climate predictions, to represent systematic model errors arising from subgrid-scale fluctuations. The previous methodology used a principal component analysis (PCA) technique based on the ansatz that the increments of the stochastic parametrization are normally distributed. In this paper, the PCA technique is replaced by a generative model technique. This enables us to avoid imposing additional constraints on the increments. The methodology is tested on a stochastic rotating shallow water model with the elevation variable of the model used as input data. The numerical simulations show that the noise is indeed non-Gaussian. The generative modelling technology gives good RMSE, CRPS score and forecast rank histogram results.

Recently, graph neural networks (GNNs) have been widely used for document classification. However, most existing methods are based on static word co-occurrence graphs without sentence-level information, which poses three challenges:(1) word ambiguity, (2) word synonymity, and (3) dynamic contextual dependency. To address these challenges, we propose a novel GNN-based sparse structure learning model for inductive document classification. Specifically, a document-level graph is initially generated by a disjoint union of sentence-level word co-occurrence graphs. Our model collects a set of trainable edges connecting disjoint words between sentences and employs structure learning to sparsely select edges with dynamic contextual dependencies. Graphs with sparse structures can jointly exploit local and global contextual information in documents through GNNs. For inductive learning, the refined document graph is further fed into a general readout function for graph-level classification and optimization in an end-to-end manner. Extensive experiments on several real-world datasets demonstrate that the proposed model outperforms most state-of-the-art results, and reveal the necessity to learn sparse structures for each document.

We consider the problem of explaining the predictions of graph neural networks (GNNs), which otherwise are considered as black boxes. Existing methods invariably focus on explaining the importance of graph nodes or edges but ignore the substructures of graphs, which are more intuitive and human-intelligible. In this work, we propose a novel method, known as SubgraphX, to explain GNNs by identifying important subgraphs. Given a trained GNN model and an input graph, our SubgraphX explains its predictions by efficiently exploring different subgraphs with Monte Carlo tree search. To make the tree search more effective, we propose to use Shapley values as a measure of subgraph importance, which can also capture the interactions among different subgraphs. To expedite computations, we propose efficient approximation schemes to compute Shapley values for graph data. Our work represents the first attempt to explain GNNs via identifying subgraphs explicitly and directly. Experimental results show that our SubgraphX achieves significantly improved explanations, while keeping computations at a reasonable level.

Residual networks (ResNets) have displayed impressive results in pattern recognition and, recently, have garnered considerable theoretical interest due to a perceived link with neural ordinary differential equations (neural ODEs). This link relies on the convergence of network weights to a smooth function as the number of layers increases. We investigate the properties of weights trained by stochastic gradient descent and their scaling with network depth through detailed numerical experiments. We observe the existence of scaling regimes markedly different from those assumed in neural ODE literature. Depending on certain features of the network architecture, such as the smoothness of the activation function, one may obtain an alternative ODE limit, a stochastic differential equation or neither of these. These findings cast doubts on the validity of the neural ODE model as an adequate asymptotic description of deep ResNets and point to an alternative class of differential equations as a better description of the deep network limit.

Graph Neural Networks (GNNs), which generalize deep neural networks to graph-structured data, have drawn considerable attention and achieved state-of-the-art performance in numerous graph related tasks. However, existing GNN models mainly focus on designing graph convolution operations. The graph pooling (or downsampling) operations, that play an important role in learning hierarchical representations, are usually overlooked. In this paper, we propose a novel graph pooling operator, called Hierarchical Graph Pooling with Structure Learning (HGP-SL), which can be integrated into various graph neural network architectures. HGP-SL incorporates graph pooling and structure learning into a unified module to generate hierarchical representations of graphs. More specifically, the graph pooling operation adaptively selects a subset of nodes to form an induced subgraph for the subsequent layers. To preserve the integrity of graph's topological information, we further introduce a structure learning mechanism to learn a refined graph structure for the pooled graph at each layer. By combining HGP-SL operator with graph neural networks, we perform graph level representation learning with focus on graph classification task. Experimental results on six widely used benchmarks demonstrate the effectiveness of our proposed model.

Deep convolutional neural networks (CNNs) have recently achieved great success in many visual recognition tasks. However, existing deep neural network models are computationally expensive and memory intensive, hindering their deployment in devices with low memory resources or in applications with strict latency requirements. Therefore, a natural thought is to perform model compression and acceleration in deep networks without significantly decreasing the model performance. During the past few years, tremendous progress has been made in this area. In this paper, we survey the recent advanced techniques for compacting and accelerating CNNs model developed. These techniques are roughly categorized into four schemes: parameter pruning and sharing, low-rank factorization, transferred/compact convolutional filters, and knowledge distillation. Methods of parameter pruning and sharing will be described at the beginning, after that the other techniques will be introduced. For each scheme, we provide insightful analysis regarding the performance, related applications, advantages, and drawbacks etc. Then we will go through a few very recent additional successful methods, for example, dynamic capacity networks and stochastic depths networks. After that, we survey the evaluation matrix, the main datasets used for evaluating the model performance and recent benchmarking efforts. Finally, we conclude this paper, discuss remaining challenges and possible directions on this topic.

Attention Model has now become an important concept in neural networks that has been researched within diverse application domains. This survey provides a structured and comprehensive overview of the developments in modeling attention. In particular, we propose a taxonomy which groups existing techniques into coherent categories. We review the different neural architectures in which attention has been incorporated, and also show how attention improves interpretability of neural models. Finally, we discuss some applications in which modeling attention has a significant impact. We hope this survey will provide a succinct introduction to attention models and guide practitioners while developing approaches for their applications.

Recently, graph neural networks (GNNs) have revolutionized the field of graph representation learning through effectively learned node embeddings, and achieved state-of-the-art results in tasks such as node classification and link prediction. However, current GNN methods are inherently flat and do not learn hierarchical representations of graphs---a limitation that is especially problematic for the task of graph classification, where the goal is to predict the label associated with an entire graph. Here we propose DiffPool, a differentiable graph pooling module that can generate hierarchical representations of graphs and can be combined with various graph neural network architectures in an end-to-end fashion. DiffPool learns a differentiable soft cluster assignment for nodes at each layer of a deep GNN, mapping nodes to a set of clusters, which then form the coarsened input for the next GNN layer. Our experimental results show that combining existing GNN methods with DiffPool yields an average improvement of 5-10% accuracy on graph classification benchmarks, compared to all existing pooling approaches, achieving a new state-of-the-art on four out of five benchmark data sets.

This paper proposes a method to modify traditional convolutional neural networks (CNNs) into interpretable CNNs, in order to clarify knowledge representations in high conv-layers of CNNs. In an interpretable CNN, each filter in a high conv-layer represents a certain object part. We do not need any annotations of object parts or textures to supervise the learning process. Instead, the interpretable CNN automatically assigns each filter in a high conv-layer with an object part during the learning process. Our method can be applied to different types of CNNs with different structures. The clear knowledge representation in an interpretable CNN can help people understand the logics inside a CNN, i.e., based on which patterns the CNN makes the decision. Experiments showed that filters in an interpretable CNN were more semantically meaningful than those in traditional CNNs.

Deep neural networks (DNNs) have been found to be vulnerable to adversarial examples resulting from adding small-magnitude perturbations to inputs. Such adversarial examples can mislead DNNs to produce adversary-selected results. Different attack strategies have been proposed to generate adversarial examples, but how to produce them with high perceptual quality and more efficiently requires more research efforts. In this paper, we propose AdvGAN to generate adversarial examples with generative adversarial networks (GANs), which can learn and approximate the distribution of original instances. For AdvGAN, once the generator is trained, it can generate adversarial perturbations efficiently for any instance, so as to potentially accelerate adversarial training as defenses. We apply AdvGAN in both semi-whitebox and black-box attack settings. In semi-whitebox attacks, there is no need to access the original target model after the generator is trained, in contrast to traditional white-box attacks. In black-box attacks, we dynamically train a distilled model for the black-box model and optimize the generator accordingly. Adversarial examples generated by AdvGAN on different target models have high attack success rate under state-of-the-art defenses compared to other attacks. Our attack has placed the first with 92.76% accuracy on a public MNIST black-box attack challenge.

北京阿比特科技有限公司