亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

We use fixed point theory to analyze nonnegative neural networks, which we define as neural networks that map nonnegative vectors to nonnegative vectors. We first show that nonnegative neural networks with nonnegative weights and biases can be recognized as monotonic and (weakly) scalable mappings within the framework of nonlinear Perron-Frobenius theory. This fact enables us to provide conditions for the existence of fixed points of nonnegative neural networks having inputs and outputs of the same dimension, and these conditions are weaker than those recently obtained using arguments in convex analysis. Furthermore, we prove that the shape of the fixed point set of nonnegative neural networks with nonnegative weights and biases is an interval, which under mild conditions degenerates to a point. These results are then used to obtain the existence of fixed points of more general nonnegative neural networks. From a practical perspective, our results contribute to the understanding of the behavior of autoencoders, and we also offer valuable mathematical machinery for future developments in deep equilibrium models.

相關內容

神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(Neural Networks)是世界(jie)上三個最古老(lao)的(de)(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)建模(mo)學(xue)(xue)會的(de)(de)(de)(de)(de)檔案期刊:國際(ji)神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)學(xue)(xue)會(INNS)、歐洲(zhou)神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)學(xue)(xue)會(ENNS)和(he)(he)(he)(he)(he)日本神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)學(xue)(xue)會(JNNS)。神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)提供了(le)一(yi)個論(lun)壇,以(yi)發(fa)(fa)(fa)展(zhan)和(he)(he)(he)(he)(he)培(pei)育一(yi)個國際(ji)社會的(de)(de)(de)(de)(de)學(xue)(xue)者和(he)(he)(he)(he)(he)實踐者感興(xing)(xing)趣的(de)(de)(de)(de)(de)所有方面的(de)(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)和(he)(he)(he)(he)(he)相關方法(fa)的(de)(de)(de)(de)(de)計(ji)算(suan)智能(neng)。神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)歡迎(ying)高(gao)質量論(lun)文的(de)(de)(de)(de)(de)提交(jiao),有助于(yu)全面的(de)(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)研究(jiu),從行為(wei)和(he)(he)(he)(he)(he)大腦建模(mo),學(xue)(xue)習(xi)算(suan)法(fa),通過數學(xue)(xue)和(he)(he)(he)(he)(he)計(ji)算(suan)分析(xi),系統(tong)的(de)(de)(de)(de)(de)工(gong)(gong)程和(he)(he)(he)(he)(he)技(ji)術(shu)應(ying)用(yong),大量使用(yong)神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)的(de)(de)(de)(de)(de)概念和(he)(he)(he)(he)(he)技(ji)術(shu)。這(zhe)一(yi)獨特而廣泛的(de)(de)(de)(de)(de)范圍促進(jin)了(le)生(sheng)物(wu)和(he)(he)(he)(he)(he)技(ji)術(shu)研究(jiu)之間的(de)(de)(de)(de)(de)思想交(jiao)流,并(bing)有助于(yu)促進(jin)對生(sheng)物(wu)啟發(fa)(fa)(fa)的(de)(de)(de)(de)(de)計(ji)算(suan)智能(neng)感興(xing)(xing)趣的(de)(de)(de)(de)(de)跨(kua)學(xue)(xue)科(ke)社區的(de)(de)(de)(de)(de)發(fa)(fa)(fa)展(zhan)。因(yin)此(ci),神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)編委會代(dai)表的(de)(de)(de)(de)(de)專家領域包括(kuo)心理(li)學(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)生(sheng)物(wu)學(xue)(xue),計(ji)算(suan)機科(ke)學(xue)(xue),工(gong)(gong)程,數學(xue)(xue),物(wu)理(li)。該雜(za)志發(fa)(fa)(fa)表文章(zhang)、信(xin)件(jian)和(he)(he)(he)(he)(he)評(ping)論(lun)以(yi)及給編輯(ji)的(de)(de)(de)(de)(de)信(xin)件(jian)、社論(lun)、時事、軟(ruan)件(jian)調查和(he)(he)(he)(he)(he)專利信(xin)息。文章(zhang)發(fa)(fa)(fa)表在五(wu)個部(bu)分之一(yi):認知科(ke)學(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)科(ke)學(xue)(xue),學(xue)(xue)習(xi)系統(tong),數學(xue)(xue)和(he)(he)(he)(he)(he)計(ji)算(suan)分析(xi)、工(gong)(gong)程和(he)(he)(he)(he)(he)應(ying)用(yong)。 官網(wang)(wang)地址:

We develop an inferential toolkit for analyzing object-valued responses, which correspond to data situated in general metric spaces, paired with Euclidean predictors within the conformal framework. To this end we introduce conditional profile average transport costs, where we compare distance profiles that correspond to one-dimensional distributions of probability mass falling into balls of increasing radius through the optimal transport cost when moving from one distance profile to another. The average transport cost to transport a given distance profile to all others is crucial for statistical inference in metric spaces and underpins the proposed conditional profile scores. A key feature of the proposed approach is to utilize the distribution of conditional profile average transport costs as conformity score for general metric space-valued responses, which facilitates the construction of prediction sets by the split conformal algorithm. We derive the uniform convergence rate of the proposed conformity score estimators and establish asymptotic conditional validity for the prediction sets. The finite sample performance for synthetic data in various metric spaces demonstrates that the proposed conditional profile score outperforms existing methods in terms of both coverage level and size of the resulting prediction sets, even in the special case of scalar and thus Euclidean responses. We also demonstrate the practical utility of conditional profile scores for network data from New York taxi trips and for compositional data reflecting energy sourcing of U.S. states.

Neural networks can be thought of as applying a transformation to an input dataset. The way in which they change the topology of such a dataset often holds practical significance for many tasks, particularly those demanding non-homeomorphic mappings for optimal solutions, such as classification problems. In this work, we leverage the fact that neural networks are equivalent to continuous piecewise-affine maps, whose rank can be used to pinpoint regions in the input space that undergo non-homeomorphic transformations, leading to alterations in the topological structure of the input dataset. Our approach enables us to make use of the relative homology sequence, with which one can study the homology groups of the quotient of a manifold $\mathcal{M}$ and a subset $A$, assuming some minimal properties on these spaces. As a proof of principle, we empirically investigate the presence of low-rank (topology-changing) affine maps as a function of network width and mean weight. We show that in randomly initialized narrow networks, there will be regions in which the (co)homology groups of a data manifold can change. As the width increases, the homology groups of the input manifold become more likely to be preserved. We end this part of our work by constructing highly non-random wide networks that do not have this property and relating this non-random regime to Dale's principle, which is a defining characteristic of biological neural networks. Finally, we study simple feedforward networks trained on MNIST, as well as on toy classification and regression tasks, and show that networks manipulate the topology of data differently depending on the continuity of the task they are trained on.

In backbone networks, it is fundamental to quickly protect traffic against any unexpected event, such as failures or congestions, which may impact Quality of Service (QoS). Standard solutions based on Segment Routing (SR), such as Topology-Independent Loop-Free Alternate (TI-LFA), are used in practice to handle failures, but no distributed solutions exist for distributed and tactical congestion mitigation. A promising approach leveraging SR has been recently proposed to quickly steer traffic away from congested links over alternative paths. As the pre-computation of alternative paths plays a paramount role to efficiently mitigating congestions, we investigate the associated path computation problem aiming at maximizing the amount of traffic that can be rerouted as well as the resilience against any 1-link failure. In particular, we focus on two variants of this problem. First, we maximize the residual flow after all possible failures. We show that the problem is NP-Hard, and we solve it via a Benders decomposition algorithm. Then, to provide a practical and scalable solution, we solve a relaxed variant problem, that maximizes, instead of flow, the number of surviving alternative paths after all possible failures. We provide a polynomial algorithm. Through numerical experiments, we compare the two variants and show that they allow to increase the amount of rerouted traffic and the resiliency of the network after any 1-link failure.

Traditional supervised learning aims to learn an unknown mapping by fitting a function to a set of input-output pairs with a fixed dimension. The fitted function is then defined on inputs of the same dimension. However, in many settings, the unknown mapping takes inputs in any dimension; examples include graph parameters defined on graphs of any size and physics quantities defined on an arbitrary number of particles. We leverage a newly-discovered phenomenon in algebraic topology, called representation stability, to define equivariant neural networks that can be trained with data in a fixed dimension and then extended to accept inputs in any dimension. Our approach is user-friendly, requiring only the network architecture and the groups for equivariance, and can be combined with any training procedure. We provide a simple open-source implementation of our methods and offer preliminary numerical experiments.

Physics-informed neural networks (PINN) is a extremely powerful paradigm used to solve equations encountered in scientific computing applications. An important part of the procedure is the minimization of the equation residual which includes, when the equation is time-dependent, a time sampling. It was argued in the literature that the sampling need not be uniform but should overweight initial time instants, but no rigorous explanation was provided for these choice. In this paper we take some prototypical examples and, under standard hypothesis concerning the neural network convergence, we show that the optimal time sampling follows a truncated exponential distribution. In particular we explain when the time sampling is best to be uniform and when it should not be. The findings are illustrated with numerical examples on linear equation, Burgers' equation and the Lorenz system.

Domain decomposition provides an effective way to tackle the dilemma of physics-informed neural networks (PINN) which struggle to accurately and efficiently solve partial differential equations (PDEs) in the whole domain, but the lack of efficient tools for dealing with the interfaces between two adjacent sub-domains heavily hinders the training effects, even leads to the discontinuity of the learned solutions. In this paper, we propose a symmetry group based domain decomposition strategy to enhance the PINN for solving the forward and inverse problems of the PDEs possessing a Lie symmetry group. Specifically, for the forward problem, we first deploy the symmetry group to generate the dividing-lines having known solution information which can be adjusted flexibly and are used to divide the whole training domain into a finite number of non-overlapping sub-domains, then utilize the PINN and the symmetry-enhanced PINN methods to learn the solutions in each sub-domain and finally stitch them to the overall solution of PDEs. For the inverse problem, we first utilize the symmetry group acting on the data of the initial and boundary conditions to generate labeled data in the interior domain of PDEs and then find the undetermined parameters as well as the solution by only training the neural networks in a sub-domain. Consequently, the proposed method can predict high-accuracy solutions of PDEs which are failed by the vanilla PINN in the whole domain and the extended physics-informed neural network in the same sub-domains. Numerical results of the Korteweg-de Vries equation with a translation symmetry and the nonlinear viscous fluid equation with a scaling symmetry show that the accuracies of the learned solutions are improved largely.

Convolutional neural networks (CNNs) trained with cross-entropy loss have proven to be extremely successful in classifying images. In recent years, much work has been done to also improve the theoretical understanding of neural networks. Nevertheless, it seems limited when these networks are trained with cross-entropy loss, mainly because of the unboundedness of the target function. In this paper, we aim to fill this gap by analyzing the rate of the excess risk of a CNN classifier trained by cross-entropy loss. Under suitable assumptions on the smoothness and structure of the a posteriori probability, it is shown that these classifiers achieve a rate of convergence which is independent of the dimension of the image. These rates are in line with the practical observations about CNNs.

A neural architecture with randomly initialized weights, in the infinite width limit, is equivalent to a Gaussian Random Field whose covariance function is the so-called Neural Network Gaussian Process kernel (NNGP). We prove that a reproducing kernel Hilbert space (RKHS) defined by the NNGP contains only functions that can be approximated by the architecture. To achieve a certain approximation error the required number of neurons in each layer is defined by the RKHS norm of the target function. Moreover, the approximation can be constructed from a supervised dataset by a random multi-layer representation of an input vector, together with training of the last layer's weights. For a 2-layer NN and a domain equal to an $n-1$-dimensional sphere in ${\mathbb R}^n$, we compare the number of neurons required by Barron's theorem and by the multi-layer features construction. We show that if eigenvalues of the integral operator of the NNGP decay slower than $k^{-n-\frac{2}{3}}$ where $k$ is an order of an eigenvalue, then our theorem guarantees a more succinct neural network approximation than Barron's theorem. We also make some computational experiments to verify our theoretical findings. Our experiments show that realistic neural networks easily learn target functions even when both theorems do not give any guarantees.

We hypothesize that due to the greedy nature of learning in multi-modal deep neural networks, these models tend to rely on just one modality while under-fitting the other modalities. Such behavior is counter-intuitive and hurts the models' generalization, as we observe empirically. To estimate the model's dependence on each modality, we compute the gain on the accuracy when the model has access to it in addition to another modality. We refer to this gain as the conditional utilization rate. In the experiments, we consistently observe an imbalance in conditional utilization rates between modalities, across multiple tasks and architectures. Since conditional utilization rate cannot be computed efficiently during training, we introduce a proxy for it based on the pace at which the model learns from each modality, which we refer to as the conditional learning speed. We propose an algorithm to balance the conditional learning speeds between modalities during training and demonstrate that it indeed addresses the issue of greedy learning. The proposed algorithm improves the model's generalization on three datasets: Colored MNIST, Princeton ModelNet40, and NVIDIA Dynamic Hand Gesture.

Recent advances in 3D fully convolutional networks (FCN) have made it feasible to produce dense voxel-wise predictions of volumetric images. In this work, we show that a multi-class 3D FCN trained on manually labeled CT scans of several anatomical structures (ranging from the large organs to thin vessels) can achieve competitive segmentation results, while avoiding the need for handcrafting features or training class-specific models. To this end, we propose a two-stage, coarse-to-fine approach that will first use a 3D FCN to roughly define a candidate region, which will then be used as input to a second 3D FCN. This reduces the number of voxels the second FCN has to classify to ~10% and allows it to focus on more detailed segmentation of the organs and vessels. We utilize training and validation sets consisting of 331 clinical CT images and test our models on a completely unseen data collection acquired at a different hospital that includes 150 CT scans, targeting three anatomical organs (liver, spleen, and pancreas). In challenging organs such as the pancreas, our cascaded approach improves the mean Dice score from 68.5 to 82.2%, achieving the highest reported average score on this dataset. We compare with a 2D FCN method on a separate dataset of 240 CT scans with 18 classes and achieve a significantly higher performance in small organs and vessels. Furthermore, we explore fine-tuning our models to different datasets. Our experiments illustrate the promise and robustness of current 3D FCN based semantic segmentation of medical images, achieving state-of-the-art results. Our code and trained models are available for download: //github.com/holgerroth/3Dunet_abdomen_cascade.

北京阿比特科技有限公司