亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

We derive conditions for the existence of fixed points of nonnegative neural networks, an important research objective to understand the behavior of neural networks in modern applications involving autoencoders and loop unrolling techniques, among others. In particular, we show that neural networks with nonnegative inputs and nonnegative parameters can be recognized as monotonic and (weakly) scalable functions within the framework of nonlinear Perron-Frobenius theory. This fact enables us to derive conditions for the existence of a nonempty fixed point set of the nonnegative neural networks, and these conditions are weaker than those obtained recently using arguments in convex analysis, which are typically based on the assumption of nonexpansivity of the activation functions. Furthermore, we prove that the shape of the fixed point set of monotonic and weakly scalable neural networks is often an interval, which degenerates to a point for the case of scalable networks. The chief results of this paper are verified in numerical simulations, where we consider an autoencoder-type network that first compresses angular power spectra in massive MIMO systems, and, second, reconstruct the input spectra from the compressed signals.

相關內容

神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)(Neural Networks)是世界上三個最古(gu)老的(de)(de)(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)建模(mo)學(xue)會的(de)(de)(de)(de)檔案期刊:國際神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)學(xue)會(INNS)、歐(ou)洲神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)學(xue)會(ENNS)和(he)(he)日本神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)學(xue)會(JNNS)。神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)提供(gong)了(le)一(yi)個論壇,以發(fa)(fa)展(zhan)和(he)(he)培(pei)育一(yi)個國際社會的(de)(de)(de)(de)學(xue)者和(he)(he)實(shi)踐者感興趣的(de)(de)(de)(de)所有方面(mian)的(de)(de)(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)和(he)(he)相關(guan)方法的(de)(de)(de)(de)計(ji)算(suan)智能(neng)。神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)歡(huan)迎高質量論文的(de)(de)(de)(de)提交,有助(zhu)于全面(mian)的(de)(de)(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)研究,從行(xing)為和(he)(he)大(da)腦建模(mo),學(xue)習算(suan)法,通過(guo)數學(xue)和(he)(he)計(ji)算(suan)分析,系統的(de)(de)(de)(de)工(gong)程(cheng)(cheng)和(he)(he)技(ji)術應用(yong),大(da)量使用(yong)神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)的(de)(de)(de)(de)概(gai)念和(he)(he)技(ji)術。這一(yi)獨特(te)而廣泛的(de)(de)(de)(de)范圍促進(jin)了(le)生物(wu)(wu)和(he)(he)技(ji)術研究之(zhi)間的(de)(de)(de)(de)思想交流,并有助(zhu)于促進(jin)對生物(wu)(wu)啟發(fa)(fa)的(de)(de)(de)(de)計(ji)算(suan)智能(neng)感興趣的(de)(de)(de)(de)跨(kua)學(xue)科(ke)社區(qu)的(de)(de)(de)(de)發(fa)(fa)展(zhan)。因此,神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)絡(luo)編(bian)委會代表(biao)的(de)(de)(de)(de)專家領(ling)域(yu)包括心理(li)學(xue),神(shen)(shen)(shen)經(jing)(jing)(jing)生物(wu)(wu)學(xue),計(ji)算(suan)機(ji)科(ke)學(xue),工(gong)程(cheng)(cheng),數學(xue),物(wu)(wu)理(li)。該雜志(zhi)發(fa)(fa)表(biao)文章(zhang)、信(xin)件(jian)和(he)(he)評(ping)論以及給編(bian)輯的(de)(de)(de)(de)信(xin)件(jian)、社論、時事、軟件(jian)調查和(he)(he)專利信(xin)息。文章(zhang)發(fa)(fa)表(biao)在五(wu)個部分之(zhi)一(yi):認知科(ke)學(xue),神(shen)(shen)(shen)經(jing)(jing)(jing)科(ke)學(xue),學(xue)習系統,數學(xue)和(he)(he)計(ji)算(suan)分析、工(gong)程(cheng)(cheng)和(he)(he)應用(yong)。 官網(wang)(wang)地址(zhi):

Graph Convolutional Networks (GCNs) are one of the most popular architectures that are used to solve classification problems accompanied by graphical information. We present a rigorous theoretical understanding of the effects of graph convolutions in multi-layer networks. We study these effects through the node classification problem of a non-linearly separable Gaussian mixture model coupled with a stochastic block model. First, we show that a single graph convolution expands the regime of the distance between the means where multi-layer networks can classify the data by a factor of at least $1/\sqrt[4]{\mathbb{E}{\rm deg}}$, where $\mathbb{E}{\rm deg}$ denotes the expected degree of a node. Second, we show that with a slightly stronger graph density, two graph convolutions improve this factor to at least $1/\sqrt[4]{n}$, where $n$ is the number of nodes in the graph. Finally, we provide both theoretical and empirical insights into the performance of graph convolutions placed in different combinations among the layers of a network, concluding that the performance is mutually similar for all combinations of the placement. We present extensive experiments on both synthetic and real-world data that illustrate our results.

In this paper we present a deep learning method to predict the temporal evolution of dissipative dynamic systems. We propose using both geometric and thermodynamic inductive biases to improve accuracy and generalization of the resulting integration scheme. The first is achieved with Graph Neural Networks, which induces a non-Euclidean geometrical prior with permutation invariant node and edge update functions. The second bias is forced by learning the GENERIC structure of the problem, an extension of the Hamiltonian formalism, to model more general non-conservative dynamics. Several examples are provided in both Eulerian and Lagrangian description in the context of fluid and solid mechanics respectively, achieving relative mean errors of less than 3% in all the tested examples. Two ablation studies are provided based on recent works in both physics-informed and geometric deep learning.

Traditional nonnegative matrix factorization (NMF) learns a new feature representation on the whole data space, which means treating all features equally. However, a subspace is often sufficient for accurate representation in practical applications, and redundant features can be invalid or even harmful. For example, if a camera has some sensors destroyed, then the corresponding pixels in the photos from this camera are not helpful to identify the content, which means only the subspace consisting of remaining pixels is worthy of attention. This paper proposes a new NMF method by introducing adaptive weights to identify key features in the original space so that only a subspace involves generating the new representation. Two strategies are proposed to achieve this: the fuzzier weighted technique and entropy regularized weighted technique, both of which result in an iterative solution with a simple form. Experimental results on several real-world datasets demonstrated that the proposed methods can generate a more accurate feature representation than existing methods. The code developed in this study is available at //github.com/WNMF1/FWNMF-ERWNMF.

The shift towards end-to-end deep learning has brought unprecedented advances in many areas of computer vision. However, deep neural networks are trained on images with resolutions that rarely exceed $1,000 \times 1,000$ pixels. The growing use of scanners that create images with extremely high resolutions (average can be $100,000 \times 100,000$ pixels) thereby presents novel challenges to the field. Most of the published methods preprocess high-resolution images into a set of smaller patches, imposing an a priori belief on the best properties of the extracted patches (magnification, field of view, location, etc.). Herein, we introduce Magnifying Networks (MagNets) as an alternative deep learning solution for gigapixel image analysis that does not rely on a preprocessing stage nor requires the processing of billions of pixels. MagNets can learn to dynamically retrieve any part of a gigapixel image, at any magnification level and field of view, in an end-to-end fashion with minimal ground truth (a single global, slide-level label). Our results on the publicly available Camelyon16 and Camelyon17 datasets corroborate to the effectiveness and efficiency of MagNets and the proposed optimization framework for whole slide image classification. Importantly, MagNets process far less patches from each slide than any of the existing approaches ($10$ to $300$ times less).

Existing inferential methods for small area data involve a trade-off between maintaining area-level frequentist coverage rates and improving inferential precision via the incorporation of indirect information. In this article, we propose a method to obtain an area-level prediction region for a future observation which mitigates this trade-off. The proposed method takes a conformal prediction approach in which the conformity measure is the posterior predictive density of a working model that incorporates indirect information. The resulting prediction region has guaranteed frequentist coverage regardless of the working model, and, if the working model assumptions are accurate, the region has minimum expected volume compared to other regions with the same coverage rate. When constructed under a normal working model, we prove such a prediction region is an interval and construct an efficient algorithm to obtain the exact interval. We illustrate the performance of our method through simulation studies and an application to EPA radon survey data.

The stochastic gradient Langevin Dynamics is one of the most fundamental algorithms to solve sampling problems and non-convex optimization appearing in several machine learning applications. Especially, its variance reduced versions have nowadays gained particular attention. In this paper, we study two variants of this kind, namely, the Stochastic Variance Reduced Gradient Langevin Dynamics and the Stochastic Recursive Gradient Langevin Dynamics. We prove their convergence to the objective distribution in terms of KL-divergence under the sole assumptions of smoothness and Log-Sobolev inequality which are weaker conditions than those used in prior works for these algorithms. With the batch size and the inner loop length set to $\sqrt{n}$, the gradient complexity to achieve an $\epsilon$-precision is $\tilde{O}((n+dn^{1/2}\epsilon^{-1})\gamma^2 L^2\alpha^{-2})$, which is an improvement from any previous analyses. We also show some essential applications of our result to non-convex optimization.

Designers reportedly struggle with design optimization tasks where they are asked to find a combination of design parameters that maximizes a given set of objectives. In HCI, design optimization problems are often exceedingly complex, involving multiple objectives and expensive empirical evaluations. Model-based computational design algorithms assist designers by generating design examples during design, however they assume a model of the interaction domain. Black box methods for assistance, on the other hand, can work with any design problem. However, virtually all empirical studies of this human-in-the-loop approach have been carried out by either researchers or end-users. The question stands out if such methods can help designers in realistic tasks. In this paper, we study Bayesian optimization as an algorithmic method to guide the design optimization process. It operates by proposing to a designer which design candidate to try next, given previous observations. We report observations from a comparative study with 40 novice designers who were tasked to optimize a complex 3D touch interaction technique. The optimizer helped designers explore larger proportions of the design space and arrive at a better solution, however they reported lower agency and expressiveness. Designers guided by an optimizer reported lower mental effort but also felt less creative and less in charge of the progress. We conclude that human-in-the-loop optimization can support novice designers in cases where agency is not critical.

Music Structure Analysis (MSA) consists in segmenting a music piece in several distinct sections. We approach MSA within a compression framework, under the hypothesis that the structure is more easily revealed by a simplified representation of the original content of the song. More specifically, under the hypothesis that MSA is correlated with similarities occurring at the bar scale, this article introduces the use of linear and non-linear compression schemes on barwise audio signals. Compressed representations capture the most salient components of the different bars in the song and are then used to infer the song structure using a dynamic programming algorithm. This work explores both low-rank approximation models such as Principal Component Analysis or Nonnegative Matrix Factorization and "piece-specific" Auto-Encoding Neural Networks, with the objective to learn latent representations specific to a given song. Such approaches do not rely on supervision nor annotations, which are well-known to be tedious to collect and possibly ambiguous in MSA description. In our experiments, several unsupervised compression schemes achieve a level of performance comparable to that of state-of-the-art supervised methods (for 3s tolerance) on the RWC-Pop dataset, showcasing the importance of the barwise compression processing for MSA.

Dynamic neural network is an emerging research topic in deep learning. Compared to static models which have fixed computational graphs and parameters at the inference stage, dynamic networks can adapt their structures or parameters to different inputs, leading to notable advantages in terms of accuracy, computational efficiency, adaptiveness, etc. In this survey, we comprehensively review this rapidly developing area by dividing dynamic networks into three main categories: 1) instance-wise dynamic models that process each instance with data-dependent architectures or parameters; 2) spatial-wise dynamic networks that conduct adaptive computation with respect to different spatial locations of image data and 3) temporal-wise dynamic models that perform adaptive inference along the temporal dimension for sequential data such as videos and texts. The important research problems of dynamic networks, e.g., architecture design, decision making scheme, optimization technique and applications, are reviewed systematically. Finally, we discuss the open problems in this field together with interesting future research directions.

When and why can a neural network be successfully trained? This article provides an overview of optimization algorithms and theory for training neural networks. First, we discuss the issue of gradient explosion/vanishing and the more general issue of undesirable spectrum, and then discuss practical solutions including careful initialization and normalization methods. Second, we review generic optimization methods used in training neural networks, such as SGD, adaptive gradient methods and distributed methods, and theoretical results for these algorithms. Third, we review existing research on the global issues of neural network training, including results on bad local minima, mode connectivity, lottery ticket hypothesis and infinite-width analysis.

北京阿比特科技有限公司