亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

In recent years, there has been an intense debate about how learning in biological neural networks (BNNs) differs from learning in artificial neural networks. It is often argued that the updating of connections in the brain relies only on local information, and therefore a stochastic gradient-descent type optimization method cannot be used. In this paper, we study a stochastic model for supervised learning in BNNs. We show that a (continuous) gradient step occurs approximately when each learning opportunity is processed by many local updates. This result suggests that stochastic gradient descent may indeed play a role in optimizing BNNs.

相關內容

神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)(Neural Networks)是世界上三(san)個最(zui)古老的(de)(de)(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)建(jian)模(mo)學(xue)(xue)(xue)(xue)會的(de)(de)(de)(de)(de)檔案(an)期刊:國際(ji)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)學(xue)(xue)(xue)(xue)會(INNS)、歐(ou)洲(zhou)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)學(xue)(xue)(xue)(xue)會(ENNS)和(he)(he)(he)日本(ben)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)學(xue)(xue)(xue)(xue)會(JNNS)。神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)提供(gong)了一(yi)個論(lun)壇,以發(fa)(fa)展(zhan)和(he)(he)(he)培育一(yi)個國際(ji)社(she)會的(de)(de)(de)(de)(de)學(xue)(xue)(xue)(xue)者(zhe)和(he)(he)(he)實踐者(zhe)感興(xing)趣的(de)(de)(de)(de)(de)所有(you)(you)(you)方面(mian)(mian)的(de)(de)(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)和(he)(he)(he)相關方法(fa)的(de)(de)(de)(de)(de)計(ji)算(suan)(suan)(suan)智能。神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)歡迎高質量論(lun)文(wen)的(de)(de)(de)(de)(de)提交(jiao)(jiao),有(you)(you)(you)助于全面(mian)(mian)的(de)(de)(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)研究,從行為(wei)和(he)(he)(he)大腦建(jian)模(mo),學(xue)(xue)(xue)(xue)習(xi)算(suan)(suan)(suan)法(fa),通(tong)過(guo)數(shu)(shu)學(xue)(xue)(xue)(xue)和(he)(he)(he)計(ji)算(suan)(suan)(suan)分析(xi)(xi),系統的(de)(de)(de)(de)(de)工(gong)程(cheng)和(he)(he)(he)技(ji)(ji)術應(ying)用,大量使(shi)用神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)的(de)(de)(de)(de)(de)概(gai)念和(he)(he)(he)技(ji)(ji)術。這一(yi)獨(du)特而廣泛的(de)(de)(de)(de)(de)范圍(wei)促(cu)進了生物(wu)和(he)(he)(he)技(ji)(ji)術研究之(zhi)間的(de)(de)(de)(de)(de)思想交(jiao)(jiao)流,并有(you)(you)(you)助于促(cu)進對(dui)生物(wu)啟發(fa)(fa)的(de)(de)(de)(de)(de)計(ji)算(suan)(suan)(suan)智能感興(xing)趣的(de)(de)(de)(de)(de)跨(kua)學(xue)(xue)(xue)(xue)科(ke)社(she)區的(de)(de)(de)(de)(de)發(fa)(fa)展(zhan)。因(yin)此,神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)絡(luo)編(bian)(bian)委(wei)會代(dai)表(biao)的(de)(de)(de)(de)(de)專家領域包括心理學(xue)(xue)(xue)(xue),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)生物(wu)學(xue)(xue)(xue)(xue),計(ji)算(suan)(suan)(suan)機(ji)科(ke)學(xue)(xue)(xue)(xue),工(gong)程(cheng),數(shu)(shu)學(xue)(xue)(xue)(xue),物(wu)理。該(gai)雜志發(fa)(fa)表(biao)文(wen)章、信(xin)件(jian)和(he)(he)(he)評論(lun)以及給編(bian)(bian)輯的(de)(de)(de)(de)(de)信(xin)件(jian)、社(she)論(lun)、時(shi)事、軟件(jian)調查和(he)(he)(he)專利信(xin)息。文(wen)章發(fa)(fa)表(biao)在五(wu)個部分之(zhi)一(yi):認知科(ke)學(xue)(xue)(xue)(xue),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)科(ke)學(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)習(xi)系統,數(shu)(shu)學(xue)(xue)(xue)(xue)和(he)(he)(he)計(ji)算(suan)(suan)(suan)分析(xi)(xi)、工(gong)程(cheng)和(he)(he)(he)應(ying)用。 官網(wang)地址:

Sign language recognition (SLR) has recently achieved a breakthrough in performance thanks to deep neural networks trained on large annotated sign datasets. Of the many different sign languages, these annotated datasets are only available for a select few. Since acquiring gloss-level labels on sign language videos is difficult, learning by transferring knowledge from existing annotated sources is useful for recognition in under-resourced sign languages. This study provides a publicly available cross-dataset transfer learning benchmark from two existing public Turkish SLR datasets. We use a temporal graph convolution-based sign language recognition approach to evaluate five supervised transfer learning approaches and experiment with closed-set and partial-set cross-dataset transfer learning. Experiments demonstrate that improvement over finetuning based transfer learning is possible with specialized supervised transfer learning methods.

Bayesian approaches for training deep neural networks (BNNs) have received significant interest and have been effectively utilized in a wide range of applications. There have been several studies on the properties of posterior concentrations of BNNs. However, most of these studies only demonstrate results in BNN models with sparse or heavy-tailed priors. Surprisingly, no theoretical results currently exist for BNNs using Gaussian priors, which are the most commonly used one. The lack of theory arises from the absence of approximation results of Deep Neural Networks (DNNs) that are non-sparse and have bounded parameters. In this paper, we present a new approximation theory for non-sparse DNNs with bounded parameters. Additionally, based on the approximation theory, we show that BNNs with non-sparse general priors can achieve near-minimax optimal posterior concentration rates to the true model.

The Adam optimizer, often used in Machine Learning for neural network training, corresponds to an underlying ordinary differential equation (ODE) in the limit of very small learning rates. This work shows that the classical Adam algorithm is a first order implicit-explicit (IMEX) Euler discretization of the underlying ODE. Employing the time discretization point of view, we propose new extensions of the Adam scheme obtained by using higher order IMEX methods to solve the ODE. Based on this approach, we derive a new optimization algorithm for neural network training that performs better than classical Adam on several regression and classification problems.

Conformal inference is a fundamental and versatile tool that provides distribution-free guarantees for many machine learning tasks. We consider the transductive setting, where decisions are made on a test sample of $m$ new points, giving rise to $m$ conformal $p$-values. While classical results only concern their marginal distribution, we show that their joint distribution follows a P\'olya urn model, and establish a concentration inequality for their empirical distribution function. The results hold for arbitrary exchangeable scores, including adaptive ones that can use the covariates of the test+calibration samples at training stage for increased accuracy. We demonstrate the usefulness of these theoretical results through uniform, in-probability guarantees for two machine learning tasks of current interest: interval prediction for transductive transfer learning and novelty detection based on two-class classification.

The use of Artificial Intelligence (AI) based on data-driven algorithms has become ubiquitous in today's society. Yet, in many cases and especially when stakes are high, humans still make final decisions. The critical question, therefore, is whether AI helps humans make better decisions as compared to a human alone or AI an alone. We introduce a new methodological framework that can be used to answer experimentally this question with no additional assumptions. We measure a decision maker's ability to make correct decisions using standard classification metrics based on the baseline potential outcome. We consider a single-blinded experimental design, in which the provision of AI-generated recommendations is randomized across cases with a human making final decisions. Under this experimental design, we show how to compare the performance of three alternative decision-making systems--human-alone, human-with-AI, and AI-alone. We apply the proposed methodology to the data from our own randomized controlled trial of a pretrial risk assessment instrument. We find that AI recommendations do not improve the classification accuracy of a judge's decision to impose cash bail. Our analysis also shows that AI-alone decisions generally perform worse than human decisions with or without AI assistance. Finally, AI recommendations tend to impose cash bail on non-white arrestees more often than necessary when compared to white arrestees.

Deep neural network based recommendation systems have achieved great success as information filtering techniques in recent years. However, since model training from scratch requires sufficient data, deep learning-based recommendation methods still face the bottlenecks of insufficient data and computational inefficiency. Meta-learning, as an emerging paradigm that learns to improve the learning efficiency and generalization ability of algorithms, has shown its strength in tackling the data sparsity issue. Recently, a growing number of studies on deep meta-learning based recommenddation systems have emerged for improving the performance under recommendation scenarios where available data is limited, e.g. user cold-start and item cold-start. Therefore, this survey provides a timely and comprehensive overview of current deep meta-learning based recommendation methods. Specifically, we propose a taxonomy to discuss existing methods according to recommendation scenarios, meta-learning techniques, and meta-knowledge representations, which could provide the design space for meta-learning based recommendation methods. For each recommendation scenario, we further discuss technical details about how existing methods apply meta-learning to improve the generalization ability of recommendation models. Finally, we also point out several limitations in current research and highlight some promising directions for future research in this area.

Graph neural networks (GNNs) have been proven to be effective in various network-related tasks. Most existing GNNs usually exploit the low-frequency signals of node features, which gives rise to one fundamental question: is the low-frequency information all we need in the real world applications? In this paper, we first present an experimental investigation assessing the roles of low-frequency and high-frequency signals, where the results clearly show that exploring low-frequency signal only is distant from learning an effective node representation in different scenarios. How can we adaptively learn more information beyond low-frequency information in GNNs? A well-informed answer can help GNNs enhance the adaptability. We tackle this challenge and propose a novel Frequency Adaptation Graph Convolutional Networks (FAGCN) with a self-gating mechanism, which can adaptively integrate different signals in the process of message passing. For a deeper understanding, we theoretically analyze the roles of low-frequency signals and high-frequency signals on learning node representations, which further explains why FAGCN can perform well on different types of networks. Extensive experiments on six real-world networks validate that FAGCN not only alleviates the over-smoothing problem, but also has advantages over the state-of-the-arts.

Recently, deep multiagent reinforcement learning (MARL) has become a highly active research area as many real-world problems can be inherently viewed as multiagent systems. A particularly interesting and widely applicable class of problems is the partially observable cooperative multiagent setting, in which a team of agents learns to coordinate their behaviors conditioning on their private observations and commonly shared global reward signals. One natural solution is to resort to the centralized training and decentralized execution paradigm. During centralized training, one key challenge is the multiagent credit assignment: how to allocate the global rewards for individual agent policies for better coordination towards maximizing system-level's benefits. In this paper, we propose a new method called Q-value Path Decomposition (QPD) to decompose the system's global Q-values into individual agents' Q-values. Unlike previous works which restrict the representation relation of the individual Q-values and the global one, we leverage the integrated gradient attribution technique into deep MARL to directly decompose global Q-values along trajectory paths to assign credits for agents. We evaluate QPD on the challenging StarCraft II micromanagement tasks and show that QPD achieves the state-of-the-art performance in both homogeneous and heterogeneous multiagent scenarios compared with existing cooperative MARL algorithms.

With the rise of knowledge graph (KG), question answering over knowledge base (KBQA) has attracted increasing attention in recent years. Despite much research has been conducted on this topic, it is still challenging to apply KBQA technology in industry because business knowledge and real-world questions can be rather complicated. In this paper, we present AliMe-KBQA, a bold attempt to apply KBQA in the E-commerce customer service field. To handle real knowledge and questions, we extend the classic "subject-predicate-object (SPO)" structure with property hierarchy, key-value structure and compound value type (CVT), and enhance traditional KBQA with constraints recognition and reasoning ability. We launch AliMe-KBQA in the Marketing Promotion scenario for merchants during the "Double 11" period in 2018 and other such promotional events afterwards. Online results suggest that AliMe-KBQA is not only able to gain better resolution and improve customer satisfaction, but also becomes the preferred knowledge management method by business knowledge staffs since it offers a more convenient and efficient management experience.

Machine Learning has been the quintessential solution for many AI problems, but learning is still heavily dependent on the specific training data. Some learning models can be incorporated with a prior knowledge in the Bayesian set up, but these learning models do not have the ability to access any organised world knowledge on demand. In this work, we propose to enhance learning models with world knowledge in the form of Knowledge Graph (KG) fact triples for Natural Language Processing (NLP) tasks. Our aim is to develop a deep learning model that can extract relevant prior support facts from knowledge graphs depending on the task using attention mechanism. We introduce a convolution-based model for learning representations of knowledge graph entity and relation clusters in order to reduce the attention space. We show that the proposed method is highly scalable to the amount of prior information that has to be processed and can be applied to any generic NLP task. Using this method we show significant improvement in performance for text classification with News20, DBPedia datasets and natural language inference with Stanford Natural Language Inference (SNLI) dataset. We also demonstrate that a deep learning model can be trained well with substantially less amount of labeled training data, when it has access to organised world knowledge in the form of knowledge graph.

北京阿比特科技有限公司