亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Hardware failures are a growing challenge for machine learning accelerators, many of which are based on systolic arrays. When a permanent hardware failure occurs in a systolic array, existing solutions include localizing and isolating the faulty processing element (PE), using a redundant PE for re-execution, or in some extreme cases decommissioning the entire accelerator for further investigation. In this paper, we propose novel algorithmic approaches that mitigate permanent hardware faults in neural network (NN) accelerators by uniquely integrating the behavior of the faulty component instead of bypassing it. In doing so, we aim for a more sustainable use of the accelerator where faulty hardware is neither bypassed nor discarded, instead being given a second life. We first introduce a CUDA-accelerated systolic array simulator in PyTorch, which enabled us to quantify the impact of permanent faults appearing on links connecting two PEs or in weight registers, where one bit is stuck at 0 or 1 in the float32, float16, or bfloat16 representation. We then propose several algorithmic mitigation techniques for a subset of stuck-at faults, such as Invertible Scaling or Shifting of activations and weights, or fine tuning with the faulty behavior. Notably, the proposed techniques do not require any hardware modification, instead relying on existing components of widely used systolic array based accelerators, such as normalization, activation, and storage units. Extensive experimental evaluations using fully connected and convolutional NNs trained on MNIST, CIFAR-10 and ImageNet show that the proposed fault-tolerant approach matches or gets very close to the original fault-free accuracy.

相關內容

神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡(Neural Networks)是(shi)世界上三(san)個(ge)最(zui)古老的(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)建(jian)模(mo)學(xue)(xue)(xue)會(hui)(hui)的(de)(de)(de)(de)檔案期刊:國(guo)際(ji)神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡學(xue)(xue)(xue)會(hui)(hui)(INNS)、歐洲神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡學(xue)(xue)(xue)會(hui)(hui)(ENNS)和(he)日本神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡學(xue)(xue)(xue)會(hui)(hui)(JNNS)。神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡提供了一(yi)個(ge)論壇,以(yi)發(fa)展(zhan)和(he)培育一(yi)個(ge)國(guo)際(ji)社會(hui)(hui)的(de)(de)(de)(de)學(xue)(xue)(xue)者(zhe)和(he)實踐者(zhe)感(gan)興趣的(de)(de)(de)(de)所有方面的(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡和(he)相(xiang)關方法的(de)(de)(de)(de)計算(suan)(suan)(suan)(suan)智(zhi)能(neng)。神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡歡迎高(gao)質量論文(wen)的(de)(de)(de)(de)提交,有助于全(quan)面的(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡研究,從行為和(he)大腦建(jian)模(mo),學(xue)(xue)(xue)習算(suan)(suan)(suan)(suan)法,通(tong)過數(shu)(shu)學(xue)(xue)(xue)和(he)計算(suan)(suan)(suan)(suan)分(fen)析(xi),系統(tong)的(de)(de)(de)(de)工程和(he)技(ji)(ji)術應用,大量使用神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡的(de)(de)(de)(de)概念和(he)技(ji)(ji)術。這一(yi)獨特(te)而廣(guang)泛的(de)(de)(de)(de)范圍促進了生物和(he)技(ji)(ji)術研究之(zhi)間的(de)(de)(de)(de)思想交流,并有助于促進對生物啟發(fa)的(de)(de)(de)(de)計算(suan)(suan)(suan)(suan)智(zhi)能(neng)感(gan)興趣的(de)(de)(de)(de)跨學(xue)(xue)(xue)科社區的(de)(de)(de)(de)發(fa)展(zhan)。因此(ci),神(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)絡編委會(hui)(hui)代表(biao)(biao)的(de)(de)(de)(de)專(zhuan)家領域包括心理學(xue)(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)生物學(xue)(xue)(xue),計算(suan)(suan)(suan)(suan)機科學(xue)(xue)(xue),工程,數(shu)(shu)學(xue)(xue)(xue),物理。該雜志發(fa)表(biao)(biao)文(wen)章(zhang)、信(xin)件和(he)評論以(yi)及給編輯(ji)的(de)(de)(de)(de)信(xin)件、社論、時事(shi)、軟件調查(cha)和(he)專(zhuan)利信(xin)息。文(wen)章(zhang)發(fa)表(biao)(biao)在五(wu)個(ge)部分(fen)之(zhi)一(yi):認知科學(xue)(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)科學(xue)(xue)(xue),學(xue)(xue)(xue)習系統(tong),數(shu)(shu)學(xue)(xue)(xue)和(he)計算(suan)(suan)(suan)(suan)分(fen)析(xi)、工程和(he)應用。 官(guan)網(wang)(wang)(wang)地址:

As artificial intelligence (AI) models continue to scale up, they are becoming more capable and integrated into various forms of decision-making systems. For models involved in moral decision-making, also known as artificial moral agents (AMA), interpretability provides a way to trust and understand the agent's internal reasoning mechanisms for effective use and error correction. In this paper, we provide an overview of this rapidly-evolving sub-field of AI interpretability, introduce the concept of the Minimum Level of Interpretability (MLI) and recommend an MLI for various types of agents, to aid their safe deployment in real-world settings.

In pace with developments in the research field of artificial intelligence, knowledge graphs (KGs) have attracted a surge of interest from both academia and industry. As a representation of semantic relations between entities, KGs have proven to be particularly relevant for natural language processing (NLP), experiencing a rapid spread and wide adoption within recent years. Given the increasing amount of research work in this area, several KG-related approaches have been surveyed in the NLP research community. However, a comprehensive study that categorizes established topics and reviews the maturity of individual research streams remains absent to this day. Contributing to closing this gap, we systematically analyzed 507 papers from the literature on KGs in NLP. Our survey encompasses a multifaceted review of tasks, research types, and contributions. As a result, we present a structured overview of the research landscape, provide a taxonomy of tasks, summarize our findings, and highlight directions for future work.

In the past decade, we have witnessed the rise of deep learning to dominate the field of artificial intelligence. Advances in artificial neural networks alongside corresponding advances in hardware accelerators with large memory capacity, together with the availability of large datasets enabled researchers and practitioners alike to train and deploy sophisticated neural network models that achieve state-of-the-art performance on tasks across several fields spanning computer vision, natural language processing, and reinforcement learning. However, as these neural networks become bigger, more complex, and more widely used, fundamental problems with current deep learning models become more apparent. State-of-the-art deep learning models are known to suffer from issues that range from poor robustness, inability to adapt to novel task settings, to requiring rigid and inflexible configuration assumptions. Ideas from collective intelligence, in particular concepts from complex systems such as self-organization, emergent behavior, swarm optimization, and cellular systems tend to produce solutions that are robust, adaptable, and have less rigid assumptions about the environment configuration. It is therefore natural to see these ideas incorporated into newer deep learning methods. In this review, we will provide a historical context of neural network research's involvement with complex systems, and highlight several active areas in modern deep learning research that incorporate the principles of collective intelligence to advance its current capabilities. To facilitate a bi-directional flow of ideas, we also discuss work that utilize modern deep learning models to help advance complex systems research. We hope this review can serve as a bridge between complex systems and deep learning communities to facilitate the cross pollination of ideas and foster new collaborations across disciplines.

In contrast to batch learning where all training data is available at once, continual learning represents a family of methods that accumulate knowledge and learn continuously with data available in sequential order. Similar to the human learning process with the ability of learning, fusing, and accumulating new knowledge coming at different time steps, continual learning is considered to have high practical significance. Hence, continual learning has been studied in various artificial intelligence tasks. In this paper, we present a comprehensive review of the recent progress of continual learning in computer vision. In particular, the works are grouped by their representative techniques, including regularization, knowledge distillation, memory, generative replay, parameter isolation, and a combination of the above techniques. For each category of these techniques, both its characteristics and applications in computer vision are presented. At the end of this overview, several subareas, where continuous knowledge accumulation is potentially helpful while continual learning has not been well studied, are discussed.

Data augmentation, the artificial creation of training data for machine learning by transformations, is a widely studied research field across machine learning disciplines. While it is useful for increasing the generalization capabilities of a model, it can also address many other challenges and problems, from overcoming a limited amount of training data over regularizing the objective to limiting the amount data used to protect privacy. Based on a precise description of the goals and applications of data augmentation (C1) and a taxonomy for existing works (C2), this survey is concerned with data augmentation methods for textual classification and aims to achieve a concise and comprehensive overview for researchers and practitioners (C3). Derived from the taxonomy, we divided more than 100 methods into 12 different groupings and provide state-of-the-art references expounding which methods are highly promising (C4). Finally, research perspectives that may constitute a building block for future work are given (C5).

Despite its great success, machine learning can have its limits when dealing with insufficient training data. A potential solution is the additional integration of prior knowledge into the training process which leads to the notion of informed machine learning. In this paper, we present a structured overview of various approaches in this field. We provide a definition and propose a concept for informed machine learning which illustrates its building blocks and distinguishes it from conventional machine learning. We introduce a taxonomy that serves as a classification framework for informed machine learning approaches. It considers the source of knowledge, its representation, and its integration into the machine learning pipeline. Based on this taxonomy, we survey related research and describe how different knowledge representations such as algebraic equations, logic rules, or simulation results can be used in learning systems. This evaluation of numerous papers on the basis of our taxonomy uncovers key methods in the field of informed machine learning.

Sampling methods (e.g., node-wise, layer-wise, or subgraph) has become an indispensable strategy to speed up training large-scale Graph Neural Networks (GNNs). However, existing sampling methods are mostly based on the graph structural information and ignore the dynamicity of optimization, which leads to high variance in estimating the stochastic gradients. The high variance issue can be very pronounced in extremely large graphs, where it results in slow convergence and poor generalization. In this paper, we theoretically analyze the variance of sampling methods and show that, due to the composite structure of empirical risk, the variance of any sampling method can be decomposed into \textit{embedding approximation variance} in the forward stage and \textit{stochastic gradient variance} in the backward stage that necessities mitigating both types of variance to obtain faster convergence rate. We propose a decoupled variance reduction strategy that employs (approximate) gradient information to adaptively sample nodes with minimal variance, and explicitly reduces the variance introduced by embedding approximation. We show theoretically and empirically that the proposed method, even with smaller mini-batch sizes, enjoys a faster convergence rate and entails a better generalization compared to the existing methods.

Spectral clustering (SC) is a popular clustering technique to find strongly connected communities on a graph. SC can be used in Graph Neural Networks (GNNs) to implement pooling operations that aggregate nodes belonging to the same cluster. However, the eigendecomposition of the Laplacian is expensive and, since clustering results are graph-specific, pooling methods based on SC must perform a new optimization for each new sample. In this paper, we propose a graph clustering approach that addresses these limitations of SC. We formulate a continuous relaxation of the normalized minCUT problem and train a GNN to compute cluster assignments that minimize this objective. Our GNN-based implementation is differentiable, does not require to compute the spectral decomposition, and learns a clustering function that can be quickly evaluated on out-of-sample graphs. From the proposed clustering method, we design a graph pooling operator that overcomes some important limitations of state-of-the-art graph pooling techniques and achieves the best performance in several supervised and unsupervised tasks.

Neural machine translation (NMT) is a deep learning based approach for machine translation, which yields the state-of-the-art translation performance in scenarios where large-scale parallel corpora are available. Although the high-quality and domain-specific translation is crucial in the real world, domain-specific corpora are usually scarce or nonexistent, and thus vanilla NMT performs poorly in such scenarios. Domain adaptation that leverages both out-of-domain parallel corpora as well as monolingual corpora for in-domain translation, is very important for domain-specific translation. In this paper, we give a comprehensive survey of the state-of-the-art domain adaptation techniques for NMT.

Recently, deep learning has achieved very promising results in visual object tracking. Deep neural networks in existing tracking methods require a lot of training data to learn a large number of parameters. However, training data is not sufficient for visual object tracking as annotations of a target object are only available in the first frame of a test sequence. In this paper, we propose to learn hierarchical features for visual object tracking by using tree structure based Recursive Neural Networks (RNN), which have fewer parameters than other deep neural networks, e.g. Convolutional Neural Networks (CNN). First, we learn RNN parameters to discriminate between the target object and background in the first frame of a test sequence. Tree structure over local patches of an exemplar region is randomly generated by using a bottom-up greedy search strategy. Given the learned RNN parameters, we create two dictionaries regarding target regions and corresponding local patches based on the learned hierarchical features from both top and leaf nodes of multiple random trees. In each of the subsequent frames, we conduct sparse dictionary coding on all candidates to select the best candidate as the new target location. In addition, we online update two dictionaries to handle appearance changes of target objects. Experimental results demonstrate that our feature learning algorithm can significantly improve tracking performance on benchmark datasets.

北京阿比特科技有限公司