亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Spiking Neural Networks (SNNs) are well known as a promising energy-efficient alternative to conventional artificial neural networks. Subject to the preconceived impression that SNNs are sparse firing, the analysis and optimization of inherent redundancy in SNNs have been largely overlooked, thus the potential advantages of spike-based neuromorphic computing in accuracy and energy efficiency are interfered. In this work, we pose and focus on three key questions regarding the inherent redundancy in SNNs. We argue that the redundancy is induced by the spatio-temporal invariance of SNNs, which enhances the efficiency of parameter utilization but also invites lots of noise spikes. Further, we analyze the effect of spatio-temporal invariance on the spatio-temporal dynamics and spike firing of SNNs. Then, motivated by these analyses, we propose an Advance Spatial Attention (ASA) module to harness SNNs' redundancy, which can adaptively optimize their membrane potential distribution by a pair of individual spatial attention sub-modules. In this way, noise spike features are accurately regulated. Experimental results demonstrate that the proposed method can significantly drop the spike firing with better performance than state-of-the-art SNN baselines. Our code is available in \url{//github.com/BICLab/ASA-SNN}.

相關內容

神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)(Neural Networks)是世(shi)界上三(san)個最古老的(de)神(shen)(shen)經(jing)(jing)(jing)(jing)建模學(xue)(xue)(xue)(xue)(xue)(xue)會的(de)檔案期刊:國(guo)際神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)(xue)會(INNS)、歐洲神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)(xue)會(ENNS)和(he)日本神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)(xue)會(JNNS)。神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)提供了(le)一個論(lun)(lun)壇,以發(fa)展和(he)培育一個國(guo)際社會的(de)學(xue)(xue)(xue)(xue)(xue)(xue)者(zhe)和(he)實(shi)踐者(zhe)感(gan)(gan)興趣的(de)所有(you)方面(mian)的(de)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)和(he)相關方法的(de)計(ji)(ji)(ji)算(suan)智(zhi)能。神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)歡迎高質量論(lun)(lun)文(wen)的(de)提交,有(you)助(zhu)于全面(mian)的(de)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)研(yan)究,從(cong)行(xing)為和(he)大腦(nao)建模,學(xue)(xue)(xue)(xue)(xue)(xue)習(xi)算(suan)法,通過數(shu)學(xue)(xue)(xue)(xue)(xue)(xue)和(he)計(ji)(ji)(ji)算(suan)分(fen)(fen)析,系(xi)統的(de)工程和(he)技(ji)(ji)術應(ying)用(yong),大量使用(yong)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)的(de)概念和(he)技(ji)(ji)術。這(zhe)一獨特而廣泛的(de)范圍促(cu)進(jin)了(le)生(sheng)物(wu)(wu)和(he)技(ji)(ji)術研(yan)究之間的(de)思想交流,并有(you)助(zhu)于促(cu)進(jin)對生(sheng)物(wu)(wu)啟發(fa)的(de)計(ji)(ji)(ji)算(suan)智(zhi)能感(gan)(gan)興趣的(de)跨(kua)學(xue)(xue)(xue)(xue)(xue)(xue)科(ke)社區的(de)發(fa)展。因此(ci),神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)(wang)絡(luo)編委會代(dai)表(biao)的(de)專家(jia)領域包括心理(li)學(xue)(xue)(xue)(xue)(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)(jing)生(sheng)物(wu)(wu)學(xue)(xue)(xue)(xue)(xue)(xue),計(ji)(ji)(ji)算(suan)機科(ke)學(xue)(xue)(xue)(xue)(xue)(xue),工程,數(shu)學(xue)(xue)(xue)(xue)(xue)(xue),物(wu)(wu)理(li)。該雜(za)志發(fa)表(biao)文(wen)章、信件(jian)和(he)評(ping)論(lun)(lun)以及給編輯的(de)信件(jian)、社論(lun)(lun)、時(shi)事、軟件(jian)調查和(he)專利信息。文(wen)章發(fa)表(biao)在五個部分(fen)(fen)之一:認知(zhi)科(ke)學(xue)(xue)(xue)(xue)(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)(jing)科(ke)學(xue)(xue)(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)(xue)(xue)習(xi)系(xi)統,數(shu)學(xue)(xue)(xue)(xue)(xue)(xue)和(he)計(ji)(ji)(ji)算(suan)分(fen)(fen)析、工程和(he)應(ying)用(yong)。 官(guan)網(wang)(wang)(wang)(wang)地址:

In copula models the marginal distributions and copula function are specified separately. We treat these as two modules in a modular Bayesian inference framework, and propose conducting modified Bayesian inference by ``cutting feedback''. Cutting feedback limits the influence of potentially misspecified modules in posterior inference. We consider two types of cuts. The first limits the influence of a misspecified copula on inference for the marginals, which is a Bayesian analogue of the popular Inference for Margins (IFM) estimator. The second limits the influence of misspecified marginals on inference for the copula parameters by using a rank likelihood to define the cut model. We establish that if only one of the modules is misspecified, then the appropriate cut posterior gives accurate uncertainty quantification asymptotically for the parameters in the other module. Computation of the cut posteriors is difficult, and new variational inference methods to do so are proposed. The efficacy of the new methodology is demonstrated using both simulated data and a substantive multivariate time series application from macroeconomic forecasting. In the latter, cutting feedback from misspecified marginals to a 1096 dimension copula improves posterior inference and predictive accuracy greatly, compared to conventional Bayesian inference.

The Internet of Things (IoT) consistently generates vast amounts of data, sparking increasing concern over the protection of data privacy and the limitation of data misuse. Federated learning (FL) facilitates collaborative capabilities among multiple parties by sharing machine learning (ML) model parameters instead of raw user data, and it has recently gained significant attention for its potential in privacy preservation and learning efficiency enhancement. In this paper, we highlight the digital ethics concerns that arise when human-centric devices serve as clients in FL. More specifically, challenges of game dynamics, fairness, incentive, and continuity arise in FL due to differences in perspectives and objectives between clients and the server. We analyze these challenges and their solutions from the perspectives of both the client and the server, and through the viewpoints of centralized and decentralized FL. Finally, we explore the opportunities in FL for human-centric IoT as directions for future development.

Neural networks often suffer from a feature preference problem, where they tend to overly rely on specific features to solve a task while disregarding other features, even if those neglected features are essential for the task. Feature preference problems have primarily been investigated in classification task. However, we observe that feature preference occurs in high-dimensional regression task, specifically, source separation. To mitigate feature preference in source separation, we propose FEAture BAlancing by Suppressing Easy feature (FEABASE). This approach enables efficient data utilization by learning hidden information about the neglected feature. We evaluate our method in a multi-channel source separation task, where feature preference between spatial feature and timbre feature appears.

Fictitious Play (FP) is a simple and natural dynamic for repeated play with many applications in game theory and multi-agent reinforcement learning. It was introduced by Brown (1949,1951) and its convergence properties for two-player zero-sum games was established later by Robinson (1951). Potential games Monderer and Shapley (1996b) is another class of games which exhibit the FP property (Monderer and Shapley (1996a)), i.e., FP dynamics converges to a Nash equilibrium if all agents follows it. Nevertheless, except for two-player zero-sum games and for specific instances of payoff matrices (Abernethy et al. (2021)) or for adversarial tie-breaking rules (Daskalakis and Pan (2014)), the convergence rate of FP is unknown. In this work, we focus on the rate of convergence of FP when applied to potential games and more specifically identical payoff games. We prove that FP can take exponential time (in the number of strategies) to reach a Nash equilibrium, even if the game is restricted to two agents and for arbitrary tie-breaking rules. To prove this, we recursively construct a two-player coordination game with a unique Nash equilibrium. Moreover, every approximate Nash equilibrium in the constructed game must be close to the pure Nash equilibrium in $\ell_1$-distance.

We investigate the unsupervised constituency parsing task, which organizes words and phrases of a sentence into a hierarchical structure without using linguistically annotated data. We observe that existing unsupervised parsers capture differing aspects of parsing structures, which can be leveraged to enhance unsupervised parsing performance. To this end, we propose a notion of "tree averaging," based on which we further propose a novel ensemble method for unsupervised parsing. To improve inference efficiency, we further distill the ensemble knowledge into a student model; such an ensemble-then-distill process is an effective approach to mitigate the over-smoothing problem existing in common multi-teacher distilling methods. Experiments show that our method surpasses all previous approaches, consistently demonstrating its effectiveness and robustness across various runs, with different ensemble components, and under domain-shift conditions.

In this work, we propose and analyse a weak Galerkin method for the electrical impedance tomography based on a bounded variation regularization. We use the complete electrode model as the forward system that is approximated by a weak Galerkin method with lowest order. The error estimates are studied for the forward problem, which are used to establish the convergence of this weak Galerkin algorithm for the inverse problem. Numerical examples are presented to verify the effectiveness and efficiency of the weak Galerkin algorithm for the electrical impedance tomography.

Large Language Models (LLMs) have shown excellent generalization capabilities that have led to the development of numerous models. These models propose various new architectures, tweaking existing architectures with refined training strategies, increasing context length, using high-quality training data, and increasing training time to outperform baselines. Analyzing new developments is crucial for identifying changes that enhance training stability and improve generalization in LLMs. This survey paper comprehensively analyses the LLMs architectures and their categorization, training strategies, training datasets, and performance evaluations and discusses future research directions. Moreover, the paper also discusses the basic building blocks and concepts behind LLMs, followed by a complete overview of LLMs, including their important features and functions. Finally, the paper summarizes significant findings from LLM research and consolidates essential architectural and training strategies for developing advanced LLMs. Given the continuous advancements in LLMs, we intend to regularly update this paper by incorporating new sections and featuring the latest LLM models.

Transformer is a type of deep neural network mainly based on self-attention mechanism which is originally applied in natural language processing field. Inspired by the strong representation ability of transformer, researchers propose to extend transformer for computer vision tasks. Transformer-based models show competitive and even better performance on various visual benchmarks compared to other network types such as convolutional networks and recurrent networks. In this paper we provide a literature review of these visual transformer models by categorizing them in different tasks and analyze the advantages and disadvantages of these methods. In particular, the main categories include the basic image classification, high-level vision, low-level vision and video processing. Self-attention in computer vision is also briefly revisited as self-attention is the base component in transformer. Efficient transformer methods are included for pushing transformer into real applications. Finally, we give a discussion about the further research directions for visual transformer.

Graph Neural Networks (GNNs) have recently become increasingly popular due to their ability to learn complex systems of relations or interactions arising in a broad spectrum of problems ranging from biology and particle physics to social networks and recommendation systems. Despite the plethora of different models for deep learning on graphs, few approaches have been proposed thus far for dealing with graphs that present some sort of dynamic nature (e.g. evolving features or connectivity over time). In this paper, we present Temporal Graph Networks (TGNs), a generic, efficient framework for deep learning on dynamic graphs represented as sequences of timed events. Thanks to a novel combination of memory modules and graph-based operators, TGNs are able to significantly outperform previous approaches being at the same time more computationally efficient. We furthermore show that several previous models for learning on dynamic graphs can be cast as specific instances of our framework. We perform a detailed ablation study of different components of our framework and devise the best configuration that achieves state-of-the-art performance on several transductive and inductive prediction tasks for dynamic graphs.

Deep neural networks (DNNs) have been found to be vulnerable to adversarial examples resulting from adding small-magnitude perturbations to inputs. Such adversarial examples can mislead DNNs to produce adversary-selected results. Different attack strategies have been proposed to generate adversarial examples, but how to produce them with high perceptual quality and more efficiently requires more research efforts. In this paper, we propose AdvGAN to generate adversarial examples with generative adversarial networks (GANs), which can learn and approximate the distribution of original instances. For AdvGAN, once the generator is trained, it can generate adversarial perturbations efficiently for any instance, so as to potentially accelerate adversarial training as defenses. We apply AdvGAN in both semi-whitebox and black-box attack settings. In semi-whitebox attacks, there is no need to access the original target model after the generator is trained, in contrast to traditional white-box attacks. In black-box attacks, we dynamically train a distilled model for the black-box model and optimize the generator accordingly. Adversarial examples generated by AdvGAN on different target models have high attack success rate under state-of-the-art defenses compared to other attacks. Our attack has placed the first with 92.76% accuracy on a public MNIST black-box attack challenge.

北京阿比特科技有限公司