亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Bayesian Neural Networks (BNNs) offer a principled and natural framework for proper uncertainty quantification in the context of deep learning. They address the typical challenges associated with conventional deep learning methods, such as data insatiability, ad-hoc nature, and susceptibility to overfitting. However, their implementation typically either relies on Markov chain Monte Carlo (MCMC) methods, which are characterized by their computational intensity and inefficiency in a high-dimensional space, or variational inference methods, which tend to underestimate uncertainty. To address this issue, we propose a novel Calibration-Emulation-Sampling (CES) strategy to significantly enhance the computational efficiency of BNN. In this framework, during the initial calibration stage, we collect a small set of samples from the parameter space. These samples serve as training data for the emulator, which approximates the map between parameters and posterior probability. The trained emulator is then used for sampling from the posterior distribution at substantially higher speed compared to the standard BNN. Using simulated and real data, we demonstrate that our proposed method improves computational efficiency of BNN, while maintaining similar performance in terms of prediction accuracy and uncertainty quantification.

相關內容

神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)(Neural Networks)是(shi)世界上三個最古老的(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)建模學(xue)(xue)會(hui)的(de)(de)(de)(de)檔案(an)期(qi)刊:國際神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(INNS)、歐(ou)洲(zhou)神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(ENNS)和(he)日本神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(JNNS)。神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)提供(gong)了一個論(lun)(lun)壇(tan),以發(fa)展和(he)培育一個國際社(she)(she)會(hui)的(de)(de)(de)(de)學(xue)(xue)者(zhe)和(he)實(shi)踐者(zhe)感(gan)興趣的(de)(de)(de)(de)所有方(fang)面的(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)和(he)相關(guan)方(fang)法的(de)(de)(de)(de)計(ji)(ji)算(suan)智(zhi)能。神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)歡(huan)迎高質(zhi)量論(lun)(lun)文的(de)(de)(de)(de)提交,有助于全面的(de)(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)研(yan)究(jiu),從行為和(he)大腦建模,學(xue)(xue)習(xi)算(suan)法,通過數學(xue)(xue)和(he)計(ji)(ji)算(suan)分(fen)析,系(xi)統的(de)(de)(de)(de)工程和(he)技(ji)術(shu)(shu)應用(yong),大量使用(yong)神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)的(de)(de)(de)(de)概念和(he)技(ji)術(shu)(shu)。這一獨特而廣(guang)泛的(de)(de)(de)(de)范圍促進了生(sheng)物(wu)(wu)(wu)和(he)技(ji)術(shu)(shu)研(yan)究(jiu)之(zhi)間的(de)(de)(de)(de)思想(xiang)交流(liu),并有助于促進對生(sheng)物(wu)(wu)(wu)啟發(fa)的(de)(de)(de)(de)計(ji)(ji)算(suan)智(zhi)能感(gan)興趣的(de)(de)(de)(de)跨學(xue)(xue)科社(she)(she)區的(de)(de)(de)(de)發(fa)展。因此,神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)編(bian)委(wei)會(hui)代表的(de)(de)(de)(de)專(zhuan)(zhuan)家領域包括心理學(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)生(sheng)物(wu)(wu)(wu)學(xue)(xue),計(ji)(ji)算(suan)機科學(xue)(xue),工程,數學(xue)(xue),物(wu)(wu)(wu)理。該雜志發(fa)表文章、信(xin)件和(he)評論(lun)(lun)以及給(gei)編(bian)輯的(de)(de)(de)(de)信(xin)件、社(she)(she)論(lun)(lun)、時事、軟件調查和(he)專(zhuan)(zhuan)利信(xin)息。文章發(fa)表在(zai)五個部分(fen)之(zhi)一:認知(zhi)科學(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)(jing)(jing)科學(xue)(xue),學(xue)(xue)習(xi)系(xi)統,數學(xue)(xue)和(he)計(ji)(ji)算(suan)分(fen)析、工程和(he)應用(yong)。 官(guan)網(wang)(wang)地址(zhi):

Graph Neural Networks (GNNs) have shown promising results on a broad spectrum of applications. Most empirical studies of GNNs directly take the observed graph as input, assuming the observed structure perfectly depicts the accurate and complete relations between nodes. However, graphs in the real world are inevitably noisy or incomplete, which could even exacerbate the quality of graph representations. In this work, we propose a novel Variational Information Bottleneck guided Graph Structure Learning framework, namely VIB-GSL, in the perspective of information theory. VIB-GSL advances the Information Bottleneck (IB) principle for graph structure learning, providing a more elegant and universal framework for mining underlying task-relevant relations. VIB-GSL learns an informative and compressive graph structure to distill the actionable information for specific downstream tasks. VIB-GSL deduces a variational approximation for irregular graph data to form a tractable IB objective function, which facilitates training stability. Extensive experimental results demonstrate that the superior effectiveness and robustness of VIB-GSL.

Recent contrastive representation learning methods rely on estimating mutual information (MI) between multiple views of an underlying context. E.g., we can derive multiple views of a given image by applying data augmentation, or we can split a sequence into views comprising the past and future of some step in the sequence. Contrastive lower bounds on MI are easy to optimize, but have a strong underestimation bias when estimating large amounts of MI. We propose decomposing the full MI estimation problem into a sum of smaller estimation problems by splitting one of the views into progressively more informed subviews and by applying the chain rule on MI between the decomposed views. This expression contains a sum of unconditional and conditional MI terms, each measuring modest chunks of the total MI, which facilitates approximation via contrastive bounds. To maximize the sum, we formulate a contrastive lower bound on the conditional MI which can be approximated efficiently. We refer to our general approach as Decomposed Estimation of Mutual Information (DEMI). We show that DEMI can capture a larger amount of MI than standard non-decomposed contrastive bounds in a synthetic setting, and learns better representations in a vision domain and for dialogue generation.

Recently, contrastive learning (CL) has emerged as a successful method for unsupervised graph representation learning. Most graph CL methods first perform stochastic augmentation on the input graph to obtain two graph views and maximize the agreement of representations in the two views. Despite the prosperous development of graph CL methods, the design of graph augmentation schemes -- a crucial component in CL -- remains rarely explored. We argue that the data augmentation schemes should preserve intrinsic structures and attributes of graphs, which will force the model to learn representations that are insensitive to perturbation on unimportant nodes and edges. However, most existing methods adopt uniform data augmentation schemes, like uniformly dropping edges and uniformly shuffling features, leading to suboptimal performance. In this paper, we propose a novel graph contrastive representation learning method with adaptive augmentation that incorporates various priors for topological and semantic aspects of the graph. Specifically, on the topology level, we design augmentation schemes based on node centrality measures to highlight important connective structures. On the node attribute level, we corrupt node features by adding more noise to unimportant node features, to enforce the model to recognize underlying semantic information. We perform extensive experiments of node classification on a variety of real-world datasets. Experimental results demonstrate that our proposed method consistently outperforms existing state-of-the-art baselines and even surpasses some supervised counterparts, which validates the effectiveness of the proposed contrastive framework with adaptive augmentation.

Graph Neural Networks (GNNs), which generalize deep neural networks to graph-structured data, have drawn considerable attention and achieved state-of-the-art performance in numerous graph related tasks. However, existing GNN models mainly focus on designing graph convolution operations. The graph pooling (or downsampling) operations, that play an important role in learning hierarchical representations, are usually overlooked. In this paper, we propose a novel graph pooling operator, called Hierarchical Graph Pooling with Structure Learning (HGP-SL), which can be integrated into various graph neural network architectures. HGP-SL incorporates graph pooling and structure learning into a unified module to generate hierarchical representations of graphs. More specifically, the graph pooling operation adaptively selects a subset of nodes to form an induced subgraph for the subsequent layers. To preserve the integrity of graph's topological information, we further introduce a structure learning mechanism to learn a refined graph structure for the pooled graph at each layer. By combining HGP-SL operator with graph neural networks, we perform graph level representation learning with focus on graph classification task. Experimental results on six widely used benchmarks demonstrate the effectiveness of our proposed model.

Incompleteness is a common problem for existing knowledge graphs (KGs), and the completion of KG which aims to predict links between entities is challenging. Most existing KG completion methods only consider the direct relation between nodes and ignore the relation paths which contain useful information for link prediction. Recently, a few methods take relation paths into consideration but pay less attention to the order of relations in paths which is important for reasoning. In addition, these path-based models always ignore nonlinear contributions of path features for link prediction. To solve these problems, we propose a novel KG completion method named OPTransE. Instead of embedding both entities of a relation into the same latent space as in previous methods, we project the head entity and the tail entity of each relation into different spaces to guarantee the order of relations in the path. Meanwhile, we adopt a pooling strategy to extract nonlinear and complex features of different paths to further improve the performance of link prediction. Experimental results on two benchmark datasets show that the proposed model OPTransE performs better than state-of-the-art methods.

It is important to detect anomalous inputs when deploying machine learning systems. The use of larger and more complex inputs in deep learning magnifies the difficulty of distinguishing between anomalous and in-distribution examples. At the same time, diverse image and text data are available in enormous quantities. We propose leveraging these data to improve deep anomaly detection by training anomaly detectors against an auxiliary dataset of outliers, an approach we call Outlier Exposure (OE). This enables anomaly detectors to generalize and detect unseen anomalies. In extensive experiments on natural language processing and small- and large-scale vision tasks, we find that Outlier Exposure significantly improves detection performance. We also observe that cutting-edge generative models trained on CIFAR-10 may assign higher likelihoods to SVHN images than to CIFAR-10 images; we use OE to mitigate this issue. We also analyze the flexibility and robustness of Outlier Exposure, and identify characteristics of the auxiliary dataset that improve performance.

Recently, graph neural networks (GNNs) have revolutionized the field of graph representation learning through effectively learned node embeddings, and achieved state-of-the-art results in tasks such as node classification and link prediction. However, current GNN methods are inherently flat and do not learn hierarchical representations of graphs---a limitation that is especially problematic for the task of graph classification, where the goal is to predict the label associated with an entire graph. Here we propose DiffPool, a differentiable graph pooling module that can generate hierarchical representations of graphs and can be combined with various graph neural network architectures in an end-to-end fashion. DiffPool learns a differentiable soft cluster assignment for nodes at each layer of a deep GNN, mapping nodes to a set of clusters, which then form the coarsened input for the next GNN layer. Our experimental results show that combining existing GNN methods with DiffPool yields an average improvement of 5-10% accuracy on graph classification benchmarks, compared to all existing pooling approaches, achieving a new state-of-the-art on four out of five benchmark data sets.

Multi-relation Question Answering is a challenging task, due to the requirement of elaborated analysis on questions and reasoning over multiple fact triples in knowledge base. In this paper, we present a novel model called Interpretable Reasoning Network that employs an interpretable, hop-by-hop reasoning process for question answering. The model dynamically decides which part of an input question should be analyzed at each hop; predicts a relation that corresponds to the current parsed results; utilizes the predicted relation to update the question representation and the state of the reasoning process; and then drives the next-hop reasoning. Experiments show that our model yields state-of-the-art results on two datasets. More interestingly, the model can offer traceable and observable intermediate predictions for reasoning analysis and failure diagnosis, thereby allowing manual manipulation in predicting the final answer.

Aspect based sentiment analysis (ABSA) can provide more detailed information than general sentiment analysis, because it aims to predict the sentiment polarities of the given aspects or entities in text. We summarize previous approaches into two subtasks: aspect-category sentiment analysis (ACSA) and aspect-term sentiment analysis (ATSA). Most previous approaches employ long short-term memory and attention mechanisms to predict the sentiment polarity of the concerned targets, which are often complicated and need more training time. We propose a model based on convolutional neural networks and gating mechanisms, which is more accurate and efficient. First, the novel Gated Tanh-ReLU Units can selectively output the sentiment features according to the given aspect or entity. The architecture is much simpler than attention layer used in the existing models. Second, the computations of our model could be easily parallelized during training, because convolutional layers do not have time dependency as in LSTM layers, and gating units also work independently. The experiments on SemEval datasets demonstrate the efficiency and effectiveness of our models.

We propose a new method for event extraction (EE) task based on an imitation learning framework, specifically, inverse reinforcement learning (IRL) via generative adversarial network (GAN). The GAN estimates proper rewards according to the difference between the actions committed by the expert (or ground truth) and the agent among complicated states in the environment. EE task benefits from these dynamic rewards because instances and labels yield to various extents of difficulty and the gains are expected to be diverse -- e.g., an ambiguous but correctly detected trigger or argument should receive high gains -- while the traditional RL models usually neglect such differences and pay equal attention on all instances. Moreover, our experiments also demonstrate that the proposed framework outperforms state-of-the-art methods, without explicit feature engineering.

北京阿比特科技有限公司