亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Input gradients have a pivotal role in a variety of applications, including adversarial attack algorithms for evaluating model robustness, explainable AI techniques for generating Saliency Maps, and counterfactual explanations. However, Saliency Maps generated by traditional neural networks are often noisy and provide limited insights. In this paper, we demonstrate that, on the contrary, the Saliency Maps of 1-Lipschitz neural networks, learnt with the dual loss of an optimal transportation problem, exhibit desirable XAI properties: They are highly concentrated on the essential parts of the image with low noise, significantly outperforming state-of-the-art explanation approaches across various models and metrics. We also prove that these maps align unprecedentedly well with human explanations on ImageNet. To explain the particularly beneficial properties of the Saliency Map for such models, we prove this gradient encodes both the direction of the transportation plan and the direction towards the nearest adversarial attack. Following the gradient down to the decision boundary is no longer considered an adversarial attack, but rather a counterfactual explanation that explicitly transports the input from one class to another. Thus, Learning with such a loss jointly optimizes the classification objective and the alignment of the gradient , i.e. the Saliency Map, to the transportation plan direction. These networks were previously known to be certifiably robust by design, and we demonstrate that they scale well for large problems and models, and are tailored for explainability using a fast and straightforward method.

相關內容

神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(Neural Networks)是世(shi)界(jie)上三個(ge)最(zui)古老(lao)的(de)(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)建模學(xue)(xue)會(hui)的(de)(de)(de)檔案期刊:國際(ji)(ji)神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(INNS)、歐洲(zhou)神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(ENNS)和(he)(he)(he)日本神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(JNNS)。神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)提(ti)(ti)供了一(yi)個(ge)論(lun)(lun)(lun)壇,以發(fa)展和(he)(he)(he)培(pei)育一(yi)個(ge)國際(ji)(ji)社(she)會(hui)的(de)(de)(de)學(xue)(xue)者(zhe)和(he)(he)(he)實(shi)踐者(zhe)感興(xing)趣(qu)的(de)(de)(de)所(suo)有(you)方(fang)面(mian)的(de)(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)和(he)(he)(he)相關(guan)方(fang)法(fa)的(de)(de)(de)計(ji)算(suan)智(zhi)能。神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)歡迎高質(zhi)量論(lun)(lun)(lun)文的(de)(de)(de)提(ti)(ti)交(jiao),有(you)助于(yu)全面(mian)的(de)(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)研究,從行為和(he)(he)(he)大(da)腦建模,學(xue)(xue)習算(suan)法(fa),通過數學(xue)(xue)和(he)(he)(he)計(ji)算(suan)分(fen)析(xi),系統的(de)(de)(de)工程(cheng)和(he)(he)(he)技(ji)(ji)術應用(yong),大(da)量使用(yong)神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)的(de)(de)(de)概念和(he)(he)(he)技(ji)(ji)術。這一(yi)獨特而廣(guang)泛的(de)(de)(de)范圍(wei)促進了生物(wu)和(he)(he)(he)技(ji)(ji)術研究之間的(de)(de)(de)思想交(jiao)流,并(bing)有(you)助于(yu)促進對生物(wu)啟發(fa)的(de)(de)(de)計(ji)算(suan)智(zhi)能感興(xing)趣(qu)的(de)(de)(de)跨(kua)學(xue)(xue)科社(she)區(qu)的(de)(de)(de)發(fa)展。因此(ci),神(shen)(shen)(shen)經(jing)(jing)(jing)網(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)編(bian)委會(hui)代(dai)表的(de)(de)(de)專家領域(yu)包括心(xin)理(li)學(xue)(xue),神(shen)(shen)(shen)經(jing)(jing)(jing)生物(wu)學(xue)(xue),計(ji)算(suan)機科學(xue)(xue),工程(cheng),數學(xue)(xue),物(wu)理(li)。該雜志發(fa)表文章(zhang)、信件和(he)(he)(he)評論(lun)(lun)(lun)以及給編(bian)輯(ji)的(de)(de)(de)信件、社(she)論(lun)(lun)(lun)、時事(shi)、軟件調查和(he)(he)(he)專利信息。文章(zhang)發(fa)表在五(wu)個(ge)部分(fen)之一(yi):認知科學(xue)(xue),神(shen)(shen)(shen)經(jing)(jing)(jing)科學(xue)(xue),學(xue)(xue)習系統,數學(xue)(xue)和(he)(he)(he)計(ji)算(suan)分(fen)析(xi)、工程(cheng)和(he)(he)(he)應用(yong)。 官網(wang)(wang)(wang)(wang)(wang)地址:

Deep reinforcement learning algorithms are usually impeded by sampling inefficiency, heavily depending on multiple interactions with the environment to acquire accurate decision-making capabilities. In contrast, humans rely on their hippocampus to retrieve relevant information from past experiences of relevant tasks, which guides their decision-making when learning a new task, rather than exclusively depending on environmental interactions. Nevertheless, designing a hippocampus-like module for an agent to incorporate past experiences into established reinforcement learning algorithms presents two challenges. The first challenge involves selecting the most relevant past experiences for the current task, and the second challenge is integrating such experiences into the decision network. To address these challenges, we propose a novel method that utilizes a retrieval network based on task-conditioned hypernetwork, which adapts the retrieval network's parameters depending on the task. At the same time, a dynamic modification mechanism enhances the collaborative efforts between the retrieval and decision networks. We evaluate the proposed method on the MiniGrid environment.The experimental results demonstrate that our proposed method significantly outperforms strong baselines.

Diffusion probabilistic models (DPMs) are a powerful class of generative models known for their ability to generate high-fidelity image samples. A major challenge in the implementation of DPMs is the slow sampling process. In this work, we bring a high-efficiency sampler for DPMs. Specifically, we propose a score-based exact solution paradigm for the diffusion ODEs corresponding to the sampling process of DPMs, which introduces a new perspective on developing numerical algorithms for solving diffusion ODEs. To achieve an efficient sampler, we propose a recursive derivative estimation (RDE) method to reduce the estimation error. With our proposed solution paradigm and RDE method, we propose the score-integrand solver with the convergence order guarantee as efficient solver (SciRE-Solver) for solving diffusion ODEs. The SciRE-Solver attains state-of-the-art (SOTA) sampling performance with a limited number of score function evaluations (NFE) on both discrete-time and continuous-time DPMs in comparison to existing training-free sampling algorithms. Such as, we achieve $3.48$ FID with $12$ NFE and $2.42$ FID with $20$ NFE for continuous-time DPMs on CIFAR10, respectively. Different from other samplers, SciRE-Solver has the promising potential to surpass the FIDs achieved in the original papers of some pre-trained models with just fewer NFEs. For example, we reach SOTA value of $2.40$ FID with $100$ NFE for continuous-time DPM and of $3.15$ FID with $84$ NFE for discrete-time DPM on CIFAR-10, as well as of $2.17$ ($2.02$) FID with $18$ ($50$) NFE for discrete-time DPM on CelebA 64$\times$64.

In a traditional Gaussian graphical model, data homogeneity is routinely assumed with no extra variables affecting the conditional independence. In modern genomic datasets, there is an abundance of auxiliary information, which often gets under-utilized in determining the joint dependency structure. In this article, we consider a Bayesian approach to model undirected graphs underlying heterogeneous multivariate observations with additional assistance from covariates. Building on product partition models, we propose a novel covariate-dependent Gaussian graphical model that allows graphs to vary with covariates so that observations whose covariates are similar share a similar undirected graph. To efficiently embed Gaussian graphical models into our proposed framework, we explore both Gaussian likelihood and pseudo-likelihood functions. For Gaussian likelihood, a G-Wishart distribution is used as a natural conjugate prior, and for the pseudo-likelihood, a product of Gaussian-conditionals is used. Moreover, the proposed model has large prior support and is flexible to approximate any $\nu$-H\"{o}lder conditional variance-covariance matrices with $\nu\in(0,1]$. We further show that based on the theory of fractional likelihood, the rate of posterior contraction is minimax optimal assuming the true density to be a Gaussian mixture with a known number of components. The efficacy of the approach is demonstrated via simulation studies and an analysis of a protein network for a breast cancer dataset assisted by mRNA gene expression as covariates.

Packet scheduling is a fundamental networking task that recently received renewed attention in the context of programmable data planes. Programmable packet scheduling systems such as those based on Push-In First-Out (PIFO) abstraction enabled flexible scheduling policies, but are too resource-expensive for large-scale line rate operation. This prompted research into practical programmable schedulers (e.g., SP-PIFO, AIFO) approximating PIFO behavior on regular hardware. Yet, their scalability remains limited due to extensive number of memory operations. To address this, we design an effective yet resource-efficient packet scheduler, Range-In First-Out (RIFO), which uses only three mutable memory cells and one FIFO queue per PIFO queue. RIFO is based on multi-criteria decision-making principles and uses small guaranteed admission buffers. Our large-scale simulations in Netbench demonstrate that despite using fewer resources, RIFO generally achieves competitive flow completion times across all studied workloads, and is especially effective in workloads with a significant share of large flows, reducing flow completion time up to 2.9x in Datamining workloads compared to state-of-the-art solutions. Our prototype implementation using P4 on Tofino switches requires only 650 lines of code, is scalable, and runs at line rate.

The stability, robustness, accuracy, and efficiency of space-time finite element methods crucially depend on the choice of approximation spaces for test and trial functions. This is especially true for high-order, mixed finite element methods which often must satisfy an inf-sup condition in order to ensure stability. With this in mind, the primary objective of this paper and a companion paper is to provide a wide range of explicitly stated, conforming, finite element spaces in four-dimensions. In this paper, we construct explicit high-order conforming finite elements on 4-cubes (tesseracts); our construction uses tools from the recently developed `Finite Element Exterior Calculus'. With a focus on practical implementation, we provide details including Piola-type transformations, and explicit expressions for the volumetric, facet, face, edge, and vertex degrees of freedom. In addition, we establish important theoretical properties, such as the exactness of the finite element sequences, and the unisolvence of the degrees of freedom.

Reasoning ability is one of the most crucial capabilities of a foundation model, signifying its capacity to address complex reasoning tasks. Chain-of-Thought (CoT) technique is widely regarded as one of the effective methods for enhancing the reasoning ability of foundation models and has garnered significant attention. However, the reasoning process of CoT is linear, step-by-step, similar to personal logical reasoning, suitable for solving general and slightly complicated problems. On the contrary, the thinking pattern of an expert owns two prominent characteristics that cannot be handled appropriately in CoT, i.e., high-order multi-hop reasoning and multimodal comparative judgement. Therefore, the core motivation of this paper is transcending CoT to construct a reasoning paradigm that can think like an expert. The hyperedge of a hypergraph could connect various vertices, making it naturally suitable for modelling high-order relationships. Inspired by this, this paper innovatively proposes a multimodal Hypergraph-of-Thought (HoT) reasoning paradigm, which enables the foundation models to possess the expert-level ability of high-order multi-hop reasoning and multimodal comparative judgement. Specifically, a textual hypergraph-of-thought is constructed utilizing triple as the primary thought to model higher-order relationships, and a hyperedge-of-thought is generated through multi-hop walking paths to achieve multi-hop inference. Furthermore, we devise a visual hypergraph-of-thought to interact with the textual hypergraph-of-thought via Cross-modal Co-Attention Graph Learning for multimodal comparative verification. Experimentations on the ScienceQA benchmark demonstrate the proposed HoT-based T5 outperforms CoT-based GPT3.5 and chatGPT, which is on par with CoT-based GPT4 with a lower model size.

The existence of representative datasets is a prerequisite of many successful artificial intelligence and machine learning models. However, the subsequent application of these models often involves scenarios that are inadequately represented in the data used for training. The reasons for this are manifold and range from time and cost constraints to ethical considerations. As a consequence, the reliable use of these models, especially in safety-critical applications, is a huge challenge. Leveraging additional, already existing sources of knowledge is key to overcome the limitations of purely data-driven approaches, and eventually to increase the generalization capability of these models. Furthermore, predictions that conform with knowledge are crucial for making trustworthy and safe decisions even in underrepresented scenarios. This work provides an overview of existing techniques and methods in the literature that combine data-based models with existing knowledge. The identified approaches are structured according to the categories integration, extraction and conformity. Special attention is given to applications in the field of autonomous driving.

Current models for event causality identification (ECI) mainly adopt a supervised framework, which heavily rely on labeled data for training. Unfortunately, the scale of current annotated datasets is relatively limited, which cannot provide sufficient support for models to capture useful indicators from causal statements, especially for handing those new, unseen cases. To alleviate this problem, we propose a novel approach, shortly named CauSeRL, which leverages external causal statements for event causality identification. First of all, we design a self-supervised framework to learn context-specific causal patterns from external causal statements. Then, we adopt a contrastive transfer strategy to incorporate the learned context-specific causal patterns into the target ECI model. Experimental results show that our method significantly outperforms previous methods on EventStoryLine and Causal-TimeBank (+2.0 and +3.4 points on F1 value respectively).

Generative commonsense reasoning which aims to empower machines to generate sentences with the capacity of reasoning over a set of concepts is a critical bottleneck for text generation. Even the state-of-the-art pre-trained language generation models struggle at this task and often produce implausible and anomalous sentences. One reason is that they rarely consider incorporating the knowledge graph which can provide rich relational information among the commonsense concepts. To promote the ability of commonsense reasoning for text generation, we propose a novel knowledge graph augmented pre-trained language generation model KG-BART, which encompasses the complex relations of concepts through the knowledge graph and produces more logical and natural sentences as output. Moreover, KG-BART can leverage the graph attention to aggregate the rich concept semantics that enhances the model generalization on unseen concept sets. Experiments on benchmark CommonGen dataset verify the effectiveness of our proposed approach by comparing with several strong pre-trained language generation models, particularly KG-BART outperforms BART by 5.80, 4.60, in terms of BLEU-3, 4. Moreover, we also show that the generated context by our model can work as background scenarios to benefit downstream commonsense QA tasks.

We introduce a multi-task setup of identifying and classifying entities, relations, and coreference clusters in scientific articles. We create SciERC, a dataset that includes annotations for all three tasks and develop a unified framework called Scientific Information Extractor (SciIE) for with shared span representations. The multi-task setup reduces cascading errors between tasks and leverages cross-sentence relations through coreference links. Experiments show that our multi-task model outperforms previous models in scientific information extraction without using any domain-specific features. We further show that the framework supports construction of a scientific knowledge graph, which we use to analyze information in scientific literature.

北京阿比特科技有限公司