亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Advancement in finite element methods have become essential in various disciplines, and in particular for Computational Fluid Dynamics (CFD), driving research efforts for improved precision and efficiency. While Convolutional Neural Networks (CNNs) have found success in CFD by mapping meshes into images, recent attention has turned to leveraging Graph Neural Networks (GNNs) for direct mesh processing. This paper introduces a novel model merging Self-Attention with Message Passing in GNNs, achieving a 15\% reduction in RMSE on the well known flow past a cylinder benchmark. Furthermore, a dynamic mesh pruning technique based on Self-Attention is proposed, that leads to a robust GNN-based multigrid approach, also reducing RMSE by 15\%. Additionally, a new self-supervised training method based on BERT is presented, resulting in a 25\% RMSE reduction. The paper includes an ablation study and outperforms state-of-the-art models on several challenging datasets, promising advancements similar to those recently achieved in natural language and image processing. Finally, the paper introduces a dataset with meshes larger than existing ones by at least an order of magnitude. Code and Datasets will be released at //github.com/DonsetPG/multigrid-gnn.

相關內容

神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(Neural Networks)是世(shi)界上三個(ge)(ge)最(zui)古老的(de)(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)建(jian)模學(xue)(xue)(xue)會(hui)(hui)(hui)的(de)(de)(de)(de)檔(dang)案(an)期刊:國際(ji)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡學(xue)(xue)(xue)會(hui)(hui)(hui)(INNS)、歐洲神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡學(xue)(xue)(xue)會(hui)(hui)(hui)(ENNS)和(he)日本神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡學(xue)(xue)(xue)會(hui)(hui)(hui)(JNNS)。神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡提供了一個(ge)(ge)論(lun)壇,以發展和(he)培育一個(ge)(ge)國際(ji)社會(hui)(hui)(hui)的(de)(de)(de)(de)學(xue)(xue)(xue)者(zhe)(zhe)和(he)實(shi)踐(jian)者(zhe)(zhe)感興(xing)趣的(de)(de)(de)(de)所有方面(mian)(mian)的(de)(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡和(he)相關方法的(de)(de)(de)(de)計(ji)(ji)算(suan)(suan)智(zhi)能(neng)(neng)。神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡歡(huan)迎高質(zhi)量(liang)論(lun)文的(de)(de)(de)(de)提交(jiao)(jiao),有助于(yu)全面(mian)(mian)的(de)(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡研(yan)究,從行為和(he)大(da)腦建(jian)模,學(xue)(xue)(xue)習(xi)算(suan)(suan)法,通(tong)過(guo)數學(xue)(xue)(xue)和(he)計(ji)(ji)算(suan)(suan)分析,系(xi)統的(de)(de)(de)(de)工程和(he)技術(shu)應(ying)用,大(da)量(liang)使用神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡的(de)(de)(de)(de)概念和(he)技術(shu)。這一獨特而廣泛的(de)(de)(de)(de)范圍(wei)促(cu)進(jin)了生物(wu)和(he)技術(shu)研(yan)究之間的(de)(de)(de)(de)思(si)想交(jiao)(jiao)流(liu),并(bing)有助于(yu)促(cu)進(jin)對生物(wu)啟發的(de)(de)(de)(de)計(ji)(ji)算(suan)(suan)智(zhi)能(neng)(neng)感興(xing)趣的(de)(de)(de)(de)跨學(xue)(xue)(xue)科(ke)社區的(de)(de)(de)(de)發展。因此,神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡編委會(hui)(hui)(hui)代表(biao)的(de)(de)(de)(de)專家領(ling)域包括心理(li)學(xue)(xue)(xue),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)生物(wu)學(xue)(xue)(xue),計(ji)(ji)算(suan)(suan)機科(ke)學(xue)(xue)(xue),工程,數學(xue)(xue)(xue),物(wu)理(li)。該雜志發表(biao)文章、信(xin)件(jian)(jian)和(he)評論(lun)以及給(gei)編輯的(de)(de)(de)(de)信(xin)件(jian)(jian)、社論(lun)、時事、軟件(jian)(jian)調查(cha)和(he)專利信(xin)息。文章發表(biao)在五個(ge)(ge)部分之一:認(ren)知科(ke)學(xue)(xue)(xue),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)科(ke)學(xue)(xue)(xue),學(xue)(xue)(xue)習(xi)系(xi)統,數學(xue)(xue)(xue)和(he)計(ji)(ji)算(suan)(suan)分析、工程和(he)應(ying)用。 官網(wang)地址:

Retrieval-augmented generation (RAG) has emerged as a critical mechanism in contemporary NLP to support Large Language Models(LLMs) in systematically accessing richer factual context. However, the integration of RAG mechanisms brings its inherent challenges, as LLMs need to deal with potentially noisy contexts. Recent studies have shown that LLMs still struggle to critically analyse RAG-based in-context information, a limitation that may lead to incorrect inferences and hallucinations. In this paper, we investigate how to elicit critical reasoning in RAG via contrastive explanations. In particular, we propose Contrastive-RAG (C-RAG), a framework that (i) retrieves relevant documents given a query, (ii) selects and exemplifies relevant passages, and (iii) generates explanations that explicitly contrast the relevance of the passages to (iv) support the final answer. We show the impact of C-RAG building contrastive reasoning demonstrations from LLMs to instruct smaller models for retrieval-augmented tasks. Extensive experiments demonstrate that C-RAG improves state-of-the-art RAG models while (a) requiring significantly fewer prompts and demonstrations and (b) being robust to perturbations in the retrieved documents.

Recent artificial intelligence (AI) systems have reached milestones in "grand challenges" ranging from Go to protein-folding. The capability to retrieve medical knowledge, reason over it, and answer medical questions comparably to physicians has long been viewed as one such grand challenge. Large language models (LLMs) have catalyzed significant progress in medical question answering; Med-PaLM was the first model to exceed a "passing" score in US Medical Licensing Examination (USMLE) style questions with a score of 67.2% on the MedQA dataset. However, this and other prior work suggested significant room for improvement, especially when models' answers were compared to clinicians' answers. Here we present Med-PaLM 2, which bridges these gaps by leveraging a combination of base LLM improvements (PaLM 2), medical domain finetuning, and prompting strategies including a novel ensemble refinement approach. Med-PaLM 2 scored up to 86.5% on the MedQA dataset, improving upon Med-PaLM by over 19% and setting a new state-of-the-art. We also observed performance approaching or exceeding state-of-the-art across MedMCQA, PubMedQA, and MMLU clinical topics datasets. We performed detailed human evaluations on long-form questions along multiple axes relevant to clinical applications. In pairwise comparative ranking of 1066 consumer medical questions, physicians preferred Med-PaLM 2 answers to those produced by physicians on eight of nine axes pertaining to clinical utility (p < 0.001). We also observed significant improvements compared to Med-PaLM on every evaluation axis (p < 0.001) on newly introduced datasets of 240 long-form "adversarial" questions to probe LLM limitations. While further studies are necessary to validate the efficacy of these models in real-world settings, these results highlight rapid progress towards physician-level performance in medical question answering.

Graph Neural Networks (GNNs) have shown promising results on a broad spectrum of applications. Most empirical studies of GNNs directly take the observed graph as input, assuming the observed structure perfectly depicts the accurate and complete relations between nodes. However, graphs in the real world are inevitably noisy or incomplete, which could even exacerbate the quality of graph representations. In this work, we propose a novel Variational Information Bottleneck guided Graph Structure Learning framework, namely VIB-GSL, in the perspective of information theory. VIB-GSL advances the Information Bottleneck (IB) principle for graph structure learning, providing a more elegant and universal framework for mining underlying task-relevant relations. VIB-GSL learns an informative and compressive graph structure to distill the actionable information for specific downstream tasks. VIB-GSL deduces a variational approximation for irregular graph data to form a tractable IB objective function, which facilitates training stability. Extensive experimental results demonstrate that the superior effectiveness and robustness of VIB-GSL.

Recent contrastive representation learning methods rely on estimating mutual information (MI) between multiple views of an underlying context. E.g., we can derive multiple views of a given image by applying data augmentation, or we can split a sequence into views comprising the past and future of some step in the sequence. Contrastive lower bounds on MI are easy to optimize, but have a strong underestimation bias when estimating large amounts of MI. We propose decomposing the full MI estimation problem into a sum of smaller estimation problems by splitting one of the views into progressively more informed subviews and by applying the chain rule on MI between the decomposed views. This expression contains a sum of unconditional and conditional MI terms, each measuring modest chunks of the total MI, which facilitates approximation via contrastive bounds. To maximize the sum, we formulate a contrastive lower bound on the conditional MI which can be approximated efficiently. We refer to our general approach as Decomposed Estimation of Mutual Information (DEMI). We show that DEMI can capture a larger amount of MI than standard non-decomposed contrastive bounds in a synthetic setting, and learns better representations in a vision domain and for dialogue generation.

Graph Neural Networks (GNNs) have recently become increasingly popular due to their ability to learn complex systems of relations or interactions arising in a broad spectrum of problems ranging from biology and particle physics to social networks and recommendation systems. Despite the plethora of different models for deep learning on graphs, few approaches have been proposed thus far for dealing with graphs that present some sort of dynamic nature (e.g. evolving features or connectivity over time). In this paper, we present Temporal Graph Networks (TGNs), a generic, efficient framework for deep learning on dynamic graphs represented as sequences of timed events. Thanks to a novel combination of memory modules and graph-based operators, TGNs are able to significantly outperform previous approaches being at the same time more computationally efficient. We furthermore show that several previous models for learning on dynamic graphs can be cast as specific instances of our framework. We perform a detailed ablation study of different components of our framework and devise the best configuration that achieves state-of-the-art performance on several transductive and inductive prediction tasks for dynamic graphs.

Translational distance-based knowledge graph embedding has shown progressive improvements on the link prediction task, from TransE to the latest state-of-the-art RotatE. However, N-1, 1-N and N-N predictions still remain challenging. In this work, we propose a novel translational distance-based approach for knowledge graph link prediction. The proposed method includes two-folds, first we extend the RotatE from 2D complex domain to high dimension space with orthogonal transforms to model relations for better modeling capacity. Second, the graph context is explicitly modeled via two directed context representations. These context representations are used as part of the distance scoring function to measure the plausibility of the triples during training and inference. The proposed approach effectively improves prediction accuracy on the difficult N-1, 1-N and N-N cases for knowledge graph link prediction task. The experimental results show that it achieves better performance on two benchmark data sets compared to the baseline RotatE, especially on data set (FB15k-237) with many high in-degree connection nodes.

Video captioning is a challenging task that requires a deep understanding of visual scenes. State-of-the-art methods generate captions using either scene-level or object-level information but without explicitly modeling object interactions. Thus, they often fail to make visually grounded predictions, and are sensitive to spurious correlations. In this paper, we propose a novel spatio-temporal graph model for video captioning that exploits object interactions in space and time. Our model builds interpretable links and is able to provide explicit visual grounding. To avoid unstable performance caused by the variable number of objects, we further propose an object-aware knowledge distillation mechanism, in which local object information is used to regularize global scene features. We demonstrate the efficacy of our approach through extensive experiments on two benchmarks, showing our approach yields competitive performance with interpretable predictions.

Reasoning with knowledge expressed in natural language and Knowledge Bases (KBs) is a major challenge for Artificial Intelligence, with applications in machine reading, dialogue, and question answering. General neural architectures that jointly learn representations and transformations of text are very data-inefficient, and it is hard to analyse their reasoning process. These issues are addressed by end-to-end differentiable reasoning systems such as Neural Theorem Provers (NTPs), although they can only be used with small-scale symbolic KBs. In this paper we first propose Greedy NTPs (GNTPs), an extension to NTPs addressing their complexity and scalability limitations, thus making them applicable to real-world datasets. This result is achieved by dynamically constructing the computation graph of NTPs and including only the most promising proof paths during inference, thus obtaining orders of magnitude more efficient models. Then, we propose a novel approach for jointly reasoning over KBs and textual mentions, by embedding logic facts and natural language sentences in a shared embedding space. We show that GNTPs perform on par with NTPs at a fraction of their cost while achieving competitive link prediction results on large datasets, providing explanations for predictions, and inducing interpretable models. Source code, datasets, and supplementary material are available online at //github.com/uclnlp/gntp.

Aspect level sentiment classification aims to identify the sentiment expressed towards an aspect given a context sentence. Previous neural network based methods largely ignore the syntax structure in one sentence. In this paper, we propose a novel target-dependent graph attention network (TD-GAT) for aspect level sentiment classification, which explicitly utilizes the dependency relationship among words. Using the dependency graph, it propagates sentiment features directly from the syntactic context of an aspect target. In our experiments, we show our method outperforms multiple baselines with GloVe embeddings. We also demonstrate that using BERT representations further substantially boosts the performance.

We propose a novel single shot object detection network named Detection with Enriched Semantics (DES). Our motivation is to enrich the semantics of object detection features within a typical deep detector, by a semantic segmentation branch and a global activation module. The segmentation branch is supervised by weak segmentation ground-truth, i.e., no extra annotation is required. In conjunction with that, we employ a global activation module which learns relationship between channels and object classes in a self-supervised manner. Comprehensive experimental results on both PASCAL VOC and MS COCO detection datasets demonstrate the effectiveness of the proposed method. In particular, with a VGG16 based DES, we achieve an mAP of 81.7 on VOC2007 test and an mAP of 32.8 on COCO test-dev with an inference speed of 31.5 milliseconds per image on a Titan Xp GPU. With a lower resolution version, we achieve an mAP of 79.7 on VOC2007 with an inference speed of 13.0 milliseconds per image.

北京阿比特科技有限公司