亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Deep neural networks (DNNs) have been widely used to solve partial differential equations (PDEs) in recent years. In this work, a novel deep learning-based framework named Particle Weak-form based Neural Networks (ParticleWNN) is developed for solving PDEs in the weak form. In this framework, the trial space is defined as the space of DNNs, while the test space consists of functions compactly supported in extremely small regions, centered around particles. To facilitate the training of neural networks, an R-adaptive strategy is designed to adaptively modify the radius of regions during training. The ParticleWNN inherits the benefits of weak/variational formulation, requiring less regularity of the solution and a small number of quadrature points for computing integrals. Additionally, due to the special construction of the test functions, ParticleWNN enables parallel implementation and integral calculations only in extremely small regions. This framework is particularly desirable for solving problems with high-dimensional and complex domains. The efficiency and accuracy of ParticleWNN are demonstrated through several numerical examples, showcasing its superiority over state-of-the-art methods. The source code for the numerical examples presented in this paper is available at //github.com/yaohua32/ParticleWNN.

相關內容

神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(Neural Networks)是(shi)世界上三(san)個最古老的(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)建模(mo)學(xue)(xue)(xue)(xue)(xue)會(hui)的(de)(de)(de)檔案期刊(kan):國際(ji)(ji)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)學(xue)(xue)(xue)(xue)(xue)會(hui)(INNS)、歐洲神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)學(xue)(xue)(xue)(xue)(xue)會(hui)(ENNS)和(he)(he)(he)日本神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)學(xue)(xue)(xue)(xue)(xue)會(hui)(JNNS)。神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)提(ti)供了一個論(lun)壇,以發(fa)展和(he)(he)(he)培育一個國際(ji)(ji)社(she)(she)會(hui)的(de)(de)(de)學(xue)(xue)(xue)(xue)(xue)者和(he)(he)(he)實踐者感興趣的(de)(de)(de)所有(you)方面的(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)和(he)(he)(he)相關方法(fa)的(de)(de)(de)計(ji)(ji)算(suan)智(zhi)能。神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)歡迎(ying)高(gao)質量(liang)論(lun)文(wen)的(de)(de)(de)提(ti)交,有(you)助于(yu)全面的(de)(de)(de)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)研(yan)究,從行為和(he)(he)(he)大腦建模(mo),學(xue)(xue)(xue)(xue)(xue)習(xi)算(suan)法(fa),通過數學(xue)(xue)(xue)(xue)(xue)和(he)(he)(he)計(ji)(ji)算(suan)分析(xi),系(xi)統的(de)(de)(de)工程(cheng)和(he)(he)(he)技術應(ying)用(yong),大量(liang)使用(yong)神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)的(de)(de)(de)概念(nian)和(he)(he)(he)技術。這一獨(du)特而廣泛的(de)(de)(de)范圍促進了生物(wu)和(he)(he)(he)技術研(yan)究之(zhi)間(jian)的(de)(de)(de)思想交流,并有(you)助于(yu)促進對生物(wu)啟(qi)發(fa)的(de)(de)(de)計(ji)(ji)算(suan)智(zhi)能感興趣的(de)(de)(de)跨(kua)學(xue)(xue)(xue)(xue)(xue)科(ke)社(she)(she)區的(de)(de)(de)發(fa)展。因(yin)此,神(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)編委會(hui)代表(biao)的(de)(de)(de)專家領域包括心理學(xue)(xue)(xue)(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)(jing)生物(wu)學(xue)(xue)(xue)(xue)(xue),計(ji)(ji)算(suan)機科(ke)學(xue)(xue)(xue)(xue)(xue),工程(cheng),數學(xue)(xue)(xue)(xue)(xue),物(wu)理。該雜志發(fa)表(biao)文(wen)章(zhang)、信(xin)件和(he)(he)(he)評論(lun)以及給編輯(ji)的(de)(de)(de)信(xin)件、社(she)(she)論(lun)、時(shi)事、軟件調查和(he)(he)(he)專利信(xin)息。文(wen)章(zhang)發(fa)表(biao)在(zai)五個部分之(zhi)一:認知科(ke)學(xue)(xue)(xue)(xue)(xue),神(shen)(shen)經(jing)(jing)(jing)(jing)科(ke)學(xue)(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)(xue)習(xi)系(xi)統,數學(xue)(xue)(xue)(xue)(xue)和(he)(he)(he)計(ji)(ji)算(suan)分析(xi)、工程(cheng)和(he)(he)(he)應(ying)用(yong)。 官網(wang)(wang)(wang)地址:

Humans generally acquire new skills without compromising the old; however, the opposite holds for Large Language Models (LLMs), e.g., from LLaMA to CodeLLaMA. To this end, we propose a new post-pretraining method for LLMs with an expansion of Transformer blocks. We tune the expanded blocks using only new corpus, efficiently and effectively improving the model's knowledge without catastrophic forgetting. In this paper, we experiment on the corpus of code and math, yielding LLaMA Pro-8.3B, a versatile foundation model initialized from LLaMA2-7B, excelling in general tasks, programming, and mathematics. LLaMA Pro and its instruction-following counterpart (LLaMA Pro-Instruct) achieve advanced performance among various benchmarks, demonstrating superiority over existing open models in the LLaMA family and the immense potential of reasoning and addressing diverse tasks as an intelligent agent. Our findings provide valuable insights into integrating natural and programming languages, laying a solid foundation for developing advanced language agents that operate effectively in various environments.

Transformer is a promising neural network learner, and has achieved great success in various machine learning tasks. Thanks to the recent prevalence of multimodal applications and big data, Transformer-based multimodal learning has become a hot topic in AI research. This paper presents a comprehensive survey of Transformer techniques oriented at multimodal data. The main contents of this survey include: (1) a background of multimodal learning, Transformer ecosystem, and the multimodal big data era, (2) a theoretical review of Vanilla Transformer, Vision Transformer, and multimodal Transformers, from a geometrically topological perspective, (3) a review of multimodal Transformer applications, via two important paradigms, i.e., for multimodal pretraining and for specific multimodal tasks, (4) a summary of the common challenges and designs shared by the multimodal Transformer models and applications, and (5) a discussion of open problems and potential research directions for the community.

Graph neural networks (GNNs) have been a hot spot of recent research and are widely utilized in diverse applications. However, with the use of huger data and deeper models, an urgent demand is unsurprisingly made to accelerate GNNs for more efficient execution. In this paper, we provide a comprehensive survey on acceleration methods for GNNs from an algorithmic perspective. We first present a new taxonomy to classify existing acceleration methods into five categories. Based on the classification, we systematically discuss these methods and highlight their correlations. Next, we provide comparisons from aspects of the efficiency and characteristics of these methods. Finally, we suggest some promising prospects for future research.

Recently many efforts have been devoted to applying graph neural networks (GNNs) to molecular property prediction which is a fundamental task for computational drug and material discovery. One of major obstacles to hinder the successful prediction of molecule property by GNNs is the scarcity of labeled data. Though graph contrastive learning (GCL) methods have achieved extraordinary performance with insufficient labeled data, most focused on designing data augmentation schemes for general graphs. However, the fundamental property of a molecule could be altered with the augmentation method (like random perturbation) on molecular graphs. Whereas, the critical geometric information of molecules remains rarely explored under the current GNN and GCL architectures. To this end, we propose a novel graph contrastive learning method utilizing the geometry of the molecule across 2D and 3D views, which is named GeomGCL. Specifically, we first devise a dual-view geometric message passing network (GeomMPNN) to adaptively leverage the rich information of both 2D and 3D graphs of a molecule. The incorporation of geometric properties at different levels can greatly facilitate the molecular representation learning. Then a novel geometric graph contrastive scheme is designed to make both geometric views collaboratively supervise each other to improve the generalization ability of GeomMPNN. We evaluate GeomGCL on various downstream property prediction tasks via a finetune process. Experimental results on seven real-life molecular datasets demonstrate the effectiveness of our proposed GeomGCL against state-of-the-art baselines.

Deep neural networks (DNNs) have become a proven and indispensable machine learning tool. As a black-box model, it remains difficult to diagnose what aspects of the model's input drive the decisions of a DNN. In countless real-world domains, from legislation and law enforcement to healthcare, such diagnosis is essential to ensure that DNN decisions are driven by aspects appropriate in the context of its use. The development of methods and studies enabling the explanation of a DNN's decisions has thus blossomed into an active, broad area of research. A practitioner wanting to study explainable deep learning may be intimidated by the plethora of orthogonal directions the field has taken. This complexity is further exacerbated by competing definitions of what it means ``to explain'' the actions of a DNN and to evaluate an approach's ``ability to explain''. This article offers a field guide to explore the space of explainable deep learning aimed at those uninitiated in the field. The field guide: i) Introduces three simple dimensions defining the space of foundational methods that contribute to explainable deep learning, ii) discusses the evaluations for model explanations, iii) places explainability in the context of other related deep learning research areas, and iv) finally elaborates on user-oriented explanation designing and potential future directions on explainable deep learning. We hope the guide is used as an easy-to-digest starting point for those just embarking on research in this field.

It has been shown that deep neural networks are prone to overfitting on biased training data. Towards addressing this issue, meta-learning employs a meta model for correcting the training bias. Despite the promising performances, super slow training is currently the bottleneck in the meta learning approaches. In this paper, we introduce a novel Faster Meta Update Strategy (FaMUS) to replace the most expensive step in the meta gradient computation with a faster layer-wise approximation. We empirically find that FaMUS yields not only a reasonably accurate but also a low-variance approximation of the meta gradient. We conduct extensive experiments to verify the proposed method on two tasks. We show our method is able to save two-thirds of the training time while still maintaining the comparable or achieving even better generalization performance. In particular, our method achieves the state-of-the-art performance on both synthetic and realistic noisy labels, and obtains promising performance on long-tailed recognition on standard benchmarks.

Deep convolutional neural networks (CNNs) have recently achieved great success in many visual recognition tasks. However, existing deep neural network models are computationally expensive and memory intensive, hindering their deployment in devices with low memory resources or in applications with strict latency requirements. Therefore, a natural thought is to perform model compression and acceleration in deep networks without significantly decreasing the model performance. During the past few years, tremendous progress has been made in this area. In this paper, we survey the recent advanced techniques for compacting and accelerating CNNs model developed. These techniques are roughly categorized into four schemes: parameter pruning and sharing, low-rank factorization, transferred/compact convolutional filters, and knowledge distillation. Methods of parameter pruning and sharing will be described at the beginning, after that the other techniques will be introduced. For each scheme, we provide insightful analysis regarding the performance, related applications, advantages, and drawbacks etc. Then we will go through a few very recent additional successful methods, for example, dynamic capacity networks and stochastic depths networks. After that, we survey the evaluation matrix, the main datasets used for evaluating the model performance and recent benchmarking efforts. Finally, we conclude this paper, discuss remaining challenges and possible directions on this topic.

Graph convolutional networks (GCNs) have recently become one of the most powerful tools for graph analytics tasks in numerous applications, ranging from social networks and natural language processing to bioinformatics and chemoinformatics, thanks to their ability to capture the complex relationships between concepts. At present, the vast majority of GCNs use a neighborhood aggregation framework to learn a continuous and compact vector, then performing a pooling operation to generalize graph embedding for the classification task. These approaches have two disadvantages in the graph classification task: (1)when only the largest sub-graph structure ($k$-hop neighbor) is used for neighborhood aggregation, a large amount of early-stage information is lost during the graph convolution step; (2) simple average/sum pooling or max pooling utilized, which loses the characteristics of each node and the topology between nodes. In this paper, we propose a novel framework called, dual attention graph convolutional networks (DAGCN) to address these problems. DAGCN automatically learns the importance of neighbors at different hops using a novel attention graph convolution layer, and then employs a second attention component, a self-attention pooling layer, to generalize the graph representation from the various aspects of a matrix graph embedding. The dual attention network is trained in an end-to-end manner for the graph classification task. We compare our model with state-of-the-art graph kernels and other deep learning methods. The experimental results show that our framework not only outperforms other baselines but also achieves a better rate of convergence.

Convolutional networks (ConvNets) have achieved great successes in various challenging vision tasks. However, the performance of ConvNets would degrade when encountering the domain shift. The domain adaptation is more significant while challenging in the field of biomedical image analysis, where cross-modality data have largely different distributions. Given that annotating the medical data is especially expensive, the supervised transfer learning approaches are not quite optimal. In this paper, we propose an unsupervised domain adaptation framework with adversarial learning for cross-modality biomedical image segmentations. Specifically, our model is based on a dilated fully convolutional network for pixel-wise prediction. Moreover, we build a plug-and-play domain adaptation module (DAM) to map the target input to features which are aligned with source domain feature space. A domain critic module (DCM) is set up for discriminating the feature space of both domains. We optimize the DAM and DCM via an adversarial loss without using any target domain label. Our proposed method is validated by adapting a ConvNet trained with MRI images to unpaired CT data for cardiac structures segmentations, and achieved very promising results.

ASR (automatic speech recognition) systems like Siri, Alexa, Google Voice or Cortana has become quite popular recently. One of the key techniques enabling the practical use of such systems in people's daily life is deep learning. Though deep learning in computer vision is known to be vulnerable to adversarial perturbations, little is known whether such perturbations are still valid on the practical speech recognition. In this paper, we not only demonstrate such attacks can happen in reality, but also show that the attacks can be systematically conducted. To minimize users' attention, we choose to embed the voice commands into a song, called CommandSong. In this way, the song carrying the command can spread through radio, TV or even any media player installed in the portable devices like smartphones, potentially impacting millions of users in long distance. In particular, we overcome two major challenges: minimizing the revision of a song in the process of embedding commands, and letting the CommandSong spread through the air without losing the voice "command". Our evaluation demonstrates that we can craft random songs to "carry" any commands and the modify is extremely difficult to be noticed. Specially, the physical attack that we play the CommandSongs over the air and record them can success with 94 percentage.

北京阿比特科技有限公司