亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

While the popularity of physics-informed neural networks (PINNs) is steadily rising, to this date, PINNs have not been successful in simulating multi-scale and singular perturbation problems. In this work, we present a new training paradigm referred to as "gradient boosting" (GB), which significantly enhances the performance of physics informed neural networks (PINNs). Rather than learning the solution of a given PDE using a single neural network directly, our algorithm employs a sequence of neural networks to achieve a superior outcome. This approach allows us to solve problems presenting great challenges for traditional PINNs. Our numerical experiments demonstrate the effectiveness of our algorithm through various benchmarks, including comparisons with finite element methods and PINNs. Furthermore, this work also unlocks the door to employing ensemble learning techniques in PINNs, providing opportunities for further improvement in solving PDEs.

相關內容

神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)(Neural Networks)是世界上(shang)三個(ge)(ge)最古(gu)老的(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)建(jian)模學(xue)(xue)(xue)(xue)(xue)(xue)(xue)會(hui)的(de)(de)(de)檔案期刊:國(guo)際神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)學(xue)(xue)(xue)(xue)(xue)(xue)(xue)會(hui)(INNS)、歐洲(zhou)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)學(xue)(xue)(xue)(xue)(xue)(xue)(xue)會(hui)(ENNS)和(he)(he)日本神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)學(xue)(xue)(xue)(xue)(xue)(xue)(xue)會(hui)(JNNS)。神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)提供(gong)了一個(ge)(ge)論(lun)(lun)(lun)壇,以(yi)發(fa)(fa)(fa)展和(he)(he)培育一個(ge)(ge)國(guo)際社(she)會(hui)的(de)(de)(de)學(xue)(xue)(xue)(xue)(xue)(xue)(xue)者和(he)(he)實(shi)踐(jian)者感(gan)(gan)興趣的(de)(de)(de)所有(you)方面的(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)和(he)(he)相關方法的(de)(de)(de)計(ji)算(suan)智能。神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)歡迎(ying)高(gao)質量(liang)論(lun)(lun)(lun)文(wen)的(de)(de)(de)提交,有(you)助于全面的(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)研究,從行為和(he)(he)大腦建(jian)模,學(xue)(xue)(xue)(xue)(xue)(xue)(xue)習算(suan)法,通過數(shu)學(xue)(xue)(xue)(xue)(xue)(xue)(xue)和(he)(he)計(ji)算(suan)分(fen)(fen)析,系統(tong)的(de)(de)(de)工程和(he)(he)技術(shu)應(ying)用,大量(liang)使用神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)的(de)(de)(de)概(gai)念和(he)(he)技術(shu)。這一獨特而(er)廣泛的(de)(de)(de)范(fan)圍促(cu)進(jin)了生(sheng)物(wu)和(he)(he)技術(shu)研究之間的(de)(de)(de)思想(xiang)交流,并有(you)助于促(cu)進(jin)對生(sheng)物(wu)啟發(fa)(fa)(fa)的(de)(de)(de)計(ji)算(suan)智能感(gan)(gan)興趣的(de)(de)(de)跨學(xue)(xue)(xue)(xue)(xue)(xue)(xue)科(ke)社(she)區(qu)的(de)(de)(de)發(fa)(fa)(fa)展。因此,神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)絡(luo)(luo)編委會(hui)代表(biao)的(de)(de)(de)專家領(ling)域(yu)包括(kuo)心理學(xue)(xue)(xue)(xue)(xue)(xue)(xue),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)生(sheng)物(wu)學(xue)(xue)(xue)(xue)(xue)(xue)(xue),計(ji)算(suan)機(ji)科(ke)學(xue)(xue)(xue)(xue)(xue)(xue)(xue),工程,數(shu)學(xue)(xue)(xue)(xue)(xue)(xue)(xue),物(wu)理。該雜志(zhi)發(fa)(fa)(fa)表(biao)文(wen)章、信(xin)件和(he)(he)評論(lun)(lun)(lun)以(yi)及給編輯的(de)(de)(de)信(xin)件、社(she)論(lun)(lun)(lun)、時事、軟件調查和(he)(he)專利信(xin)息。文(wen)章發(fa)(fa)(fa)表(biao)在五(wu)個(ge)(ge)部(bu)分(fen)(fen)之一:認知(zhi)科(ke)學(xue)(xue)(xue)(xue)(xue)(xue)(xue),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)科(ke)學(xue)(xue)(xue)(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)(xue)(xue)(xue)習系統(tong),數(shu)學(xue)(xue)(xue)(xue)(xue)(xue)(xue)和(he)(he)計(ji)算(suan)分(fen)(fen)析、工程和(he)(he)應(ying)用。 官網(wang)地(di)址:

Despite the advancement of machine learning techniques in recent years, state-of-the-art systems lack robustness to "real world" events, where the input distributions and tasks encountered by the deployed systems will not be limited to the original training context, and systems will instead need to adapt to novel distributions and tasks while deployed. This critical gap may be addressed through the development of "Lifelong Learning" systems that are capable of 1) Continuous Learning, 2) Transfer and Adaptation, and 3) Scalability. Unfortunately, efforts to improve these capabilities are typically treated as distinct areas of research that are assessed independently, without regard to the impact of each separate capability on other aspects of the system. We instead propose a holistic approach, using a suite of metrics and an evaluation framework to assess Lifelong Learning in a principled way that is agnostic to specific domains or system techniques. Through five case studies, we show that this suite of metrics can inform the development of varied and complex Lifelong Learning systems. We highlight how the proposed suite of metrics quantifies performance trade-offs present during Lifelong Learning system development - both the widely discussed Stability-Plasticity dilemma and the newly proposed relationship between Sample Efficient and Robust Learning. Further, we make recommendations for the formulation and use of metrics to guide the continuing development of Lifelong Learning systems and assess their progress in the future.

Recent advances of data-driven machine learning have revolutionized fields like computer vision, reinforcement learning, and many scientific and engineering domains. In many real-world and scientific problems, systems that generate data are governed by physical laws. Recent work shows that it provides potential benefits for machine learning models by incorporating the physical prior and collected data, which makes the intersection of machine learning and physics become a prevailing paradigm. In this survey, we present this learning paradigm called Physics-Informed Machine Learning (PIML) which is to build a model that leverages empirical data and available physical prior knowledge to improve performance on a set of tasks that involve a physical mechanism. We systematically review the recent development of physics-informed machine learning from three perspectives of machine learning tasks, representation of physical prior, and methods for incorporating physical prior. We also propose several important open research problems based on the current trends in the field. We argue that encoding different forms of physical prior into model architectures, optimizers, inference algorithms, and significant domain-specific applications like inverse engineering design and robotic control is far from fully being explored in the field of physics-informed machine learning. We believe that this study will encourage researchers in the machine learning community to actively participate in the interdisciplinary research of physics-informed machine learning.

Recently, graph neural networks have been gaining a lot of attention to simulate dynamical systems due to their inductive nature leading to zero-shot generalizability. Similarly, physics-informed inductive biases in deep-learning frameworks have been shown to give superior performance in learning the dynamics of physical systems. There is a growing volume of literature that attempts to combine these two approaches. Here, we evaluate the performance of thirteen different graph neural networks, namely, Hamiltonian and Lagrangian graph neural networks, graph neural ODE, and their variants with explicit constraints and different architectures. We briefly explain the theoretical formulation highlighting the similarities and differences in the inductive biases and graph architecture of these systems. We evaluate these models on spring, pendulum, gravitational, and 3D deformable solid systems to compare the performance in terms of rollout error, conserved quantities such as energy and momentum, and generalizability to unseen system sizes. Our study demonstrates that GNNs with additional inductive biases, such as explicit constraints and decoupling of kinetic and potential energies, exhibit significantly enhanced performance. Further, all the physics-informed GNNs exhibit zero-shot generalizability to system sizes an order of magnitude larger than the training system, thus providing a promising route to simulate large-scale realistic systems.

The adaptive processing of structured data is a long-standing research topic in machine learning that investigates how to automatically learn a mapping from a structured input to outputs of various nature. Recently, there has been an increasing interest in the adaptive processing of graphs, which led to the development of different neural network-based methodologies. In this thesis, we take a different route and develop a Bayesian Deep Learning framework for graph learning. The dissertation begins with a review of the principles over which most of the methods in the field are built, followed by a study on graph classification reproducibility issues. We then proceed to bridge the basic ideas of deep learning for graphs with the Bayesian world, by building our deep architectures in an incremental fashion. This framework allows us to consider graphs with discrete and continuous edge features, producing unsupervised embeddings rich enough to reach the state of the art on several classification tasks. Our approach is also amenable to a Bayesian nonparametric extension that automatizes the choice of almost all model's hyper-parameters. Two real-world applications demonstrate the efficacy of deep learning for graphs. The first concerns the prediction of information-theoretic quantities for molecular simulations with supervised neural models. After that, we exploit our Bayesian models to solve a malware-classification task while being robust to intra-procedural code obfuscation techniques. We conclude the dissertation with an attempt to blend the best of the neural and Bayesian worlds together. The resulting hybrid model is able to predict multimodal distributions conditioned on input graphs, with the consequent ability to model stochasticity and uncertainty better than most works. Overall, we aim to provide a Bayesian perspective into the articulated research field of deep learning for graphs.

We hypothesize that due to the greedy nature of learning in multi-modal deep neural networks, these models tend to rely on just one modality while under-fitting the other modalities. Such behavior is counter-intuitive and hurts the models' generalization, as we observe empirically. To estimate the model's dependence on each modality, we compute the gain on the accuracy when the model has access to it in addition to another modality. We refer to this gain as the conditional utilization rate. In the experiments, we consistently observe an imbalance in conditional utilization rates between modalities, across multiple tasks and architectures. Since conditional utilization rate cannot be computed efficiently during training, we introduce a proxy for it based on the pace at which the model learns from each modality, which we refer to as the conditional learning speed. We propose an algorithm to balance the conditional learning speeds between modalities during training and demonstrate that it indeed addresses the issue of greedy learning. The proposed algorithm improves the model's generalization on three datasets: Colored MNIST, Princeton ModelNet40, and NVIDIA Dynamic Hand Gesture.

The chronological order of user-item interactions can reveal time-evolving and sequential user behaviors in many recommender systems. The items that users will interact with may depend on the items accessed in the past. However, the substantial increase of users and items makes sequential recommender systems still face non-trivial challenges: (1) the hardness of modeling the short-term user interests; (2) the difficulty of capturing the long-term user interests; (3) the effective modeling of item co-occurrence patterns. To tackle these challenges, we propose a memory augmented graph neural network (MA-GNN) to capture both the long- and short-term user interests. Specifically, we apply a graph neural network to model the item contextual information within a short-term period and utilize a shared memory network to capture the long-range dependencies between items. In addition to the modeling of user interests, we employ a bilinear function to capture the co-occurrence patterns of related items. We extensively evaluate our model on five real-world datasets, comparing with several state-of-the-art methods and using a variety of performance metrics. The experimental results demonstrate the effectiveness of our model for the task of Top-K sequential recommendation.

Meta-learning extracts the common knowledge acquired from learning different tasks and uses it for unseen tasks. It demonstrates a clear advantage on tasks that have insufficient training data, e.g., few-shot learning. In most meta-learning methods, tasks are implicitly related via the shared model or optimizer. In this paper, we show that a meta-learner that explicitly relates tasks on a graph describing the relations of their output dimensions (e.g., classes) can significantly improve the performance of few-shot learning. This type of graph is usually free or cheap to obtain but has rarely been explored in previous works. We study the prototype based few-shot classification, in which a prototype is generated for each class, such that the nearest neighbor search between the prototypes produces an accurate classification. We introduce "Gated Propagation Network (GPN)", which learns to propagate messages between prototypes of different classes on the graph, so that learning the prototype of each class benefits from the data of other related classes. In GPN, an attention mechanism is used for the aggregation of messages from neighboring classes, and a gate is deployed to choose between the aggregated messages and the message from the class itself. GPN is trained on a sequence of tasks from many-shot to few-shot generated by subgraph sampling. During training, it is able to reuse and update previously achieved prototypes from the memory in a life-long learning cycle. In experiments, we change the training-test discrepancy and test task generation settings for thorough evaluations. GPN outperforms recent meta-learning methods on two benchmark datasets in all studied cases.

Graph neural networks (GNNs) are a popular class of machine learning models whose major advantage is their ability to incorporate a sparse and discrete dependency structure between data points. Unfortunately, GNNs can only be used when such a graph-structure is available. In practice, however, real-world graphs are often noisy and incomplete or might not be available at all. With this work, we propose to jointly learn the graph structure and the parameters of graph convolutional networks (GCNs) by approximately solving a bilevel program that learns a discrete probability distribution on the edges of the graph. This allows one to apply GCNs not only in scenarios where the given graph is incomplete or corrupted but also in those where a graph is not available. We conduct a series of experiments that analyze the behavior of the proposed method and demonstrate that it outperforms related methods by a significant margin.

Humans and animals have the ability to continually acquire, fine-tune, and transfer knowledge and skills throughout their lifespan. This ability, referred to as lifelong learning, is mediated by a rich set of neurocognitive mechanisms that together contribute to the development and specialization of our sensorimotor skills as well as to long-term memory consolidation and retrieval. Consequently, lifelong learning capabilities are crucial for autonomous agents interacting in the real world and processing continuous streams of information. However, lifelong learning remains a long-standing challenge for machine learning and neural network models since the continual acquisition of incrementally available information from non-stationary data distributions generally leads to catastrophic forgetting or interference. This limitation represents a major drawback for state-of-the-art deep neural network models that typically learn representations from stationary batches of training data, thus without accounting for situations in which information becomes incrementally available over time. In this review, we critically summarize the main challenges linked to lifelong learning for artificial learning systems and compare existing neural network approaches that alleviate, to different extents, catastrophic forgetting. We discuss well-established and emerging research motivated by lifelong learning factors in biological systems such as structural plasticity, memory replay, curriculum and transfer learning, intrinsic motivation, and multisensory integration.

Recently, ensemble has been applied to deep metric learning to yield state-of-the-art results. Deep metric learning aims to learn deep neural networks for feature embeddings, distances of which satisfy given constraint. In deep metric learning, ensemble takes average of distances learned by multiple learners. As one important aspect of ensemble, the learners should be diverse in their feature embeddings. To this end, we propose an attention-based ensemble, which uses multiple attention masks, so that each learner can attend to different parts of the object. We also propose a divergence loss, which encourages diversity among the learners. The proposed method is applied to the standard benchmarks of deep metric learning and experimental results show that it outperforms the state-of-the-art methods by a significant margin on image retrieval tasks.

北京阿比特科技有限公司