亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

The Cheyette model is a quasi-Gaussian volatility interest rate model widely used to price interest rate derivatives such as European and Bermudan Swaptions for which Monte Carlo simulation has become the industry standard. In low dimensions, these approaches provide accurate and robust prices for European Swaptions but, even in this computationally simple setting, they are known to underestimate the value of Bermudan Swaptions when using the state variables as regressors. This is mainly due to the use of a finite number of predetermined basis functions in the regression. Moreover, in high-dimensional settings, these approaches succumb to the Curse of Dimensionality. To address these issues, Deep-learning techniques have been used to solve the backward Stochastic Differential Equation associated with the value process for European and Bermudan Swaptions; however, these methods are constrained by training time and memory. To overcome these limitations, we propose leveraging Tensor Neural Networks as they can provide significant parameter savings while attaining the same accuracy as classical Dense Neural Networks. In this paper we rigorously benchmark the performance of Tensor Neural Networks and Dense Neural Networks for pricing European and Bermudan Swaptions, and we show that Tensor Neural Networks can be trained faster than Dense Neural Networks and provide more accurate and robust prices than their Dense counterparts.

相關內容

神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(Neural Networks)是世(shi)界上三個(ge)最(zui)古老的(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)建模學(xue)(xue)會(hui)的(de)(de)檔(dang)案期刊:國際神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)學(xue)(xue)會(hui)(INNS)、歐洲神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)學(xue)(xue)會(hui)(ENNS)和(he)日本(ben)神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)學(xue)(xue)會(hui)(JNNS)。神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)提供了(le)一(yi)個(ge)論(lun)壇,以發(fa)展和(he)培育一(yi)個(ge)國際社(she)會(hui)的(de)(de)學(xue)(xue)者和(he)實踐者感興(xing)趣的(de)(de)所(suo)有(you)(you)方(fang)面(mian)的(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)和(he)相關方(fang)法的(de)(de)計(ji)算(suan)智能(neng)。神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)歡(huan)迎高質量(liang)論(lun)文(wen)的(de)(de)提交,有(you)(you)助于(yu)全(quan)面(mian)的(de)(de)神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)研究,從行為和(he)大腦建模,學(xue)(xue)習(xi)(xi)算(suan)法,通過(guo)數(shu)學(xue)(xue)和(he)計(ji)算(suan)分(fen)析(xi),系(xi)統的(de)(de)工(gong)程(cheng)和(he)技術應用,大量(liang)使用神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)的(de)(de)概(gai)念和(he)技術。這一(yi)獨(du)特而廣泛(fan)的(de)(de)范圍(wei)促進了(le)生物和(he)技術研究之間(jian)的(de)(de)思想交流,并(bing)有(you)(you)助于(yu)促進對生物啟發(fa)的(de)(de)計(ji)算(suan)智能(neng)感興(xing)趣的(de)(de)跨學(xue)(xue)科社(she)區(qu)的(de)(de)發(fa)展。因此,神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)編(bian)委會(hui)代表的(de)(de)專家領域包(bao)括心理學(xue)(xue),神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)生物學(xue)(xue),計(ji)算(suan)機科學(xue)(xue),工(gong)程(cheng),數(shu)學(xue)(xue),物理。該雜(za)志(zhi)發(fa)表文(wen)章(zhang)(zhang)、信件和(he)評(ping)論(lun)以及給編(bian)輯(ji)的(de)(de)信件、社(she)論(lun)、時(shi)事、軟件調查和(he)專利(li)信息。文(wen)章(zhang)(zhang)發(fa)表在五個(ge)部分(fen)之一(yi):認(ren)知科學(xue)(xue),神(shen)(shen)(shen)經(jing)(jing)(jing)(jing)(jing)科學(xue)(xue),學(xue)(xue)習(xi)(xi)系(xi)統,數(shu)學(xue)(xue)和(he)計(ji)算(suan)分(fen)析(xi)、工(gong)程(cheng)和(he)應用。 官網(wang)(wang)(wang)地址:

The success of AI models relies on the availability of large, diverse, and high-quality datasets, which can be challenging to obtain due to data scarcity, privacy concerns, and high costs. Synthetic data has emerged as a promising solution by generating artificial data that mimics real-world patterns. This paper provides an overview of synthetic data research, discussing its applications, challenges, and future directions. We present empirical evidence from prior art to demonstrate its effectiveness and highlight the importance of ensuring its factuality, fidelity, and unbiasedness. We emphasize the need for responsible use of synthetic data to build more powerful, inclusive, and trustworthy language models.

Large Language Models (LLMs) have shown excellent generalization capabilities that have led to the development of numerous models. These models propose various new architectures, tweaking existing architectures with refined training strategies, increasing context length, using high-quality training data, and increasing training time to outperform baselines. Analyzing new developments is crucial for identifying changes that enhance training stability and improve generalization in LLMs. This survey paper comprehensively analyses the LLMs architectures and their categorization, training strategies, training datasets, and performance evaluations and discusses future research directions. Moreover, the paper also discusses the basic building blocks and concepts behind LLMs, followed by a complete overview of LLMs, including their important features and functions. Finally, the paper summarizes significant findings from LLM research and consolidates essential architectural and training strategies for developing advanced LLMs. Given the continuous advancements in LLMs, we intend to regularly update this paper by incorporating new sections and featuring the latest LLM models.

As artificial intelligence (AI) models continue to scale up, they are becoming more capable and integrated into various forms of decision-making systems. For models involved in moral decision-making, also known as artificial moral agents (AMA), interpretability provides a way to trust and understand the agent's internal reasoning mechanisms for effective use and error correction. In this paper, we provide an overview of this rapidly-evolving sub-field of AI interpretability, introduce the concept of the Minimum Level of Interpretability (MLI) and recommend an MLI for various types of agents, to aid their safe deployment in real-world settings.

When is heterogeneity in the composition of an autonomous robotic team beneficial and when is it detrimental? We investigate and answer this question in the context of a minimally viable model that examines the role of heterogeneous speeds in perimeter defense problems, where defenders share a total allocated speed budget. We consider two distinct problem settings and develop strategies based on dynamic programming and on local interaction rules. We present a theoretical analysis of both approaches and our results are extensively validated using simulations. Interestingly, our results demonstrate that the viability of heterogeneous teams depends on the amount of information available to the defenders. Moreover, our results suggest a universality property: across a wide range of problem parameters the optimal ratio of the speeds of the defenders remains nearly constant.

Causality can be described in terms of a structural causal model (SCM) that carries information on the variables of interest and their mechanistic relations. For most processes of interest the underlying SCM will only be partially observable, thus causal inference tries to leverage any exposed information. Graph neural networks (GNN) as universal approximators on structured input pose a viable candidate for causal learning, suggesting a tighter integration with SCM. To this effect we present a theoretical analysis from first principles that establishes a novel connection between GNN and SCM while providing an extended view on general neural-causal models. We then establish a new model class for GNN-based causal inference that is necessary and sufficient for causal effect identification. Our empirical illustration on simulations and standard benchmarks validate our theoretical proofs.

Residual networks (ResNets) have displayed impressive results in pattern recognition and, recently, have garnered considerable theoretical interest due to a perceived link with neural ordinary differential equations (neural ODEs). This link relies on the convergence of network weights to a smooth function as the number of layers increases. We investigate the properties of weights trained by stochastic gradient descent and their scaling with network depth through detailed numerical experiments. We observe the existence of scaling regimes markedly different from those assumed in neural ODE literature. Depending on certain features of the network architecture, such as the smoothness of the activation function, one may obtain an alternative ODE limit, a stochastic differential equation or neither of these. These findings cast doubts on the validity of the neural ODE model as an adequate asymptotic description of deep ResNets and point to an alternative class of differential equations as a better description of the deep network limit.

Multi-relation Question Answering is a challenging task, due to the requirement of elaborated analysis on questions and reasoning over multiple fact triples in knowledge base. In this paper, we present a novel model called Interpretable Reasoning Network that employs an interpretable, hop-by-hop reasoning process for question answering. The model dynamically decides which part of an input question should be analyzed at each hop; predicts a relation that corresponds to the current parsed results; utilizes the predicted relation to update the question representation and the state of the reasoning process; and then drives the next-hop reasoning. Experiments show that our model yields state-of-the-art results on two datasets. More interestingly, the model can offer traceable and observable intermediate predictions for reasoning analysis and failure diagnosis, thereby allowing manual manipulation in predicting the final answer.

Aspect based sentiment analysis (ABSA) can provide more detailed information than general sentiment analysis, because it aims to predict the sentiment polarities of the given aspects or entities in text. We summarize previous approaches into two subtasks: aspect-category sentiment analysis (ACSA) and aspect-term sentiment analysis (ATSA). Most previous approaches employ long short-term memory and attention mechanisms to predict the sentiment polarity of the concerned targets, which are often complicated and need more training time. We propose a model based on convolutional neural networks and gating mechanisms, which is more accurate and efficient. First, the novel Gated Tanh-ReLU Units can selectively output the sentiment features according to the given aspect or entity. The architecture is much simpler than attention layer used in the existing models. Second, the computations of our model could be easily parallelized during training, because convolutional layers do not have time dependency as in LSTM layers, and gating units also work independently. The experiments on SemEval datasets demonstrate the efficiency and effectiveness of our models.

Visual Question Answering (VQA) models have struggled with counting objects in natural images so far. We identify a fundamental problem due to soft attention in these models as a cause. To circumvent this problem, we propose a neural network component that allows robust counting from object proposals. Experiments on a toy task show the effectiveness of this component and we obtain state-of-the-art accuracy on the number category of the VQA v2 dataset without negatively affecting other categories, even outperforming ensemble models with our single model. On a difficult balanced pair metric, the component gives a substantial improvement in counting over a strong baseline by 6.6%.

Detecting carried objects is one of the requirements for developing systems to reason about activities involving people and objects. We present an approach to detect carried objects from a single video frame with a novel method that incorporates features from multiple scales. Initially, a foreground mask in a video frame is segmented into multi-scale superpixels. Then the human-like regions in the segmented area are identified by matching a set of extracted features from superpixels against learned features in a codebook. A carried object probability map is generated using the complement of the matching probabilities of superpixels to human-like regions and background information. A group of superpixels with high carried object probability and strong edge support is then merged to obtain the shape of the carried object. We applied our method to two challenging datasets, and results show that our method is competitive with or better than the state-of-the-art.

北京阿比特科技有限公司