亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Designing an incentive compatible auction that maximizes expected revenue is a central problem in Auction Design. Theoretical approaches to the problem have hit some limits in the past decades and analytical solutions are known for only a few simple settings. Computational approaches to the problem through the use of LPs have their own set of limitations. Building on the success of deep learning, a new approach was recently proposed by Duetting et al. (2019) in which the auction is modeled by a feed-forward neural network and the design problem is framed as a learning problem. The neural architectures used in that work are general purpose and do not take advantage of any of the symmetries the problem could present, such as permutation equivariance. In this work, we consider auction design problems that have permutation-equivariant symmetry and construct a neural architecture that is capable of perfectly recovering the permutation-equivariant optimal mechanism, which we show is not possible with the previous architecture. We demonstrate that permutation-equivariant architectures are not only capable of recovering previous results, they also have better generalization properties.

相關內容

神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(Neural Networks)是世界上(shang)三個最古(gu)老的(de)(de)(de)神(shen)(shen)(shen)經(jing)建(jian)模(mo)學(xue)(xue)會(hui)(hui)的(de)(de)(de)檔案(an)期刊:國際(ji)神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(hui)(INNS)、歐洲神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(hui)(ENNS)和(he)(he)日本神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)學(xue)(xue)會(hui)(hui)(JNNS)。神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)提(ti)供了(le)一個論壇,以(yi)發(fa)展(zhan)(zhan)和(he)(he)培育一個國際(ji)社會(hui)(hui)的(de)(de)(de)學(xue)(xue)者(zhe)和(he)(he)實踐(jian)者(zhe)感興(xing)趣(qu)的(de)(de)(de)所有(you)方(fang)面的(de)(de)(de)神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)和(he)(he)相關方(fang)法的(de)(de)(de)計(ji)算智(zhi)能(neng)。神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)歡迎高質量論文的(de)(de)(de)提(ti)交,有(you)助(zhu)于(yu)全面的(de)(de)(de)神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)研究(jiu),從行為和(he)(he)大腦建(jian)模(mo),學(xue)(xue)習算法,通過數(shu)(shu)學(xue)(xue)和(he)(he)計(ji)算分(fen)(fen)析(xi)(xi),系統的(de)(de)(de)工(gong)程(cheng)和(he)(he)技術(shu)應用(yong)(yong),大量使(shi)用(yong)(yong)神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)的(de)(de)(de)概(gai)念和(he)(he)技術(shu)。這一獨特而(er)廣泛的(de)(de)(de)范圍促(cu)進了(le)生物(wu)(wu)和(he)(he)技術(shu)研究(jiu)之間的(de)(de)(de)思(si)想交流,并有(you)助(zhu)于(yu)促(cu)進對生物(wu)(wu)啟(qi)發(fa)的(de)(de)(de)計(ji)算智(zhi)能(neng)感興(xing)趣(qu)的(de)(de)(de)跨(kua)學(xue)(xue)科(ke)社區的(de)(de)(de)發(fa)展(zhan)(zhan)。因此,神(shen)(shen)(shen)經(jing)網(wang)(wang)(wang)(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)編委會(hui)(hui)代表的(de)(de)(de)專(zhuan)家領(ling)域包括心理(li)(li)學(xue)(xue),神(shen)(shen)(shen)經(jing)生物(wu)(wu)學(xue)(xue),計(ji)算機科(ke)學(xue)(xue),工(gong)程(cheng),數(shu)(shu)學(xue)(xue),物(wu)(wu)理(li)(li)。該雜志發(fa)表文章(zhang)、信(xin)(xin)件和(he)(he)評論以(yi)及給編輯的(de)(de)(de)信(xin)(xin)件、社論、時事、軟件調查和(he)(he)專(zhuan)利信(xin)(xin)息。文章(zhang)發(fa)表在五(wu)個部(bu)分(fen)(fen)之一:認知科(ke)學(xue)(xue),神(shen)(shen)(shen)經(jing)科(ke)學(xue)(xue),學(xue)(xue)習系統,數(shu)(shu)學(xue)(xue)和(he)(he)計(ji)算分(fen)(fen)析(xi)(xi)、工(gong)程(cheng)和(he)(he)應用(yong)(yong)。 官網(wang)(wang)(wang)(wang)(wang)(wang)地址(zhi):

Reinforcement learning is generally difficult for partially observable Markov decision processes (POMDPs), which occurs when the agent's observation is partial or noisy. To seek good performance in POMDPs, one strategy is to endow the agent with a finite memory, whose update is governed by the policy. However, policy optimization is non-convex in that case and can lead to poor training performance for random initialization. The performance can be empirically improved by constraining the memory architecture, then sacrificing optimality to facilitate training. Here we study this trade-off in a two-hypothesis testing problem, akin to the two-arm bandit problem. We compare two extreme cases: (i) the random access memory where any transitions between $M$ memory states are allowed and (ii) a fixed memory where the agent can access its last $m$ actions and rewards. For (i), the probability $q$ to play the worst arm is known to be exponentially small in $M$ for the optimal policy. Our main result is to show that similar performance can be reached for (ii) as well, despite the simplicity of the memory architecture: using a conjecture on Gray-ordered binary necklaces, we find policies for which $q$ is exponentially small in $2^m$, i.e. $q\sim\alpha^{2^m}$ with $\alpha < 1$. In addition, we observe empirically that training from random initialization leads to very poor results for (i), and significantly better results for (ii) thanks to the constraints on the memory architecture.

In real-world applications, data often come in a growing manner, where the data volume and the number of classes may increase dynamically. This will bring a critical challenge for learning: given the increasing data volume or the number of classes, one has to instantaneously adjust the neural model capacity to obtain promising performance. Existing methods either ignore the growing nature of data or seek to independently search an optimal architecture for a given dataset, and thus are incapable of promptly adjusting the architectures for the changed data. To address this, we present a neural architecture adaptation method, namely Adaptation eXpert (AdaXpert), to efficiently adjust previous architectures on the growing data. Specifically, we introduce an architecture adjuster to generate a suitable architecture for each data snapshot, based on the previous architecture and the different extent between current and previous data distributions. Furthermore, we propose an adaptation condition to determine the necessity of adjustment, thereby avoiding unnecessary and time-consuming adjustments. Extensive experiments on two growth scenarios (increasing data volume and number of classes) demonstrate the effectiveness of the proposed method.

One of the key steps in Neural Architecture Search (NAS) is to estimate the performance of candidate architectures. Existing methods either directly use the validation performance or learn a predictor to estimate the performance. However, these methods can be either computationally expensive or very inaccurate, which may severely affect the search efficiency and performance. Moreover, as it is very difficult to annotate architectures with accurate performance on specific tasks, learning a promising performance predictor is often non-trivial due to the lack of labeled data. In this paper, we argue that it may not be necessary to estimate the absolute performance for NAS. On the contrary, we may need only to understand whether an architecture is better than a baseline one. However, how to exploit this comparison information as the reward and how to well use the limited labeled data remains two great challenges. In this paper, we propose a novel Contrastive Neural Architecture Search (CTNAS) method which performs architecture search by taking the comparison results between architectures as the reward. Specifically, we design and learn a Neural Architecture Comparator (NAC) to compute the probability of candidate architectures being better than a baseline one. Moreover, we present a baseline updating scheme to improve the baseline iteratively in a curriculum learning manner. More critically, we theoretically show that learning NAC is equivalent to optimizing the ranking over architectures. Extensive experiments in three search spaces demonstrate the superiority of our CTNAS over existing methods.

The essence of multivariate sequential learning is all about how to extract dependencies in data. These data sets, such as hourly medical records in intensive care units and multi-frequency phonetic time series, often time exhibit not only strong serial dependencies in the individual components (the "marginal" memory) but also non-negligible memories in the cross-sectional dependencies (the "joint" memory). Because of the multivariate complexity in the evolution of the joint distribution that underlies the data generating process, we take a data-driven approach and construct a novel recurrent network architecture, termed Memory-Gated Recurrent Networks (mGRN), with gates explicitly regulating two distinct types of memories: the marginal memory and the joint memory. Through a combination of comprehensive simulation studies and empirical experiments on a range of public datasets, we show that our proposed mGRN architecture consistently outperforms state-of-the-art architectures targeting multivariate time series.

Graph neural networks (GNN) has been successfully applied to operate on the graph-structured data. Given a specific scenario, rich human expertise and tremendous laborious trials are usually required to identify a suitable GNN architecture. It is because the performance of a GNN architecture is significantly affected by the choice of graph convolution components, such as aggregate function and hidden dimension. Neural architecture search (NAS) has shown its potential in discovering effective deep architectures for learning tasks in image and language modeling. However, existing NAS algorithms cannot be directly applied to the GNN search problem. First, the search space of GNN is different from the ones in existing NAS work. Second, the representation learning capacity of GNN architecture changes obviously with slight architecture modifications. It affects the search efficiency of traditional search methods. Third, widely used techniques in NAS such as parameter sharing might become unstable in GNN. To bridge the gap, we propose the automated graph neural networks (AGNN) framework, which aims to find an optimal GNN architecture within a predefined search space. A reinforcement learning based controller is designed to greedily validate architectures via small steps. AGNN has a novel parameter sharing strategy that enables homogeneous architectures to share parameters, based on a carefully-designed homogeneity definition. Experiments on real-world benchmark datasets demonstrate that the GNN architecture identified by AGNN achieves the best performance, comparing with existing handcrafted models and tradistional search methods.

In this paper, we propose an inverse reinforcement learning method for architecture search (IRLAS), which trains an agent to learn to search network structures that are topologically inspired by human-designed network. Most existing architecture search approaches totally neglect the topological characteristics of architectures, which results in complicated architecture with a high inference latency. Motivated by the fact that human-designed networks are elegant in topology with a fast inference speed, we propose a mirror stimuli function inspired by biological cognition theory to extract the abstract topological knowledge of an expert human-design network (ResNeXt). To avoid raising a too strong prior over the search space, we introduce inverse reinforcement learning to train the mirror stimuli function and exploit it as a heuristic guidance for architecture search, easily generalized to different architecture search algorithms. On CIFAR-10, the best architecture searched by our proposed IRLAS achieves 2.60% error rate. For ImageNet mobile setting, our model achieves a state-of-the-art top-1 accuracy 75.28%, while being 2~4x faster than most auto-generated architectures. A fast version of this model achieves 10% faster than MobileNetV2, while maintaining a higher accuracy.

Automatic neural architecture design has shown its potential in discovering powerful neural network architectures. Existing methods, no matter based on reinforcement learning or evolutionary algorithms (EA), conduct architecture search in a discrete space, which is highly inefficient. In this paper, we propose a simple and efficient method to automatic neural architecture design based on continuous optimization. We call this new approach neural architecture optimization (NAO). There are three key components in our proposed approach: (1) An encoder embeds/maps neural network architectures into a continuous space. (2) A predictor takes the continuous representation of a network as input and predicts its accuracy. (3) A decoder maps a continuous representation of a network back to its architecture. The performance predictor and the encoder enable us to perform gradient based optimization in the continuous space to find the embedding of a new architecture with potentially better accuracy. Such a better embedding is then decoded to a network by the decoder. Experiments show that the architecture discovered by our method is very competitive for image classification task on CIFAR-10 and language modeling task on PTB, outperforming or on par with the best results of previous architecture search methods with a significantly reduction of computational resources. Specifically we obtain $2.07\%$ test set error rate for CIFAR-10 image classification task and $55.9$ test set perplexity of PTB language modeling task. The best discovered architectures on both tasks are successfully transferred to other tasks such as CIFAR-100 and WikiText-2.

Currently, the neural network architecture design is mostly guided by the \emph{indirect} metric of computation complexity, i.e., FLOPs. However, the \emph{direct} metric, e.g., speed, also depends on the other factors such as memory access cost and platform characterics. Thus, this work proposes to evaluate the direct metric on the target platform, beyond only considering FLOPs. Based on a series of controlled experiments, this work derives several practical \emph{guidelines} for efficient network design. Accordingly, a new architecture is presented, called \emph{ShuffleNet V2}. Comprehensive ablation experiments verify that our model is the state-of-the-art in terms of speed and accuracy tradeoff.

Deep neural network architectures have traditionally been designed and explored with human expertise in a long-lasting trial-and-error process. This process requires huge amount of time, expertise, and resources. To address this tedious problem, we propose a novel algorithm to optimally find hyperparameters of a deep network architecture automatically. We specifically focus on designing neural architectures for medical image segmentation task. Our proposed method is based on a policy gradient reinforcement learning for which the reward function is assigned a segmentation evaluation utility (i.e., dice index). We show the efficacy of the proposed method with its low computational cost in comparison with the state-of-the-art medical image segmentation networks. We also present a new architecture design, a densely connected encoder-decoder CNN, as a strong baseline architecture to apply the proposed hyperparameter search algorithm. We apply the proposed algorithm to each layer of the baseline architectures. As an application, we train the proposed system on cine cardiac MR images from Automated Cardiac Diagnosis Challenge (ACDC) MICCAI 2017. Starting from a baseline segmentation architecture, the resulting network architecture obtains the state-of-the-art results in accuracy without performing any trial-and-error based architecture design approaches or close supervision of the hyperparameters changes.

We propose a Bayesian convolutional neural network built upon Bayes by Backprop and elaborate how this known method can serve as the fundamental construct of our novel, reliable variational inference method for convolutional neural networks. First, we show how Bayes by Backprop can be applied to convolutional layers where weights in filters have probability distributions instead of point-estimates; and second, how our proposed framework leads with various network architectures to performances comparable to convolutional neural networks with point-estimates weights. In the past, Bayes by Backprop has been successfully utilised in feedforward and recurrent neural networks, but not in convolutional ones. This work symbolises the extension of the group of Bayesian neural networks which encompasses all three aforementioned types of network architectures now.

北京阿比特科技有限公司