亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

相關內容

神(shen)經(jing)網(wang)(wang)絡(luo)(luo)(Neural Networks)是世界上三個最古老(lao)的(de)(de)(de)神(shen)經(jing)建模(mo)學(xue)(xue)(xue)會(hui)(hui)的(de)(de)(de)檔案期刊:國際神(shen)經(jing)網(wang)(wang)絡(luo)(luo)學(xue)(xue)(xue)會(hui)(hui)(INNS)、歐洲(zhou)神(shen)經(jing)網(wang)(wang)絡(luo)(luo)學(xue)(xue)(xue)會(hui)(hui)(ENNS)和(he)(he)日本神(shen)經(jing)網(wang)(wang)絡(luo)(luo)學(xue)(xue)(xue)會(hui)(hui)(JNNS)。神(shen)經(jing)網(wang)(wang)絡(luo)(luo)提供了一(yi)個論(lun)壇,以(yi)發(fa)(fa)(fa)展和(he)(he)培育一(yi)個國際社會(hui)(hui)的(de)(de)(de)學(xue)(xue)(xue)者和(he)(he)實踐(jian)者感(gan)興趣的(de)(de)(de)所有(you)(you)方(fang)(fang)面的(de)(de)(de)神(shen)經(jing)網(wang)(wang)絡(luo)(luo)和(he)(he)相關(guan)方(fang)(fang)法的(de)(de)(de)計(ji)算智能(neng)。神(shen)經(jing)網(wang)(wang)絡(luo)(luo)歡迎(ying)高質量(liang)(liang)論(lun)文的(de)(de)(de)提交,有(you)(you)助于(yu)全面的(de)(de)(de)神(shen)經(jing)網(wang)(wang)絡(luo)(luo)研究(jiu),從行為和(he)(he)大腦(nao)建模(mo),學(xue)(xue)(xue)習(xi)算法,通過數(shu)學(xue)(xue)(xue)和(he)(he)計(ji)算分析(xi)(xi),系統的(de)(de)(de)工程(cheng)和(he)(he)技術(shu)應用(yong),大量(liang)(liang)使用(yong)神(shen)經(jing)網(wang)(wang)絡(luo)(luo)的(de)(de)(de)概念和(he)(he)技術(shu)。這一(yi)獨(du)特而廣泛的(de)(de)(de)范圍(wei)促(cu)進了生(sheng)物(wu)和(he)(he)技術(shu)研究(jiu)之(zhi)間的(de)(de)(de)思想交流,并有(you)(you)助于(yu)促(cu)進對生(sheng)物(wu)啟發(fa)(fa)(fa)的(de)(de)(de)計(ji)算智能(neng)感(gan)興趣的(de)(de)(de)跨(kua)學(xue)(xue)(xue)科(ke)社區的(de)(de)(de)發(fa)(fa)(fa)展。因此,神(shen)經(jing)網(wang)(wang)絡(luo)(luo)編(bian)委會(hui)(hui)代(dai)表(biao)的(de)(de)(de)專家領(ling)域包括心理學(xue)(xue)(xue),神(shen)經(jing)生(sheng)物(wu)學(xue)(xue)(xue),計(ji)算機(ji)科(ke)學(xue)(xue)(xue),工程(cheng),數(shu)學(xue)(xue)(xue),物(wu)理。該(gai)雜志發(fa)(fa)(fa)表(biao)文章、信(xin)件(jian)和(he)(he)評論(lun)以(yi)及給編(bian)輯的(de)(de)(de)信(xin)件(jian)、社論(lun)、時(shi)事、軟(ruan)件(jian)調(diao)查和(he)(he)專利信(xin)息(xi)。文章發(fa)(fa)(fa)表(biao)在五個部分之(zhi)一(yi):認知科(ke)學(xue)(xue)(xue),神(shen)經(jing)科(ke)學(xue)(xue)(xue),學(xue)(xue)(xue)習(xi)系統,數(shu)學(xue)(xue)(xue)和(he)(he)計(ji)算分析(xi)(xi)、工程(cheng)和(he)(he)應用(yong)。 官網(wang)(wang)地址:

Common self-improvement approaches for large language models (LLMs), such as STaR (Zelikman et al., 2022), iteratively fine-tune LLMs on self-generated solutions to improve their problem-solving ability. However, these approaches discard the large amounts of incorrect solutions generated during this process, potentially neglecting valuable information in such solutions. To address this shortcoming, we propose V-STaR that utilizes both the correct and incorrect solutions generated during the self-improvement process to train a verifier using DPO that judges correctness of model-generated solutions. This verifier is used at inference time to select one solution among many candidate solutions. Running V-STaR for multiple iterations results in progressively better reasoners and verifiers, delivering a 4% to 17% test accuracy improvement over existing self-improvement and verification approaches on common code generation and math reasoning benchmarks with LLaMA2 models.

Pareto Set Learning (PSL) is a promising approach for approximating the entire Pareto front in multi-objective optimization (MOO) problems. However, existing derivative-free PSL methods are often unstable and inefficient, especially for expensive black-box MOO problems where objective function evaluations are costly. In this work, we propose to address the instability and inefficiency of existing PSL methods with a novel controllable PSL method, called Co-PSL. Particularly, Co-PSL consists of two stages: (1) warm-starting Bayesian optimization to obtain quality Gaussian Processes priors and (2) controllable Pareto set learning to accurately acquire a parametric mapping from preferences to the corresponding Pareto solutions. The former is to help stabilize the PSL process and reduce the number of expensive function evaluations. The latter is to support real-time trade-off control between conflicting objectives. Performances across synthesis and real-world MOO problems showcase the effectiveness of our Co-PSL for expensive multi-objective optimization tasks.

We propose a novel method for privacy-preserving deep neural networks (DNNs) with the Vision Transformer (ViT). The method allows us not only to train models and test with visually protected images but to also avoid the performance degradation caused from the use of encrypted images, whereas conventional methods cannot avoid the influence of image encryption. A domain adaptation method is used to efficiently fine-tune ViT with encrypted images. In experiments, the method is demonstrated to outperform conventional methods in an image classification task on the CIFAR-10 and ImageNet datasets in terms of classification accuracy.

The highly sparse activations in Spiking Neural Networks (SNNs) can provide tremendous energy efficiency benefits when carefully exploited in hardware. The behavior of sparsity in SNNs is uniquely shaped by the dataset and training hyperparameters. This work reveals novel insights into the impacts of training on hardware performance. Specifically, we explore the trade-offs between model accuracy and hardware efficiency. We focus on three key hyperparameters: surrogate gradient functions, beta, and membrane threshold. Results on an FPGA-based hardware platform show that the fast sigmoid surrogate function yields a lower firing rate with similar accuracy compared to the arctangent surrogate on the SVHN dataset. Furthermore, by cross-sweeping the beta and membrane threshold hyperparameters, we can achieve a 48% reduction in hardware-based inference latency with only 2.88% trade-off in inference accuracy compared to the default setting. Overall, this study highlights the importance of fine-tuning model hyperparameters as crucial for designing efficient SNN hardware accelerators, evidenced by the fine-tuned model achieving a 1.72x improvement in accelerator efficiency (FPS/W) compared to the most recent work.

Cloud-based large language models (LLMs) such as ChatGPT have increasingly become integral to daily operations, serving as vital tools across various applications. While these models offer substantial benefits in terms of accessibility and functionality, they also introduce significant privacy concerns: the transmission and storage of user data in cloud infrastructures pose substantial risks of data breaches and unauthorized access to sensitive information; even if the transmission and storage of data is encrypted, the LLM service provider itself still knows the real contents of the data, preventing individuals or entities from confidently using such LLM services. To address these concerns, this paper proposes a simple yet effective mechanism PromptCrypt to protect user privacy. It uses Emoji to encrypt the user inputs before sending them to LLM, effectively rendering them indecipherable to human or LLM's examination while retaining the original intent of the prompt, thus ensuring the model's performance remains unaffected. We conduct experiments on three tasks, personalized recommendation, sentiment analysis, and tabular data analysis. Experiment results reveal that PromptCrypt can encrypt personal information within prompts in such a manner that not only prevents the discernment of sensitive data by humans or LLM itself, but also maintains or even improves the precision without further tuning, achieving comparable or even better task accuracy than directly prompting the LLM without prompt encryption. These results highlight the practicality of adopting encryption measures that safeguard user privacy without compromising the functional integrity and performance of LLMs. Code and dataset are available at //github.com/agiresearch/PromptCrypt.

Natural policy gradient (NPG) methods with entropy regularization achieve impressive empirical success in reinforcement learning problems with large state-action spaces. However, their convergence properties and the impact of entropy regularization remain elusive in the function approximation regime. In this paper, we establish finite-time convergence analyses of entropy-regularized NPG with linear function approximation under softmax parameterization. In particular, we prove that entropy-regularized NPG with averaging satisfies the \emph{persistence of excitation} condition, and achieves a fast convergence rate of $\tilde{O}(1/T)$ up to a function approximation error in regularized Markov decision processes. This convergence result does not require any a priori assumptions on the policies. Furthermore, under mild regularity conditions on the concentrability coefficient and basis vectors, we prove that entropy-regularized NPG exhibits \emph{linear convergence} up to a function approximation error.

This paper proposes a novel variant of GFlowNet, genetic-guided GFlowNet (Genetic GFN), which integrates an iterative genetic search into GFlowNet. Genetic search effectively guides the GFlowNet to high-rewarded regions, addressing global over-exploration that results in training inefficiency and exploring limited regions. In addition, training strategies, such as rank-based replay training and unsupervised maximum likelihood pre-training, are further introduced to improve the sample efficiency of Genetic GFN. The proposed method shows a state-of-the-art score of 16.213, significantly outperforming the reported best score in the benchmark of 15.185, in practical molecular optimization (PMO), which is an official benchmark for sample-efficient molecular optimization. Remarkably, ours exceeds all baselines, including reinforcement learning, Bayesian optimization, generative models, GFlowNets, and genetic algorithms, in 14 out of 23 tasks.

Graph neural networks (GNNs) have emerged as a powerful paradigm for embedding-based entity alignment due to their capability of identifying isomorphic subgraphs. However, in real knowledge graphs (KGs), the counterpart entities usually have non-isomorphic neighborhood structures, which easily causes GNNs to yield different representations for them. To tackle this problem, we propose a new KG alignment network, namely AliNet, aiming at mitigating the non-isomorphism of neighborhood structures in an end-to-end manner. As the direct neighbors of counterpart entities are usually dissimilar due to the schema heterogeneity, AliNet introduces distant neighbors to expand the overlap between their neighborhood structures. It employs an attention mechanism to highlight helpful distant neighbors and reduce noises. Then, it controls the aggregation of both direct and distant neighborhood information using a gating mechanism. We further propose a relation loss to refine entity representations. We perform thorough experiments with detailed ablation studies and analyses on five entity alignment datasets, demonstrating the effectiveness of AliNet.

Pre-trained deep neural network language models such as ELMo, GPT, BERT and XLNet have recently achieved state-of-the-art performance on a variety of language understanding tasks. However, their size makes them impractical for a number of scenarios, especially on mobile and edge devices. In particular, the input word embedding matrix accounts for a significant proportion of the model's memory footprint, due to the large input vocabulary and embedding dimensions. Knowledge distillation techniques have had success at compressing large neural network models, but they are ineffective at yielding student models with vocabularies different from the original teacher models. We introduce a novel knowledge distillation technique for training a student model with a significantly smaller vocabulary as well as lower embedding and hidden state dimensions. Specifically, we employ a dual-training mechanism that trains the teacher and student models simultaneously to obtain optimal word embeddings for the student vocabulary. We combine this approach with learning shared projection matrices that transfer layer-wise knowledge from the teacher model to the student model. Our method is able to compress the BERT_BASE model by more than 60x, with only a minor drop in downstream task metrics, resulting in a language model with a footprint of under 7MB. Experimental results also demonstrate higher compression efficiency and accuracy when compared with other state-of-the-art compression techniques.

We propose a novel single shot object detection network named Detection with Enriched Semantics (DES). Our motivation is to enrich the semantics of object detection features within a typical deep detector, by a semantic segmentation branch and a global activation module. The segmentation branch is supervised by weak segmentation ground-truth, i.e., no extra annotation is required. In conjunction with that, we employ a global activation module which learns relationship between channels and object classes in a self-supervised manner. Comprehensive experimental results on both PASCAL VOC and MS COCO detection datasets demonstrate the effectiveness of the proposed method. In particular, with a VGG16 based DES, we achieve an mAP of 81.7 on VOC2007 test and an mAP of 32.8 on COCO test-dev with an inference speed of 31.5 milliseconds per image on a Titan Xp GPU. With a lower resolution version, we achieve an mAP of 79.7 on VOC2007 with an inference speed of 13.0 milliseconds per image.

北京阿比特科技有限公司