亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

The adoption of machine learning solutions is rapidly increasing across all parts of society. As the models grow larger, both training and inference of machine learning models is increasingly outsourced, e.g. to cloud service providers. This means that potentially sensitive data is processed on untrusted platforms, which bears inherent data security and privacy risks. In this work, we investigate how to protect distributed machine learning systems, focusing on deep convolutional neural networks. The most common and best-performing mixed MPC approaches are based on HE, secret sharing, and garbled circuits. They commonly suffer from large performance overheads, big accuracy losses, and communication overheads that grow linearly in the depth of the neural network. To improve on these problems, we present Dash, a fast and distributed private convolutional neural network inference scheme secure against malicious attackers. Building on arithmetic garbling gadgets [BMR16] and fancy-garbling [BCM+19], Dash is based purely on arithmetic garbled circuits. We introduce LabelTensors that allow us to leverage the massive parallelity of modern GPUs. Combined with state-of-the-art garbling optimizations, Dash outperforms previous garbling approaches up to a factor of about 100. Furthermore, we introduce an efficient scaling operation over the residues of the Chinese remainder theorem representation to arithmetic garbled circuits, which allows us to garble larger networks and achieve much higher accuracy than previous approaches. Finally, Dash requires only a single communication round per inference step, regardless of the depth of the neural network, and a very small constant online communication volume.

相關內容

神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)(Neural Networks)是世界上三(san)個(ge)(ge)最古老(lao)的(de)(de)(de)(de)神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)建(jian)模學(xue)(xue)(xue)(xue)會(hui)的(de)(de)(de)(de)檔案期(qi)刊:國際神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)學(xue)(xue)(xue)(xue)會(hui)(INNS)、歐(ou)洲神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)學(xue)(xue)(xue)(xue)會(hui)(ENNS)和(he)(he)日(ri)本神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)學(xue)(xue)(xue)(xue)會(hui)(JNNS)。神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)提供(gong)了一(yi)個(ge)(ge)論(lun)(lun)壇(tan),以發(fa)展和(he)(he)培育一(yi)個(ge)(ge)國際社(she)會(hui)的(de)(de)(de)(de)學(xue)(xue)(xue)(xue)者和(he)(he)實踐者感(gan)(gan)興(xing)趣的(de)(de)(de)(de)所(suo)有(you)方面的(de)(de)(de)(de)神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)和(he)(he)相關方法的(de)(de)(de)(de)計算(suan)智能。神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)歡(huan)迎(ying)高質量(liang)論(lun)(lun)文的(de)(de)(de)(de)提交,有(you)助于(yu)全面的(de)(de)(de)(de)神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)研(yan)究,從行為和(he)(he)大腦建(jian)模,學(xue)(xue)(xue)(xue)習(xi)算(suan)法,通過數學(xue)(xue)(xue)(xue)和(he)(he)計算(suan)分析(xi),系(xi)統(tong)的(de)(de)(de)(de)工程和(he)(he)技術應(ying)用,大量(liang)使用神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)的(de)(de)(de)(de)概(gai)念(nian)和(he)(he)技術。這一(yi)獨(du)特而廣泛的(de)(de)(de)(de)范(fan)圍促(cu)進了生(sheng)物(wu)和(he)(he)技術研(yan)究之間的(de)(de)(de)(de)思(si)想交流(liu),并有(you)助于(yu)促(cu)進對生(sheng)物(wu)啟(qi)發(fa)的(de)(de)(de)(de)計算(suan)智能感(gan)(gan)興(xing)趣的(de)(de)(de)(de)跨學(xue)(xue)(xue)(xue)科社(she)區的(de)(de)(de)(de)發(fa)展。因(yin)此,神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)網(wang)絡(luo)(luo)(luo)(luo)(luo)(luo)編委(wei)會(hui)代表(biao)(biao)的(de)(de)(de)(de)專家領域包括心理學(xue)(xue)(xue)(xue),神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)生(sheng)物(wu)學(xue)(xue)(xue)(xue),計算(suan)機科學(xue)(xue)(xue)(xue),工程,數學(xue)(xue)(xue)(xue),物(wu)理。該雜志發(fa)表(biao)(biao)文章、信件和(he)(he)評(ping)論(lun)(lun)以及(ji)給編輯的(de)(de)(de)(de)信件、社(she)論(lun)(lun)、時事、軟件調查和(he)(he)專利信息。文章發(fa)表(biao)(biao)在五個(ge)(ge)部分之一(yi):認(ren)知科學(xue)(xue)(xue)(xue),神(shen)(shen)(shen)(shen)(shen)經(jing)(jing)科學(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)習(xi)系(xi)統(tong),數學(xue)(xue)(xue)(xue)和(he)(he)計算(suan)分析(xi)、工程和(he)(he)應(ying)用。 官(guan)網(wang)地址(zhi):

Multimodal multihop question answering is a complex task that requires reasoning over multiple sources of information, such as images and text, to answer questions. While there has been significant progress in visual question answering, the multihop setting remains unexplored due to the lack of high-quality datasets. Current methods focus on single-hop question answering or a single modality, which makes them unsuitable for real-world scenarios such as analyzing multimodal educational materials, summarizing lengthy academic articles, or interpreting scientific studies that combine charts, images, and text. To address this gap, we propose a novel methodology, introducing the first framework for creating a high-quality dataset that enables training models for multimodal multihop question answering. Our approach consists of a 5-stage pipeline that involves acquiring relevant multimodal documents from Wikipedia, synthetically generating high-level questions and answers, and validating them through rigorous criteria to ensure quality data. We evaluate our methodology by training models on our synthesized dataset and testing on two benchmarks, our results demonstrate that, with an equal sample size, models trained on our synthesized data outperform those trained on human-collected data by 1.9 in exact match (EM) on average. We believe our data synthesis method will serve as a strong foundation for training and evaluating multimodal multihop question answering models.

The creation of 3D scenes has traditionally been both labor-intensive and costly, requiring designers to meticulously configure 3D assets and environments. Recent advancements in generative AI, including text-to-3D and image-to-3D methods, have dramatically reduced the complexity and cost of this process. However, current techniques for editing complex 3D scenes continue to rely on generally interactive multi-step, 2D-to-3D projection methods and diffusion-based techniques, which often lack precision in control and hamper real-time performance. In this work, we propose 3DSceneEditor, a fully 3D-based paradigm for real-time, precise editing of intricate 3D scenes using Gaussian Splatting. Unlike conventional methods, 3DSceneEditor operates through a streamlined 3D pipeline, enabling direct manipulation of Gaussians for efficient, high-quality edits based on input prompts.The proposed framework (i) integrates a pre-trained instance segmentation model for semantic labeling; (ii) employs a zero-shot grounding approach with CLIP to align target objects with user prompts; and (iii) applies scene modifications, such as object addition, repositioning, recoloring, replacing, and deletion directly on Gaussians. Extensive experimental results show that 3DSceneEditor achieves superior editing precision and speed with respect to current SOTA 3D scene editing approaches, establishing a new benchmark for efficient and interactive 3D scene customization.

Realizing scaling laws in embodied AI has become a focus. However, previous work has been scattered across diverse simulation platforms, with assets and models lacking unified interfaces, which has led to inefficiencies in research. To address this, we introduce InfiniteWorld, a unified and scalable simulator for general vision-language robot interaction built on Nvidia Isaac Sim. InfiniteWorld encompasses a comprehensive set of physics asset construction methods and generalized free robot interaction benchmarks. Specifically, we first built a unified and scalable simulation framework for embodied learning that integrates a series of improvements in generation-driven 3D asset construction, Real2Sim, automated annotation framework, and unified 3D asset processing. This framework provides a unified and scalable platform for robot interaction and learning. In addition, to simulate realistic robot interaction, we build four new general benchmarks, including scene graph collaborative exploration and open-world social mobile manipulation. The former is often overlooked as an important task for robots to explore the environment and build scene knowledge, while the latter simulates robot interaction tasks with different levels of knowledge agents based on the former. They can more comprehensively evaluate the embodied agent's capabilities in environmental understanding, task planning and execution, and intelligent interaction. We hope that this work can provide the community with a systematic asset interface, alleviate the dilemma of the lack of high-quality assets, and provide a more comprehensive evaluation of robot interactions.

Large Language Models (LLMs) based on transformers achieve cutting-edge results on a variety of applications. However, their enormous size and processing requirements make deployment on devices with constrained resources extremely difficult. Among various efficiency considerations, model binarization and Early Exit (EE) are common effective solutions. However, binarization may lead to performance loss due to reduced precision affecting gradient estimation and parameter updates. Besides, the present early-exit mechanisms are still in the nascent stages of research. To ameliorate these issues, we propose Binarized Early Exit Transformer (BEExformer), the first-ever selective learning transformer architecture to combine early exit with binarization for textual inference. It improves the binarization process through a differentiable second-order approximation to the impulse function. This enables gradient computation concerning both the sign as well as the magnitude of the weights. In contrast to absolute threshold-based EE, the proposed EE mechanism hinges on fractional reduction in entropy among intermediate transformer blocks with soft-routing loss estimation. While binarization results in 18.44 times reduction in model size, early exit reduces the FLOPs during inference by 54.85% and even improves accuracy by 5.98% through resolving the "overthinking" problem inherent in deep networks. Moreover, the proposed BEExformer simplifies training by not requiring knowledge distillation from a full-precision LLM. Extensive evaluation on the GLUE dataset and comparison with the SOTA works showcase its pareto-optimal performance-efficiency trade-off.

The increasing frequency and sophistication of cybersecurity vulnerabilities in software systems underscore the urgent need for robust and effective methods of vulnerability assessment. However, existing approaches often rely on highly technical and abstract frameworks, which hinders understanding and increases the likelihood of exploitation, resulting in severe cyberattacks. Given the growing adoption of Large Language Models (LLMs) across diverse domains, this paper explores their potential application in cybersecurity, specifically for enhancing the assessment of software vulnerabilities. We propose ChatNVD, an LLM-based cybersecurity vulnerability assessment tool leveraging the National Vulnerability Database (NVD) to provide context-rich insights and streamline vulnerability analysis for cybersecurity professionals, developers, and non-technical users. We develop three variants of ChatNVD, utilizing three prominent LLMs: GPT-4o mini by OpenAI, Llama 3 by Meta, and Gemini 1.5 Pro by Google. To evaluate their efficacy, we conduct a comparative analysis of these models using a comprehensive questionnaire comprising common security vulnerability questions, assessing their accuracy in identifying and analyzing software vulnerabilities. This study provides valuable insights into the potential of LLMs to address critical challenges in understanding and mitigation of software vulnerabilities.

Deep learning has shown great potential for modeling the physical dynamics of complex particle systems such as fluids (in Lagrangian descriptions). Existing approaches, however, require the supervision of consecutive particle properties, including positions and velocities. In this paper, we consider a partially observable scenario known as fluid dynamics grounding, that is, inferring the state transitions and interactions within the fluid particle systems from sequential visual observations of the fluid surface. We propose a differentiable two-stage network named NeuroFluid. Our approach consists of (i) a particle-driven neural renderer, which involves fluid physical properties into the volume rendering function, and (ii) a particle transition model optimized to reduce the differences between the rendered and the observed images. NeuroFluid provides the first solution to unsupervised learning of particle-based fluid dynamics by training these two models jointly. It is shown to reasonably estimate the underlying physics of fluids with different initial shapes, viscosity, and densities. It is a potential alternative approach to understanding complex fluid mechanics, such as turbulence, that are difficult to model using traditional methods of mathematical physics.

The design of deep graph models still remains to be investigated and the crucial part is how to explore and exploit the knowledge from different hops of neighbors in an efficient way. In this paper, we propose a novel RNN-like deep graph neural network architecture by incorporating AdaBoost into the computation of network; and the proposed graph convolutional network called AdaGCN~(AdaBoosting Graph Convolutional Network) has the ability to efficiently extract knowledge from high-order neighbors and integrate knowledge from different hops of neighbors into the network in an AdaBoost way. We also present the architectural difference between AdaGCN and existing graph convolutional methods to show the benefits of our proposal. Finally, extensive experiments demonstrate the state-of-the-art prediction performance and the computational advantage of our approach AdaGCN.

With the capability of modeling bidirectional contexts, denoising autoencoding based pretraining like BERT achieves better performance than pretraining approaches based on autoregressive language modeling. However, relying on corrupting the input with masks, BERT neglects dependency between the masked positions and suffers from a pretrain-finetune discrepancy. In light of these pros and cons, we propose XLNet, a generalized autoregressive pretraining method that (1) enables learning bidirectional contexts by maximizing the expected likelihood over all permutations of the factorization order and (2) overcomes the limitations of BERT thanks to its autoregressive formulation. Furthermore, XLNet integrates ideas from Transformer-XL, the state-of-the-art autoregressive model, into pretraining. Empirically, XLNet outperforms BERT on 20 tasks, often by a large margin, and achieves state-of-the-art results on 18 tasks including question answering, natural language inference, sentiment analysis, and document ranking.

This paper surveys the machine learning literature and presents machine learning as optimization models. Such models can benefit from the advancement of numerical optimization techniques which have already played a distinctive role in several machine learning settings. Particularly, mathematical optimization models are presented for commonly used machine learning approaches for regression, classification, clustering, and deep neural networks as well new emerging applications in machine teaching and empirical model learning. The strengths and the shortcomings of these models are discussed and potential research directions are highlighted.

Convolutional Neural Networks (CNNs) have gained significant traction in the field of machine learning, particularly due to their high accuracy in visual recognition. Recent works have pushed the performance of GPU implementations of CNNs to significantly improve their classification and training times. With these improvements, many frameworks have become available for implementing CNNs on both CPUs and GPUs, with no support for FPGA implementations. In this work we present a modified version of the popular CNN framework Caffe, with FPGA support. This allows for classification using CNN models and specialized FPGA implementations with the flexibility of reprogramming the device when necessary, seamless memory transactions between host and device, simple-to-use test benches, and the ability to create pipelined layer implementations. To validate the framework, we use the Xilinx SDAccel environment to implement an FPGA-based Winograd convolution engine and show that the FPGA layer can be used alongside other layers running on a host processor to run several popular CNNs (AlexNet, GoogleNet, VGG A, Overfeat). The results show that our framework achieves 50 GFLOPS across 3x3 convolutions in the benchmarks. This is achieved within a practical framework, which will aid in future development of FPGA-based CNNs.

北京阿比特科技有限公司