亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

The cyber-threat landscape has evolved tremendously in recent years, with new threat variants emerging daily, and large-scale coordinated campaigns becoming more prevalent. In this study, we propose CELEST (CollaborativE LEarning for Scalable Threat detection, a federated machine learning framework for global threat detection over HTTP, which is one of the most commonly used protocols for malware dissemination and communication. CELEST leverages federated learning in order to collaboratively train a global model across multiple clients who keep their data locally, thus providing increased privacy and confidentiality assurances. Through a novel active learning component integrated with the federated learning technique, our system continuously discovers and learns the behavior of new, evolving, and globally-coordinated cyber threats. We show that CELEST is able to expose attacks that are largely invisible to individual organizations. For instance, in one challenging attack scenario with data exfiltration malware, the global model achieves a three-fold increase in Precision-Recall AUC compared to the local model. We also design a poisoning detection and mitigation method, DTrust, specifically designed for federated learning in the collaborative threat detection domain. DTrust successfully detects poisoning clients using the feedback from participating clients to investigate and remove them from the training process. We deploy CELEST on two university networks and show that it is able to detect the malicious HTTP communication with high precision and low false positive rates. Furthermore, during its deployment, CELEST detected a set of previously unknown 42 malicious URLs and 20 malicious domains in one day, which were confirmed to be malicious by VirusTotal.

相關內容

Federated noisy label learning (FNLL) is emerging as a promising tool for privacy-preserving multi-source decentralized learning. Existing research, relying on the assumption of class-balanced global data, might be incapable to model complicated label noise, especially in medical scenarios. In this paper, we first formulate a new and more realistic federated label noise problem where global data is class-imbalanced and label noise is heterogeneous, and then propose a two-stage framework named FedNoRo for noise-robust federated learning. Specifically, in the first stage of FedNoRo, per-class loss indicators followed by Gaussian Mixture Model are deployed for noisy client identification. In the second stage, knowledge distillation and a distance-aware aggregation function are jointly adopted for noise-robust federated model updating. Experimental results on the widely-used ICH and ISIC2019 datasets demonstrate the superiority of FedNoRo against the state-of-the-art FNLL methods for addressing class imbalance and label noise heterogeneity in real-world FL scenarios.

Obtaining high-quality data for collaborative training of machine learning models can be a challenging task due to A) the regulatory concerns and B) lack of incentive to participate. The first issue can be addressed through the use of privacy enhancing technologies (PET), one of the most frequently used one being differentially private (DP) training. The second challenge can be addressed by identifying which data points can be beneficial for model training and rewarding data owners for sharing this data. However, DP in deep learning typically adversely affects atypical (often informative) data samples, making it difficult to assess the usefulness of individual contributions. In this work we investigate how to leverage gradient information to identify training samples of interest in private training settings. We show that there exist techniques which are able to provide the clients with the tools for principled data selection even in strictest privacy settings.

Generative Adversarial Networks (GAN) have led to the generation of very realistic face images, which have been used in fake social media accounts and other disinformation matters that can generate profound impacts. Therefore, the corresponding GAN-face detection techniques are under active development that can examine and expose such fake faces. In this work, we aim to provide a comprehensive review of recent progress in GAN-face detection. We focus on methods that can detect face images that are generated or synthesized from GAN models. We classify the existing detection works into four categories: (1) deep learning-based, (2) physical-based, (3) physiological-based methods, and (4) evaluation and comparison against human visual performance. For each category, we summarize the key ideas and connect them with method implementations. We also discuss open problems and suggest future research directions.

近(jin)年來(lai),網(wang)絡威脅(xie)環(huan)境發生了(le)巨大(da)的(de)變化,每天(tian)都有新的(de)威脅(xie)變體出(chu)現,大(da)規(gui)模(mo)(mo)的(de)協(xie)調活動也變得越來(lai)越普遍(bian)。在這項研(yan)究中(zhong)(zhong),我(wo)們提出(chu)了(le)CELEST(CollaborativE LEarning for Scalable Threat detection),這是(shi)一(yi)(yi)(yi)(yi)個(ge)(ge)(ge)(ge)用(yong)于HTTP全(quan)(quan)球(qiu)(qiu)威脅(xie)檢(jian)(jian)測(ce)(ce)的(de)聯(lian)(lian)合機(ji)器(qi)學(xue)習(xi)(xi)框架,HTTP是(shi)最常(chang)用(yong)的(de)惡意(yi)軟件(jian)(jian)傳播和(he)通信協(xie)議之(zhi)一(yi)(yi)(yi)(yi)。CELEST利用(yong)聯(lian)(lian)邦學(xue)習(xi)(xi),以(yi)便在本(ben)(ben)地保存數據的(de)多個(ge)(ge)(ge)(ge)客戶之(zhi)間(jian)協(xie)作(zuo)訓(xun)練一(yi)(yi)(yi)(yi)個(ge)(ge)(ge)(ge)全(quan)(quan)球(qiu)(qiu)模(mo)(mo)型。通過(guo)與聯(lian)(lian)邦學(xue)習(xi)(xi)技術(shu)相(xiang)結合的(de)新型主動學(xue)習(xi)(xi)組(zu)(zu)件(jian)(jian),我(wo)們的(de)系統(tong)不(bu)斷發現和(he)學(xue)習(xi)(xi)新的(de)、不(bu)斷發展的(de)和(he)全(quan)(quan)球(qiu)(qiu)協(xie)調的(de)網(wang)絡威脅(xie)的(de)行為。我(wo)們表明(ming),CELEST能夠暴(bao)露出(chu)單(dan)個(ge)(ge)(ge)(ge)組(zu)(zu)織(zhi)基本(ben)(ben)上看不(bu)到的(de)攻擊。例如(ru),在一(yi)(yi)(yi)(yi)個(ge)(ge)(ge)(ge)具(ju)有挑戰性的(de)數據滲透惡意(yi)軟件(jian)(jian)的(de)攻擊場(chang)景中(zhong)(zhong),與本(ben)(ben)地模(mo)(mo)型相(xiang)比,全(quan)(quan)局模(mo)(mo)型實現了(le)精(jing)準度(du)-召回(hui)AUC的(de)三倍增長。我(wo)們還(huan)設計了(le)一(yi)(yi)(yi)(yi)種中(zhong)(zhong)毒檢(jian)(jian)測(ce)(ce)和(he)緩解方法,即(ji)DTrust,專(zhuan)門為協(xie)作(zuo)威脅(xie)檢(jian)(jian)測(ce)(ce)領域的(de)聯(lian)(lian)邦學(xue)習(xi)(xi)而(er)設計。我(wo)們在兩個(ge)(ge)(ge)(ge)大(da)學(xue)網(wang)絡上部署了(le)CELEST,并表明(ming)它能夠以(yi)高(gao)精(jing)確度(du)和(he)低假陽性率檢(jian)(jian)測(ce)(ce)惡意(yi)的(de)HTTP通信。此外,在其部署過(guo)程中(zhong)(zhong),CELEST在一(yi)(yi)(yi)(yi)天(tian)內(nei)檢(jian)(jian)測(ce)(ce)到了(le)一(yi)(yi)(yi)(yi)組(zu)(zu)以(yi)前未知的(de)42個(ge)(ge)(ge)(ge)惡意(yi)URL和(he)20個(ge)(ge)(ge)(ge)惡意(yi)域名,并被VirusTotal證實為惡意(yi)的(de)。

圖1:用于 URL 表示的(de)聯合嵌入(ru)模型訓練。我們(men)為 URL、Domain 和 Referer 生成嵌入(ru)式特(te)(te)征(zheng)。除了嵌入(ru)式特(te)(te)征(zheng),我們(men)還包括數字特(te)(te)征(zheng)和分(fen)類特(te)(te)征(zheng)。

圖2:主動(dong)聯(lian)邦(bang)學習框架

付費5元查看完整內容

Time series anomaly detection has applications in a wide range of research fields and applications, including manufacturing and healthcare. The presence of anomalies can indicate novel or unexpected events, such as production faults, system defects, or heart fluttering, and is therefore of particular interest. The large size and complex patterns of time series have led researchers to develop specialised deep learning models for detecting anomalous patterns. This survey focuses on providing structured and comprehensive state-of-the-art time series anomaly detection models through the use of deep learning. It providing a taxonomy based on the factors that divide anomaly detection models into different categories. Aside from describing the basic anomaly detection technique for each category, the advantages and limitations are also discussed. Furthermore, this study includes examples of deep anomaly detection in time series across various application domains in recent years. It finally summarises open issues in research and challenges faced while adopting deep anomaly detection models.

Federated learning (FL) has been proposed to protect data privacy and virtually assemble the isolated data silos by cooperatively training models among organizations without breaching privacy and security. However, FL faces heterogeneity from various aspects, including data space, statistical, and system heterogeneity. For example, collaborative organizations without conflict of interest often come from different areas and have heterogeneous data from different feature spaces. Participants may also want to train heterogeneous personalized local models due to non-IID and imbalanced data distribution and various resource-constrained devices. Therefore, heterogeneous FL is proposed to address the problem of heterogeneity in FL. In this survey, we comprehensively investigate the domain of heterogeneous FL in terms of data space, statistical, system, and model heterogeneity. We first give an overview of FL, including its definition and categorization. Then, We propose a precise taxonomy of heterogeneous FL settings for each type of heterogeneity according to the problem setting and learning objective. We also investigate the transfer learning methodologies to tackle the heterogeneity in FL. We further present the applications of heterogeneous FL. Finally, we highlight the challenges and opportunities and envision promising future research directions toward new framework design and trustworthy approaches.

The cyber-threat landscape has evolved tremendously in recent years, with new threat variants emerging daily, and large-scale coordinated campaigns becoming more prevalent. In this study, we propose CELEST (CollaborativE LEarning for Scalable Threat detection), a federated machine learning framework for global threat detection over HTTP, which is one of the most commonly used protocols for malware dissemination and communication. CELEST leverages federated learning in order to collaboratively train a global model across multiple clients who keep their data locally, thus providing increased privacy and confidentiality assurances. Through a novel active learning component integrated with the federated learning technique, our system continuously discovers and learns the behavior of new, evolving, and globally-coordinated cyber threats. We show that CELEST is able to expose attacks that are largely invisible to individual organizations. For instance, in one challenging attack scenario with data exfiltration malware, the global model achieves a three-fold increase in Precision-Recall AUC compared to the local model. We deploy CELEST on two university networks and show that it is able to detect the malicious HTTP communication with high precision and low false positive rates. Furthermore, during its deployment, CELEST detected a set of previously unknown 42 malicious URLs and 20 malicious domains in one day, which were confirmed to be malicious by VirusTotal.

Federated learning enables multiple parties to collaboratively train a machine learning model without communicating their local data. A key challenge in federated learning is to handle the heterogeneity of local data distribution across parties. Although many studies have been proposed to address this challenge, we find that they fail to achieve high performance in image datasets with deep learning models. In this paper, we propose MOON: model-contrastive federated learning. MOON is a simple and effective federated learning framework. The key idea of MOON is to utilize the similarity between model representations to correct the local training of individual parties, i.e., conducting contrastive learning in model-level. Our extensive experiments show that MOON significantly outperforms the other state-of-the-art federated learning algorithms on various image classification tasks.

In this paper, we study the few-shot multi-label classification for user intent detection. For multi-label intent detection, state-of-the-art work estimates label-instance relevance scores and uses a threshold to select multiple associated intent labels. To determine appropriate thresholds with only a few examples, we first learn universal thresholding experience on data-rich domains, and then adapt the thresholds to certain few-shot domains with a calibration based on nonparametric learning. For better calculation of label-instance relevance score, we introduce label name embedding as anchor points in representation space, which refines representations of different classes to be well-separated from each other. Experiments on two datasets show that the proposed model significantly outperforms strong baselines in both one-shot and five-shot settings.

In recent years, mobile devices have gained increasingly development with stronger computation capability and larger storage. Some of the computation-intensive machine learning and deep learning tasks can now be run on mobile devices. To take advantage of the resources available on mobile devices and preserve users' privacy, the idea of mobile distributed machine learning is proposed. It uses local hardware resources and local data to solve machine learning sub-problems on mobile devices, and only uploads computation results instead of original data to contribute to the optimization of the global model. This architecture can not only relieve computation and storage burden on servers, but also protect the users' sensitive information. Another benefit is the bandwidth reduction, as various kinds of local data can now participate in the training process without being uploaded to the server. In this paper, we provide a comprehensive survey on recent studies of mobile distributed machine learning. We survey a number of widely-used mobile distributed machine learning methods. We also present an in-depth discussion on the challenges and future directions in this area. We believe that this survey can demonstrate a clear overview of mobile distributed machine learning and provide guidelines on applying mobile distributed machine learning to real applications.

北京阿比特科技有限公司