亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Deep neural networks have achieved outstanding performance over various tasks, but they have a critical issue: over-confident predictions even for completely unknown samples. Many studies have been proposed to successfully filter out these unknown samples, but they only considered narrow and specific tasks, referred to as misclassification detection, open-set recognition, or out-of-distribution detection. In this work, we argue that these tasks should be treated as fundamentally an identical problem because an ideal model should possess detection capability for all those tasks. Therefore, we introduce the unknown detection task, an integration of previous individual tasks, for a rigorous examination of the detection capability of deep neural networks on a wide spectrum of unknown samples. To this end, unified benchmark datasets on different scales were constructed and the unknown detection capabilities of existing popular methods were subject to comparison. We found that Deep Ensemble consistently outperforms the other approaches in detecting unknowns; however, all methods are only successful for a specific type of unknown. The reproducible code and benchmark datasets are available at //github.com/daintlab/unknown-detection-benchmarks .

相關內容

神(shen)(shen)經網(wang)(wang)絡(luo)(Neural Networks)是世界上三個最古老的(de)(de)(de)(de)(de)神(shen)(shen)經建模(mo)學(xue)(xue)(xue)(xue)(xue)會(hui)(hui)的(de)(de)(de)(de)(de)檔案期刊:國際神(shen)(shen)經網(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)會(hui)(hui)(INNS)、歐洲神(shen)(shen)經網(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)會(hui)(hui)(ENNS)和(he)(he)(he)日本神(shen)(shen)經網(wang)(wang)絡(luo)學(xue)(xue)(xue)(xue)(xue)會(hui)(hui)(JNNS)。神(shen)(shen)經網(wang)(wang)絡(luo)提供了(le)一(yi)(yi)個論(lun)壇,以發展和(he)(he)(he)培育一(yi)(yi)個國際社會(hui)(hui)的(de)(de)(de)(de)(de)學(xue)(xue)(xue)(xue)(xue)者(zhe)和(he)(he)(he)實(shi)踐者(zhe)感(gan)興(xing)趣(qu)的(de)(de)(de)(de)(de)所有(you)方面的(de)(de)(de)(de)(de)神(shen)(shen)經網(wang)(wang)絡(luo)和(he)(he)(he)相關(guan)方法的(de)(de)(de)(de)(de)計(ji)算(suan)智(zhi)能。神(shen)(shen)經網(wang)(wang)絡(luo)歡迎高(gao)質量(liang)論(lun)文(wen)的(de)(de)(de)(de)(de)提交(jiao),有(you)助于(yu)全面的(de)(de)(de)(de)(de)神(shen)(shen)經網(wang)(wang)絡(luo)研(yan)究(jiu),從行(xing)為(wei)和(he)(he)(he)大腦建模(mo),學(xue)(xue)(xue)(xue)(xue)習算(suan)法,通(tong)過數學(xue)(xue)(xue)(xue)(xue)和(he)(he)(he)計(ji)算(suan)分(fen)析(xi),系統(tong)的(de)(de)(de)(de)(de)工程(cheng)和(he)(he)(he)技術應用(yong),大量(liang)使用(yong)神(shen)(shen)經網(wang)(wang)絡(luo)的(de)(de)(de)(de)(de)概(gai)念和(he)(he)(he)技術。這一(yi)(yi)獨特(te)而廣泛的(de)(de)(de)(de)(de)范圍促進了(le)生物(wu)和(he)(he)(he)技術研(yan)究(jiu)之(zhi)間的(de)(de)(de)(de)(de)思想交(jiao)流,并有(you)助于(yu)促進對(dui)生物(wu)啟發的(de)(de)(de)(de)(de)計(ji)算(suan)智(zhi)能感(gan)興(xing)趣(qu)的(de)(de)(de)(de)(de)跨學(xue)(xue)(xue)(xue)(xue)科社區的(de)(de)(de)(de)(de)發展。因此(ci),神(shen)(shen)經網(wang)(wang)絡(luo)編委會(hui)(hui)代表(biao)的(de)(de)(de)(de)(de)專家領域包括(kuo)心理(li)學(xue)(xue)(xue)(xue)(xue),神(shen)(shen)經生物(wu)學(xue)(xue)(xue)(xue)(xue),計(ji)算(suan)機科學(xue)(xue)(xue)(xue)(xue),工程(cheng),數學(xue)(xue)(xue)(xue)(xue),物(wu)理(li)。該(gai)雜志發表(biao)文(wen)章、信件和(he)(he)(he)評論(lun)以及給編輯的(de)(de)(de)(de)(de)信件、社論(lun)、時事、軟(ruan)件調(diao)查和(he)(he)(he)專利信息。文(wen)章發表(biao)在五個部分(fen)之(zhi)一(yi)(yi):認知科學(xue)(xue)(xue)(xue)(xue),神(shen)(shen)經科學(xue)(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)(xue)習系統(tong),數學(xue)(xue)(xue)(xue)(xue)和(he)(he)(he)計(ji)算(suan)分(fen)析(xi)、工程(cheng)和(he)(he)(he)應用(yong)。 官網(wang)(wang)地址(zhi):

Data transformations (e.g. rotations, reflections, and cropping) play an important role in self-supervised learning. Typically, images are transformed into different views, and neural networks trained on tasks involving these views produce useful feature representations for downstream tasks, including anomaly detection. However, for anomaly detection beyond image data, it is often unclear which transformations to use. Here we present a simple end-to-end procedure for anomaly detection with learnable transformations. The key idea is to embed the transformed data into a semantic space such that the transformed data still resemble their untransformed form, while different transformations are easily distinguishable. Extensive experiments on time series demonstrate that our proposed method outperforms existing approaches in the one-vs.-rest setting and is competitive in the more challenging n-vs.-rest anomaly detection task. On tabular datasets from the medical and cyber-security domains, our method learns domain-specific transformations and detects anomalies more accurately than previous work.

We study backdoor poisoning attacks against image classification networks, whereby an attacker inserts a trigger into a subset of the training data, in such a way that at test time, this trigger causes the classifier to predict some target class. %There are several techniques proposed in the literature that aim to detect the attack but only a few also propose to defend against it, and they typically involve retraining the network which is not always possible in practice. We propose lightweight automated detection and correction techniques against poisoning attacks, which are based on neuron patterns mined from the network using a small set of clean and poisoned test samples with known labels. The patterns built based on the mis-classified samples are used for run-time detection of new poisoned inputs. For correction, we propose an input correction technique that uses a differential analysis to identify the trigger in the detected poisoned images, which is then reset to a neutral color. Our detection and correction are performed at run-time and input level, which is in contrast to most existing work that is focused on offline model-level defenses. We demonstrate that our technique outperforms existing defenses such as NeuralCleanse and STRIP on popular benchmarks such as MNIST, CIFAR-10, and GTSRB against the popular BadNets attack and the more complex DFST attack.

A community reveals the features and connections of its members that are different from those in other communities in a network. Detecting communities is of great significance in network analysis. Despite the classical spectral clustering and statistical inference methods, we notice a significant development of deep learning techniques for community detection in recent years with their advantages in handling high dimensional network data. Hence, a comprehensive overview of community detection's latest progress through deep learning is timely to both academics and practitioners. This survey devises and proposes a new taxonomy covering different categories of the state-of-the-art methods, including deep learning-based models upon deep neural networks, deep nonnegative matrix factorization and deep sparse filtering. The main category, i.e., deep neural networks, is further divided into convolutional networks, graph attention networks, generative adversarial networks and autoencoders. The survey also summarizes the popular benchmark data sets, model evaluation metrics, and open-source implementations to address experimentation settings. We then discuss the practical applications of community detection in various domains and point to implementation scenarios. Finally, we outline future directions by suggesting challenging topics in this fast-growing deep learning field.

Mainstream object detectors based on the fully convolutional network has achieved impressive performance. While most of them still need a hand-designed non-maximum suppression (NMS) post-processing, which impedes fully end-to-end training. In this paper, we give the analysis of discarding NMS, where the results reveal that a proper label assignment plays a crucial role. To this end, for fully convolutional detectors, we introduce a Prediction-aware One-To-One (POTO) label assignment for classification to enable end-to-end detection, which obtains comparable performance with NMS. Besides, a simple 3D Max Filtering (3DMF) is proposed to utilize the multi-scale features and improve the discriminability of convolutions in the local region. With these techniques, our end-to-end framework achieves competitive performance against many state-of-the-art detectors with NMS on COCO and CrowdHuman datasets. The code is available at //github.com/Megvii-BaseDetection/DeFCN .

In this paper, we present a comprehensive review of the imbalance problems in object detection. To analyze the problems in a systematic manner, we introduce a problem-based taxonomy. Following this taxonomy, we discuss each problem in depth and present a unifying yet critical perspective on the solutions in the literature. In addition, we identify major open issues regarding the existing imbalance problems as well as imbalance problems that have not been discussed before. Moreover, in order to keep our review up to date, we provide an accompanying webpage which catalogs papers addressing imbalance problems, according to our problem-based taxonomy. Researchers can track newer studies on this webpage available at: //github.com/kemaloksuz/ObjectDetectionImbalance .

Man-made scenes can be densely packed, containing numerous objects, often identical, positioned in close proximity. We show that precise object detection in such scenes remains a challenging frontier even for state-of-the-art object detectors. We propose a novel, deep-learning based method for precise object detection, designed for such challenging settings. Our contributions include: (1) A layer for estimating the Jaccard index as a detection quality score; (2) a novel EM merging unit, which uses our quality scores to resolve detection overlap ambiguities; finally, (3) an extensive, annotated data set, SKU-110K, representing packed retail environments, released for training and testing under such extreme settings. Detection tests on SKU-110K and counting tests on the CARPK and PUCPR+ show our method to outperform existing state-of-the-art with substantial margins. The code and data will be made available on \url{www.github.com/eg4000/SKU110K_CVPR19}.

Current top-performing object detectors depend on deep CNN backbones, such as ResNet-101 and Inception, benefiting from their powerful feature representations but suffering from high computational costs. Conversely, some lightweight model based detectors fulfil real time processing, while their accuracies are often criticized. In this paper, we explore an alternative to build a fast and accurate detector by strengthening lightweight features using a hand-crafted mechanism. Inspired by the structure of Receptive Fields (RFs) in human visual systems, we propose a novel RF Block (RFB) module, which takes the relationship between the size and eccentricity of RFs into account, to enhance the feature discriminability and robustness. We further assemble RFB to the top of SSD, constructing the RFB Net detector. To evaluate its effectiveness, experiments are conducted on two major benchmarks and the results show that RFB Net is able to reach the performance of advanced very deep detectors while keeping the real-time speed. Code is available at //github.com/ruinmessi/RFBNet.

Average precision (AP), the area under the recall-precision (RP) curve, is the standard performance measure for object detection. Despite its wide acceptance, it has a number of shortcomings, the most important of which are (i) the inability to distinguish very different RP curves, and (ii) the lack of directly measuring bounding box localization accuracy. In this paper, we propose 'Localization Recall Precision (LRP) Error', a new metric which we specifically designed for object detection. LRP Error is composed of three components related to localization, false negative (FN) rate and false positive (FP) rate. Based on LRP, we introduce the 'Optimal LRP', the minimum achievable LRP error representing the best achievable configuration of the detector in terms of recall-precision and the tightness of the boxes. In contrast to AP, which considers precisions over the entire recall domain, Optimal LRP determines the 'best' confidence score threshold for a class, which balances the trade-off between localization and recall-precision. In our experiments, we show that, for state-of-the-art object (SOTA) detectors, Optimal LRP provides richer and more discriminative information than AP. We also demonstrate that the best confidence score thresholds vary significantly among classes and detectors. Moreover, we present LRP results of a simple online video object detector which uses a SOTA still image object detector and show that the class-specific optimized thresholds increase the accuracy against the common approach of using a general threshold for all classes. We provide the source code that can compute LRP for the PASCAL VOC and MSCOCO datasets in //github.com/cancam/LRP. Our source code can easily be adapted to other datasets as well.

As we move towards large-scale object detection, it is unrealistic to expect annotated training data for all object classes at sufficient scale, and so methods capable of unseen object detection are required. We propose a novel zero-shot method based on training an end-to-end model that fuses semantic attribute prediction with visual features to propose object bounding boxes for seen and unseen classes. While we utilize semantic features during training, our method is agnostic to semantic information for unseen classes at test-time. Our method retains the efficiency and effectiveness of YOLO for objects seen during training, while improving its performance for novel and unseen objects. The ability of state-of-art detection methods to learn discriminative object features to reject background proposals also limits their performance for unseen objects. We posit that, to detect unseen objects, we must incorporate semantic information into the visual domain so that the learned visual features reflect this information and leads to improved recall rates for unseen objects. We test our method on PASCAL VOC and MS COCO dataset and observed significant improvements on the average precision of unseen classes.

Automatically creating the description of an image using any natural languages sentence like English is a very challenging task. It requires expertise of both image processing as well as natural language processing. This paper discuss about different available models for image captioning task. We have also discussed about how the advancement in the task of object recognition and machine translation has greatly improved the performance of image captioning model in recent years. In addition to that we have discussed how this model can be implemented. In the end, we have also evaluated the performance of model using standard evaluation matrices.

北京阿比特科技有限公司