亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Anomaly detection is crucial to the advanced identification of product defects such as incorrect parts, misaligned components, and damages in industrial manufacturing. Due to the rare observations and unknown types of defects, anomaly detection is considered to be challenging in machine learning. To overcome this difficulty, recent approaches utilize the common visual representations pre-trained from natural image datasets and distill the relevant features. However, existing approaches still have the discrepancy between the pre-trained feature and the target data, or require the input augmentation which should be carefully designed, particularly for the industrial dataset. In this paper, we introduce ReConPatch, which constructs discriminative features for anomaly detection by training a linear modulation of patch features extracted from the pre-trained model. ReConPatch employs contrastive representation learning to collect and distribute features in a way that produces a target-oriented and easily separable representation. To address the absence of labeled pairs for the contrastive learning, we utilize two similarity measures between data representations, pairwise and contextual similarities, as pseudo-labels. Our method achieves the state-of-the-art anomaly detection performance (99.72%) for the widely used and challenging MVTec AD dataset. Additionally, we achieved a state-of-the-art anomaly detection performance (95.8%) for the BTAD dataset.

相關內容

在(zai)數(shu)據(ju)(ju)挖(wa)掘中,異(yi)(yi)常(chang)(chang)(chang)檢(jian)測(ce)(ce)(ce)(ce)(英語:anomaly detection)對不符合預期模式(shi)(shi)或數(shu)據(ju)(ju)集(ji)(ji)中其他(ta)項(xiang)目的(de)(de)(de)項(xiang)目、事(shi)件或觀測(ce)(ce)(ce)(ce)值(zhi)的(de)(de)(de)識別。通(tong)常(chang)(chang)(chang)異(yi)(yi)常(chang)(chang)(chang)項(xiang)目會轉變(bian)成銀行欺詐、結構(gou)缺陷、醫療問(wen)題(ti)(ti)、文本錯誤等類(lei)(lei)型(xing)(xing)的(de)(de)(de)問(wen)題(ti)(ti)。異(yi)(yi)常(chang)(chang)(chang)也被稱為離群值(zhi)、新奇、噪聲(sheng)、偏差(cha)和例外。 特別是(shi)在(zai)檢(jian)測(ce)(ce)(ce)(ce)濫用與(yu)網絡入(ru)侵(qin)時,有趣性(xing)(xing)對象往(wang)往(wang)不是(shi)罕(han)見(jian)對象,但卻(que)是(shi)超(chao)出(chu)預料(liao)的(de)(de)(de)突發活動(dong)。這種模式(shi)(shi)不遵(zun)循通(tong)常(chang)(chang)(chang)統(tong)計定義中把異(yi)(yi)常(chang)(chang)(chang)點(dian)看作是(shi)罕(han)見(jian)對象,于是(shi)許(xu)多(duo)異(yi)(yi)常(chang)(chang)(chang)檢(jian)測(ce)(ce)(ce)(ce)方(fang)(fang)(fang)法(fa)(fa)(fa)(fa)(特別是(shi)無(wu)監(jian)督的(de)(de)(de)方(fang)(fang)(fang)法(fa)(fa)(fa)(fa))將對此類(lei)(lei)數(shu)據(ju)(ju)失(shi)效,除非進(jin)行了(le)合適(shi)的(de)(de)(de)聚(ju)集(ji)(ji)。相反,聚(ju)類(lei)(lei)分析算法(fa)(fa)(fa)(fa)可能(neng)(neng)可以檢(jian)測(ce)(ce)(ce)(ce)出(chu)這些模式(shi)(shi)形成的(de)(de)(de)微聚(ju)類(lei)(lei)。 有三大類(lei)(lei)異(yi)(yi)常(chang)(chang)(chang)檢(jian)測(ce)(ce)(ce)(ce)方(fang)(fang)(fang)法(fa)(fa)(fa)(fa)。[1] 在(zai)假設數(shu)據(ju)(ju)集(ji)(ji)中大多(duo)數(shu)實例都是(shi)正常(chang)(chang)(chang)的(de)(de)(de)前提下,無(wu)監(jian)督異(yi)(yi)常(chang)(chang)(chang)檢(jian)測(ce)(ce)(ce)(ce)方(fang)(fang)(fang)法(fa)(fa)(fa)(fa)能(neng)(neng)通(tong)過尋找與(yu)其他(ta)數(shu)據(ju)(ju)最(zui)不匹配的(de)(de)(de)實例來檢(jian)測(ce)(ce)(ce)(ce)出(chu)未標記測(ce)(ce)(ce)(ce)試數(shu)據(ju)(ju)的(de)(de)(de)異(yi)(yi)常(chang)(chang)(chang)。監(jian)督式(shi)(shi)異(yi)(yi)常(chang)(chang)(chang)檢(jian)測(ce)(ce)(ce)(ce)方(fang)(fang)(fang)法(fa)(fa)(fa)(fa)需(xu)要一(yi)(yi)個已經被標記“正常(chang)(chang)(chang)”與(yu)“異(yi)(yi)常(chang)(chang)(chang)”的(de)(de)(de)數(shu)據(ju)(ju)集(ji)(ji),并涉(she)及到(dao)訓練分類(lei)(lei)器(與(yu)許(xu)多(duo)其他(ta)的(de)(de)(de)統(tong)計分類(lei)(lei)問(wen)題(ti)(ti)的(de)(de)(de)關鍵區別是(shi)異(yi)(yi)常(chang)(chang)(chang)檢(jian)測(ce)(ce)(ce)(ce)的(de)(de)(de)內(nei)在(zai)不均衡(heng)性(xing)(xing))。半監(jian)督式(shi)(shi)異(yi)(yi)常(chang)(chang)(chang)檢(jian)測(ce)(ce)(ce)(ce)方(fang)(fang)(fang)法(fa)(fa)(fa)(fa)根據(ju)(ju)一(yi)(yi)個給定的(de)(de)(de)正常(chang)(chang)(chang)訓練數(shu)據(ju)(ju)集(ji)(ji)創建一(yi)(yi)個表示正常(chang)(chang)(chang)行為的(de)(de)(de)模型(xing)(xing),然(ran)后檢(jian)測(ce)(ce)(ce)(ce)由(you)學(xue)習模型(xing)(xing)生成的(de)(de)(de)測(ce)(ce)(ce)(ce)試實例的(de)(de)(de)可能(neng)(neng)性(xing)(xing)。

While explainability is a desirable characteristic of increasingly complex black-box models, modern explanation methods have been shown to be inconsistent and contradictory. The semantics of explanations is not always fully understood - to what extent do explanations "explain" a decision and to what extent do they merely advocate for a decision? Can we help humans gain insights from explanations accompanying correct predictions and not over-rely on incorrect predictions advocated for by explanations? With this perspective in mind, we introduce the notion of dissenting explanations: conflicting predictions with accompanying explanations. We first explore the advantage of dissenting explanations in the setting of model multiplicity, where multiple models with similar performance may have different predictions. In such cases, providing dissenting explanations could be done by invoking the explanations of disagreeing models. Through a pilot study, we demonstrate that dissenting explanations reduce overreliance on model predictions, without reducing overall accuracy. Motivated by the utility of dissenting explanations we present both global and local methods for their generation.

Speech enhancement aims to improve speech quality and intelligibility, especially in noisy environments where background noise degrades speech signals. Currently, deep learning methods achieve great success in speech enhancement, e.g. the representative convolutional recurrent neural network (CRN) and its variants. However, CRN typically employs consecutive downsampling and upsampling convolution for frequency modeling, which destroys the inherent structure of the signal over frequency. Additionally, convolutional layers lacks of temporal modelling abilities. To address these issues, we propose an innovative module combing a State space model and Inplace Convolution (SIC), and to replace the conventional convolution in CRN, called SICRN. Specifically, a dual-path multidimensional State space model captures the global frequencies dependency and long-term temporal dependencies. Meanwhile, the 2D-inplace convolution is used to capture the local structure, which abandons the downsampling and upsampling. Systematic evaluations on the public INTERSPEECH 2020 DNS challenge dataset demonstrate SICRN's efficacy. Compared to strong baselines, SICRN achieves performance close to state-of-the-art while having advantages in model parameters, computations, and algorithmic delay. The proposed SICRN shows great promise for improved speech enhancement.

Although the security testing of Web systems can be automated by generating crafted inputs, solutions to automate the test oracle, i.e., distinguishing correct from incorrect outputs, remain preliminary. Specifically, previous work has demonstrated the potential of metamorphic testing; indeed, security failures can be determined by metamorphic relations that turn valid inputs into malicious inputs. However, without further guidance, metamorphic relations are typically executed on a large set of inputs, which is time-consuming and thus makes metamorphic testing impractical. We propose AIM, an approach that automatically selects inputs to reduce testing costs while preserving vulnerability detection capabilities. AIM includes a clustering-based black box approach, to identify similar inputs based on their security properties. It also relies on a novel genetic algorithm able to efficiently select diverse inputs while minimizing their total cost. Further, it contains a problem-reduction component to reduce the search space and speed up the minimization process. We evaluated the effectiveness of AIM on two well-known Web systems, Jenkins and Joomla, with documented vulnerabilities. We compared AIM's results with four baselines. Overall, AIM reduced metamorphic testing time by 84% for Jenkins and 82% for Joomla, while preserving vulnerability detection. Furthermore, AIM outperformed all the considered baselines regarding vulnerability coverage.

Object detection in documents is a key step to automate the structural elements identification process in a digital or scanned document through understanding the hierarchical structure and relationships between different elements. Large and complex models, while achieving high accuracy, can be computationally expensive and memory-intensive, making them impractical for deployment on resource constrained devices. Knowledge distillation allows us to create small and more efficient models that retain much of the performance of their larger counterparts. Here we present a graph-based knowledge distillation framework to correctly identify and localize the document objects in a document image. Here, we design a structured graph with nodes containing proposal-level features and edges representing the relationship between the different proposal regions. Also, to reduce text bias an adaptive node sampling strategy is designed to prune the weight distribution and put more weightage on non-text nodes. We encode the complete graph as a knowledge representation and transfer it from the teacher to the student through the proposed distillation loss by effectively capturing both local and global information concurrently. Extensive experimentation on competitive benchmarks demonstrates that the proposed framework outperforms the current state-of-the-art approaches. The code will be available at: //github.com/ayanban011/GraphKD.

Logging practices have been extensively investigated to assist developers in writing appropriate logging statements for documenting software behaviors. Although numerous automatic logging approaches have been proposed, their performance remains unsatisfactory due to the constraint of the single-method input, without informative programming context outside the method. Specifically, we identify three inherent limitations with single-method context: limited static scope of logging statements, inconsistent logging styles, and missing type information of logging variables. To tackle these limitations, we propose SCLogger, the first contextualized logging statement generation approach with inter-method static contexts. First, SCLogger extracts inter-method contexts with static analysis to construct the contextualized prompt for language models to generate a tentative logging statement. The contextualized prompt consists of an extended static scope and sampled similar methods, ordered by the chain-of-thought (COT) strategy. Second, SCLogger refines the access of logging variables by formulating a new refinement prompt for language models, which incorporates detailed type information of variables in the tentative logging statement. The evaluation results show that SCLogger surpasses the state-of-the-art approach by 8.7% in logging position accuracy, 32.1% in level accuracy, 19.6% in variable precision, and 138.4% in text BLEU-4 score. Furthermore, SCLogger consistently boosts the performance of logging statement generation across a range of large language models, thereby showcasing the generalizability of this approach.

Compared to typical multi-sensor systems, monocular 3D object detection has attracted much attention due to its simple configuration. However, there is still a significant gap between LiDAR-based and monocular-based methods. In this paper, we find that the ill-posed nature of monocular imagery can lead to depth ambiguity. Specifically, objects with different depths can appear with the same bounding boxes and similar visual features in the 2D image. Unfortunately, the network cannot accurately distinguish different depths from such non-discriminative visual features, resulting in unstable depth training. To facilitate depth learning, we propose a simple yet effective plug-and-play module, \underline{O}ne \underline{B}ounding Box \underline{M}ultiple \underline{O}bjects (OBMO). Concretely, we add a set of suitable pseudo labels by shifting the 3D bounding box along the viewing frustum. To constrain the pseudo-3D labels to be reasonable, we carefully design two label scoring strategies to represent their quality. In contrast to the original hard depth labels, such soft pseudo labels with quality scores allow the network to learn a reasonable depth range, boosting training stability and thus improving final performance. Extensive experiments on KITTI and Waymo benchmarks show that our method significantly improves state-of-the-art monocular 3D detectors by a significant margin (The improvements under the moderate setting on KITTI validation set are $\mathbf{1.82\sim 10.91\%}$ \textbf{mAP in BEV} and $\mathbf{1.18\sim 9.36\%}$ \textbf{mAP in 3D}). Codes have been released at \url{//github.com/mrsempress/OBMO}.

Oriented object detection has been rapidly developed in the past few years, but most of these methods assume the training and testing images are under the same statistical distribution, which is far from reality. In this paper, we propose the task of domain generalized oriented object detection, which intends to explore the generalization of oriented object detectors on arbitrary unseen target domains. Learning domain generalized oriented object detectors is particularly challenging, as the cross-domain style variation not only negatively impacts the content representation, but also leads to unreliable orientation predictions. To address these challenges, we propose a generalized oriented object detector (GOOD). After style hallucination by the emerging contrastive language-image pre-training (CLIP), it consists of two key components, namely, rotation-aware content consistency learning (RAC) and style consistency learning (SEC). The proposed RAC allows the oriented object detector to learn stable orientation representation from style-diversified samples. The proposed SEC further stabilizes the generalization ability of content representation from different image styles. Extensive experiments on multiple cross-domain settings show the state-of-the-art performance of GOOD. Source code will be publicly available.

Achieving a universally high accuracy in object detection is quite challenging, and the mainstream focus in the industry currently lies on detecting specific classes of objects. However, deploying one or multiple object detection networks requires a certain amount of GPU memory for training and storage capacity for inference. This presents challenges in terms of how to effectively coordinate multiple object detection tasks under resource-constrained conditions. This paper introduces a lightweight fine-tuning strategy called Calibration side tuning, which integrates aspects of adapter tuning and side tuning to adapt the successful techniques employed in transformers for use with ResNet. The Calibration side tuning architecture that incorporates maximal transition calibration, utilizing a small number of additional parameters to enhance network performance while maintaining a smooth training process. Furthermore, this paper has conducted an analysis on multiple fine-tuning strategies and have implemented their application within ResNet, thereby expanding the research on fine-tuning strategies for object detection networks. Besides, this paper carried out extensive experiments using five benchmark datasets. The experimental results demonstrated that this method outperforms other compared state-of-the-art techniques, and a better balance between the complexity and performance of the finetune schemes is achieved.

Interpretability methods are developed to understand the working mechanisms of black-box models, which is crucial to their responsible deployment. Fulfilling this goal requires both that the explanations generated by these methods are correct and that people can easily and reliably understand them. While the former has been addressed in prior work, the latter is often overlooked, resulting in informal model understanding derived from a handful of local explanations. In this paper, we introduce explanation summary (ExSum), a mathematical framework for quantifying model understanding, and propose metrics for its quality assessment. On two domains, ExSum highlights various limitations in the current practice, helps develop accurate model understanding, and reveals easily overlooked properties of the model. We also connect understandability to other properties of explanations such as human alignment, robustness, and counterfactual minimality and plausibility.

Explainable recommendation attempts to develop models that generate not only high-quality recommendations but also intuitive explanations. The explanations may either be post-hoc or directly come from an explainable model (also called interpretable or transparent model in some context). Explainable recommendation tries to address the problem of why: by providing explanations to users or system designers, it helps humans to understand why certain items are recommended by the algorithm, where the human can either be users or system designers. Explainable recommendation helps to improve the transparency, persuasiveness, effectiveness, trustworthiness, and satisfaction of recommendation systems. In this survey, we review works on explainable recommendation in or before the year of 2019. We first highlight the position of explainable recommendation in recommender system research by categorizing recommendation problems into the 5W, i.e., what, when, who, where, and why. We then conduct a comprehensive survey of explainable recommendation on three perspectives: 1) We provide a chronological research timeline of explainable recommendation, including user study approaches in the early years and more recent model-based approaches. 2) We provide a two-dimensional taxonomy to classify existing explainable recommendation research: one dimension is the information source (or display style) of the explanations, and the other dimension is the algorithmic mechanism to generate explainable recommendations. 3) We summarize how explainable recommendation applies to different recommendation tasks, such as product recommendation, social recommendation, and POI recommendation. We also devote a section to discuss the explanation perspectives in broader IR and AI/ML research. We end the survey by discussing potential future directions to promote the explainable recommendation research area and beyond.

北京阿比特科技有限公司