亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Anomaly detection in manufacturing pipelines remains a critical challenge, intensified by the complexity and variability of industrial environments. This paper introduces AssemAI, an interpretable image-based anomaly detection system tailored for smart manufacturing pipelines. Our primary contributions include the creation of a tailored image dataset and the development of a custom object detection model, YOLO-FF, designed explicitly for anomaly detection in manufacturing assembly environments. Utilizing the preprocessed image dataset derived from an industry-focused rocket assembly pipeline, we address the challenge of imbalanced image data and demonstrate the importance of image-based methods in anomaly detection. The proposed approach leverages domain knowledge in data preparation, model development and reasoning. We compare our method against several baselines, including simple CNN and custom Visual Transformer (ViT) models, showcasing the effectiveness of our custom data preparation and pretrained CNN integration. Additionally, we incorporate explainability techniques at both user and model levels, utilizing ontology for user-friendly explanations and SCORE-CAM for in-depth feature and model analysis. Finally, the model was also deployed in a real-time setting. Our results include ablation studies on the baselines, providing a comprehensive evaluation of the proposed system. This work highlights the broader impact of advanced image-based anomaly detection in enhancing the reliability and efficiency of smart manufacturing processes.

相關內容

在數(shu)(shu)(shu)據(ju)挖掘中(zhong),異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)檢(jian)測(ce)(ce)(英語:anomaly detection)對(dui)不符合預(yu)期模(mo)式(shi)(shi)或(huo)數(shu)(shu)(shu)據(ju)集(ji)(ji)(ji)中(zhong)其(qi)他(ta)項目(mu)的(de)(de)(de)(de)(de)(de)(de)項目(mu)、事(shi)件或(huo)觀測(ce)(ce)值的(de)(de)(de)(de)(de)(de)(de)識別。通常(chang)(chang)(chang)(chang)(chang)(chang)(chang)異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)項目(mu)會轉變(bian)成銀行(xing)欺詐、結構缺(que)陷(xian)、醫療問題(ti)、文本錯誤等類型的(de)(de)(de)(de)(de)(de)(de)問題(ti)。異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)也被稱為(wei)離群值、新奇、噪(zao)聲(sheng)、偏差(cha)和例(li)(li)外。 特(te)別是(shi)(shi)(shi)在檢(jian)測(ce)(ce)濫用(yong)與(yu)(yu)網絡(luo)入侵時,有(you)趣性(xing)(xing)(xing)對(dui)象(xiang)(xiang)往往不是(shi)(shi)(shi)罕見對(dui)象(xiang)(xiang),但卻是(shi)(shi)(shi)超出(chu)預(yu)料的(de)(de)(de)(de)(de)(de)(de)突發活動(dong)。這種模(mo)式(shi)(shi)不遵循通常(chang)(chang)(chang)(chang)(chang)(chang)(chang)統計定(ding)(ding)義(yi)中(zhong)把異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)點(dian)看作是(shi)(shi)(shi)罕見對(dui)象(xiang)(xiang),于是(shi)(shi)(shi)許(xu)多異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)檢(jian)測(ce)(ce)方法(特(te)別是(shi)(shi)(shi)無監(jian)督(du)的(de)(de)(de)(de)(de)(de)(de)方法)將對(dui)此類數(shu)(shu)(shu)據(ju)失效,除非進行(xing)了合適的(de)(de)(de)(de)(de)(de)(de)聚集(ji)(ji)(ji)。相反,聚類分析算法可(ke)能可(ke)以檢(jian)測(ce)(ce)出(chu)這些模(mo)式(shi)(shi)形成的(de)(de)(de)(de)(de)(de)(de)微聚類。 有(you)三大類異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)檢(jian)測(ce)(ce)方法。[1] 在假設數(shu)(shu)(shu)據(ju)集(ji)(ji)(ji)中(zhong)大多數(shu)(shu)(shu)實(shi)例(li)(li)都是(shi)(shi)(shi)正(zheng)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)的(de)(de)(de)(de)(de)(de)(de)前提(ti)下,無監(jian)督(du)異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)檢(jian)測(ce)(ce)方法能通過尋找與(yu)(yu)其(qi)他(ta)數(shu)(shu)(shu)據(ju)最不匹(pi)配的(de)(de)(de)(de)(de)(de)(de)實(shi)例(li)(li)來檢(jian)測(ce)(ce)出(chu)未標(biao)記測(ce)(ce)試(shi)數(shu)(shu)(shu)據(ju)的(de)(de)(de)(de)(de)(de)(de)異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)。監(jian)督(du)式(shi)(shi)異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)檢(jian)測(ce)(ce)方法需要一(yi)(yi)個(ge)已經被標(biao)記“正(zheng)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)”與(yu)(yu)“異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)”的(de)(de)(de)(de)(de)(de)(de)數(shu)(shu)(shu)據(ju)集(ji)(ji)(ji),并涉及(ji)到訓練分類器(與(yu)(yu)許(xu)多其(qi)他(ta)的(de)(de)(de)(de)(de)(de)(de)統計分類問題(ti)的(de)(de)(de)(de)(de)(de)(de)關(guan)鍵區別是(shi)(shi)(shi)異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)檢(jian)測(ce)(ce)的(de)(de)(de)(de)(de)(de)(de)內在不均衡(heng)性(xing)(xing)(xing))。半監(jian)督(du)式(shi)(shi)異(yi)(yi)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)檢(jian)測(ce)(ce)方法根據(ju)一(yi)(yi)個(ge)給(gei)定(ding)(ding)的(de)(de)(de)(de)(de)(de)(de)正(zheng)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)訓練數(shu)(shu)(shu)據(ju)集(ji)(ji)(ji)創建一(yi)(yi)個(ge)表示正(zheng)常(chang)(chang)(chang)(chang)(chang)(chang)(chang)行(xing)為(wei)的(de)(de)(de)(de)(de)(de)(de)模(mo)型,然(ran)后(hou)檢(jian)測(ce)(ce)由學習模(mo)型生成的(de)(de)(de)(de)(de)(de)(de)測(ce)(ce)試(shi)實(shi)例(li)(li)的(de)(de)(de)(de)(de)(de)(de)可(ke)能性(xing)(xing)(xing)。

Existing benchmarks for summarization quality evaluation often lack diverse input scenarios, focus on narrowly defined dimensions (e.g., faithfulness), and struggle with subjective and coarse-grained annotation schemes. To address these shortcomings, we create UniSumEval benchmark, which extends the range of input context (e.g., domain, length) and provides fine-grained, multi-dimensional annotations. We use AI assistance in data creation, identifying potentially hallucinogenic input texts, and also helping human annotators reduce the difficulty of fine-grained annotation tasks. With UniSumEval, we benchmark nine latest language models as summarizers, offering insights into their performance across varying input contexts and evaluation dimensions. Furthermore, we conduct a thorough comparison of SOTA automated summary evaluators. Our benchmark data will be available at //github.com/DISL-Lab/UniSumEval-v1.0.

Dealing with atypical traffic scenarios remains a challenging task in autonomous driving. However, most anomaly detection approaches cannot be trained on raw sensor data but require exposure to outlier data and powerful semantic segmentation models trained in a supervised fashion. This limits the representation of normality to labeled data, which does not scale well. In this work, we revisit unsupervised anomaly detection and present UMAD, leveraging generative world models and unsupervised image segmentation. Our method outperforms state-of-the-art unsupervised anomaly detection.

Diffusion models have achieved remarkable success in Text-to-Image generation tasks, leading to the development of many commercial models. However, recent studies have reported that diffusion models often generate replicated images in train data when triggered by specific prompts, potentially raising social issues ranging from copyright to privacy concerns. To sidestep the memorization, there have been recent studies for developing memorization mitigation methods for diffusion models. Nevertheless, the lack of benchmarks impedes the assessment of the true effectiveness of these methods. In this work, we present MemBench, the first benchmark for evaluating image memorization mitigation methods. Our benchmark includes a large number of memorized image trigger prompts in various Text-to-Image diffusion models. Furthermore, in contrast to the prior work evaluating mitigation performance only on trigger prompts, we present metrics evaluating on both trigger prompts and general prompts, so that we can see whether mitigation methods address the memorization issue while maintaining performance for general prompts. This is an important development considering the practical applications which previous works have overlooked. Through evaluation on MemBench, we verify that the performance of existing image memorization mitigation methods is still insufficient for application to diffusion models. The code and datasets are available at //github.com/chunsanHong/MemBench\_code.

Operating effectively in complex environments while complying with specified constraints is crucial for the safe and successful deployment of robots that interact with and operate around people. In this work, we focus on generating long-horizon trajectories that adhere to novel static and temporally-extended constraints/instructions at test time. We propose a data-driven diffusion-based framework, LTLDoG, that modifies the inference steps of the reverse process given an instruction specified using finite linear temporal logic ($\text{LTL}_f$). LTLDoG leverages a satisfaction value function on $\text{LTL}_f$ and guides the sampling steps using its gradient field. This value function can also be trained to generalize to new instructions not observed during training, enabling flexible test-time adaptability. Experiments in robot navigation and manipulation illustrate that the method is able to generate trajectories that satisfy formulae that specify obstacle avoidance and visitation sequences. Code and supplementary material are available online at //github.com/clear-nus/ltldog.

Phishing attacks are a growing cybersecurity threat, leveraging deceptive techniques to steal sensitive information through malicious websites. To combat these attacks, this paper introduces PhishGuard, an optimal custom ensemble model designed to improve phishing site detection. The model combines multiple machine learning classifiers, including Random Forest, Gradient Boosting, CatBoost, and XGBoost, to enhance detection accuracy. Through advanced feature selection methods such as SelectKBest and RFECV, and optimizations like hyperparameter tuning and data balancing, the model was trained and evaluated on four publicly available datasets. PhishGuard outperformed state-of-the-art models, achieving a detection accuracy of 99.05% on one of the datasets, with similarly high results across other datasets. This research demonstrates that optimization methods in conjunction with ensemble learning greatly improve phishing detection performance.

Constrained by the lack of model interpretability and a deep understanding of human movement in traditional movement recognition machine learning methods, this study introduces a novel representation learning method based on causal inference to better understand human joint dynamics and complex behaviors. We propose a two-stage framework that combines the Peter-Clark (PC) algorithm and Kullback-Leibler (KL) divergence to identify and quantify causal relationships between joints. Our method effectively captures interactions and produces interpretable, robust representations. Experiments on the EmoPain dataset show that our causal GCN outperforms traditional GCNs in accuracy, F1 score, and recall, especially in detecting protective behaviors. The model is also highly invariant to data scale changes, enhancing its reliability in practical applications. Our approach advances human motion analysis and paves the way for more adaptive intelligent healthcare solutions.

With the rapid development of IT operations, it has become increasingly crucial to efficiently manage and analyze large volumes of data for practical applications. The techniques of Natural Language Processing (NLP) have shown remarkable capabilities for various tasks, including named entity recognition, machine translation and dialogue systems. Recently, Large Language Models (LLMs) have achieved significant improvements across various NLP downstream tasks. However, there is a lack of specialized LLMs for IT operations. In this paper, we introduce the OWL, a large language model trained on our collected OWL-Instruct dataset with a wide range of IT-related information, where the mixture-of-adapter strategy is proposed to improve the parameter-efficient tuning across different domains or tasks. Furthermore, we evaluate the performance of our OWL on the OWL-Bench established by us and open IT-related benchmarks. OWL demonstrates superior performance results on IT tasks, which outperforms existing models by significant margins. Moreover, we hope that the findings of our work will provide more insights to revolutionize the techniques of IT operations with specialized LLMs.

Generative commonsense reasoning which aims to empower machines to generate sentences with the capacity of reasoning over a set of concepts is a critical bottleneck for text generation. Even the state-of-the-art pre-trained language generation models struggle at this task and often produce implausible and anomalous sentences. One reason is that they rarely consider incorporating the knowledge graph which can provide rich relational information among the commonsense concepts. To promote the ability of commonsense reasoning for text generation, we propose a novel knowledge graph augmented pre-trained language generation model KG-BART, which encompasses the complex relations of concepts through the knowledge graph and produces more logical and natural sentences as output. Moreover, KG-BART can leverage the graph attention to aggregate the rich concept semantics that enhances the model generalization on unseen concept sets. Experiments on benchmark CommonGen dataset verify the effectiveness of our proposed approach by comparing with several strong pre-trained language generation models, particularly KG-BART outperforms BART by 5.80, 4.60, in terms of BLEU-3, 4. Moreover, we also show that the generated context by our model can work as background scenarios to benefit downstream commonsense QA tasks.

With the capability of modeling bidirectional contexts, denoising autoencoding based pretraining like BERT achieves better performance than pretraining approaches based on autoregressive language modeling. However, relying on corrupting the input with masks, BERT neglects dependency between the masked positions and suffers from a pretrain-finetune discrepancy. In light of these pros and cons, we propose XLNet, a generalized autoregressive pretraining method that (1) enables learning bidirectional contexts by maximizing the expected likelihood over all permutations of the factorization order and (2) overcomes the limitations of BERT thanks to its autoregressive formulation. Furthermore, XLNet integrates ideas from Transformer-XL, the state-of-the-art autoregressive model, into pretraining. Empirically, XLNet outperforms BERT on 20 tasks, often by a large margin, and achieves state-of-the-art results on 18 tasks including question answering, natural language inference, sentiment analysis, and document ranking.

The cross-domain recommendation technique is an effective way of alleviating the data sparsity in recommender systems by leveraging the knowledge from relevant domains. Transfer learning is a class of algorithms underlying these techniques. In this paper, we propose a novel transfer learning approach for cross-domain recommendation by using neural networks as the base model. We assume that hidden layers in two base networks are connected by cross mappings, leading to the collaborative cross networks (CoNet). CoNet enables dual knowledge transfer across domains by introducing cross connections from one base network to another and vice versa. CoNet is achieved in multi-layer feedforward networks by adding dual connections and joint loss functions, which can be trained efficiently by back-propagation. The proposed model is evaluated on two real-world datasets and it outperforms baseline models by relative improvements of 3.56\% in MRR and 8.94\% in NDCG, respectively.

北京阿比特科技有限公司