亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

An increasing amount of applications rely on data-driven models that are deployed for perception tasks across a sequence of scenes. Due to the mismatch between training and deployment data, adapting the model on the new scenes is often crucial to obtain good performance. In this work, we study continual multi-scene adaptation for the task of semantic segmentation, assuming that no ground-truth labels are available during deployment and that performance on the previous scenes should be maintained. We propose training a Semantic-NeRF network for each scene by fusing the predictions of a segmentation model and then using the view-consistent rendered semantic labels as pseudo-labels to adapt the model. Through joint training with the segmentation model, the Semantic-NeRF model effectively enables 2D-3D knowledge transfer. Furthermore, due to its compact size, it can be stored in a long-term memory and subsequently used to render data from arbitrary viewpoints to reduce forgetting. We evaluate our approach on ScanNet, where we outperform both a voxel-based baseline and a state-of-the-art unsupervised domain adaptation method.

相關內容

Privacy policies provide individuals with information about their rights and how their personal information is handled. Natural language understanding (NLU) technologies can support individuals and practitioners to understand better privacy practices described in lengthy and complex documents. However, existing efforts that use NLU technologies are limited by processing the language in a way exclusive to a single task focusing on certain privacy practices. To this end, we introduce the Privacy Policy Language Understanding Evaluation (PLUE) benchmark, a multi-task benchmark for evaluating the privacy policy language understanding across various tasks. We also collect a large corpus of privacy policies to enable privacy policy domain-specific language model pre-training. We evaluate several generic pre-trained language models and continue pre-training them on the collected corpus. We demonstrate that domain-specific continual pre-training offers performance improvements across all tasks.

Automatically generating high-quality real world 3D scenes is of enormous interest for applications such as virtual reality and robotics simulation. Towards this goal, we introduce NeuralField-LDM, a generative model capable of synthesizing complex 3D environments. We leverage Latent Diffusion Models that have been successfully utilized for efficient high-quality 2D content creation. We first train a scene auto-encoder to express a set of image and pose pairs as a neural field, represented as density and feature voxel grids that can be projected to produce novel views of the scene. To further compress this representation, we train a latent-autoencoder that maps the voxel grids to a set of latent representations. A hierarchical diffusion model is then fit to the latents to complete the scene generation pipeline. We achieve a substantial improvement over existing state-of-the-art scene generation models. Additionally, we show how NeuralField-LDM can be used for a variety of 3D content creation applications, including conditional scene generation, scene inpainting and scene style manipulation.

當前計算機視覺中流行的骨干網絡,如視覺transformer (ViT)和ResNets,經過訓練可以從2D圖像中感知世界。為更有效地理解2D骨干中的3D結構先驗,本文提出Mask3D,在自監督預訓練中利用現有的大規模RGB-D數據,將這些3D先驗嵌入到2D學習的特征表示中.與需要3D重建或多視圖對應的傳統3D對比學習范式相比,所提出方法很簡單:通過屏蔽單個RGB- D幀中的RGB和深度補丁來制定前文本重建任務。Mask3D在將3D先驗嵌入到強大的2D ViT主干中特別有效,能對各種場景理解任務進行改進的表示學習,如語義分割、實例分割和目標檢測。實驗表明,Mask3D在ScanNet、NYUv2和Cityscapes圖像理解任務上明顯優于現有的自監督3D預訓練方法,在ScanNet圖像語義分割上比最先進的Pri3D提高了+6.5% mIoU。

//www.zhuanzhi.ai/paper/2cc5e9e67bcbea75082fac9489f2e2a4

付費5元查看完整內容

Over the past decade, domain adaptation has become a widely studied branch of transfer learning that aims to improve performance on target domains by leveraging knowledge from the source domain. Conventional domain adaptation methods often assume access to both source and target domain data simultaneously, which may not be feasible in real-world scenarios due to privacy and confidentiality concerns. As a result, the research of Source-Free Domain Adaptation (SFDA) has drawn growing attention in recent years, which only utilizes the source-trained model and unlabeled target data to adapt to the target domain. Despite the rapid explosion of SFDA work, yet there has no timely and comprehensive survey in the field. To fill this gap, we provide a comprehensive survey of recent advances in SFDA and organize them into a unified categorization scheme based on the framework of transfer learning. Instead of presenting each approach independently, we modularize several components of each method to more clearly illustrate their relationships and mechanics in light of the composite properties of each method. Furthermore, we compare the results of more than 30 representative SFDA methods on three popular classification benchmarks, namely Office-31, Office-home, and VisDA, to explore the effectiveness of various technical routes and the combination effects among them. Additionally, we briefly introduce the applications of SFDA and related fields. Drawing from our analysis of the challenges facing SFDA, we offer some insights into future research directions and potential settings.

題目: Diverse Image Generation via Self-Conditioned GANs

摘要:

本文介紹了一個簡單但有效的無監督方法,以產生現實和多樣化的圖像,并且訓練了一個類條件GAN模型,而不使用手動注釋的類標簽。相反,模型的條件是標簽自動聚類在鑒別器的特征空間。集群步驟自動發現不同的模式,并顯式地要求生成器覆蓋它們。在標準模式基準測試上的實驗表明,該方法在尋址模式崩潰時優于其他幾種競爭的方法。并且該方法在ImageNet和Places365這樣的大規模數據集上也有很好的表現,與以前的方法相比,提高了圖像多樣性和標準質量指標。

付費5元查看完整內容

題目: Continual Learning of Object Instances

摘要: 我們建議實例持續學習——一種將持續學習的概念應用于區分相同對象類別的實例的任務的方法。我們特別關注car對象,并通過度量學習逐步學會區分car實例與其他實例。我們從評估當前的技術開始我們的論文。在現有的方法中,災難性遺忘是顯而易見的,我們提出了兩個補救措施。首先,通過歸一化交叉熵對度量學習進行正則化。其次,我們使用合成數據傳輸來擴充現有的模型。我們在三個大型數據集上進行了大量的實驗,使用了兩種不同的體系結構,采用了五種不同的持續學習方法,結果表明,標準化的交叉熵和合成轉移可以減少現有技術中的遺忘。

付費5元查看完整內容

The rapid advancements in machine learning, graphics processing technologies and availability of medical imaging data has led to a rapid increase in use of machine learning models in the medical domain. This was exacerbated by the rapid advancements in convolutional neural network (CNN) based architectures, which were adopted by the medical imaging community to assist clinicians in disease diagnosis. Since the grand success of AlexNet in 2012, CNNs have been increasingly used in medical image analysis to improve the efficiency of human clinicians. In recent years, three-dimensional (3D) CNNs have been employed for analysis of medical images. In this paper, we trace the history of how the 3D CNN was developed from its machine learning roots, brief mathematical description of 3D CNN and the preprocessing steps required for medical images before feeding them to 3D CNNs. We review the significant research in the field of 3D medical imaging analysis using 3D CNNs (and its variants) in different medical areas such as classification, segmentation, detection, and localization. We conclude by discussing the challenges associated with the use of 3D CNNs in the medical imaging domain (and the use of deep learning models, in general) and possible future trends in the field.

題目: Correspondence Networks with Adaptive Neighbourhood Consensus

摘要:

在這篇論文中,我們處理的任務是在包含相同類別物體的圖像之間建立緊密的視覺對應。這是一個具有挑戰性的任務,因為類內部變化很大,并且缺乏密集的像素級注釋。我們提出了一種卷積神經網絡結構,稱為鄰域自適應一致網絡(ANC-Net),它可以通過稀疏的關鍵點注釋進行端到端的訓練來應對這一挑戰。該算法的核心是非各向同性的四維卷積核,構成了鄰域自適應一致的魯棒匹配模塊。為了使所學習的特征對類內變化具有魯棒性,我們還引入了一個簡單有效的多尺度自相似模型。此外,我們提出了一種新的正交損失來加強一對一匹配的約束。我們徹底地評估了我們的方法在各種基準上的有效性,在這些基準上,它的性能大大優于最先進的方法。

付費5元查看完整內容

題目: Data Augmentation using Pre-trained Transformer Models

簡介:

基于語言模型的預訓練模型,如BERT,在不同的NLP任務中提供了顯著的收益。在本文中,我們研究了不同類型的基于自回歸模型(GPT-2)、自編碼器模型(BERT)和seq2seq模型(BART)等用于條件數據增強的預訓練變壓器模型。我們表明,將類標簽前置到文本序列提供了一種簡單而有效的方法來設置預訓練模型的條件,以便進行數據擴充。在三個分類基準上,預先訓練的Seq2Seq模型優于其他模型。此外,我們還探討了不同的基于預訓練模型的數據擴充在數據多樣性方面是如何不同的,以及這些方法如何很好地保存類標簽信息。

付費5元查看完整內容
北京阿比特科技有限公司