亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

Training semantic segmentation models on multiple datasets has sparked a lot of recent interest in the computer vision community. This interest has been motivated by expensive annotations and a desire to achieve proficiency across multiple visual domains. However, established datasets have mutually incompatible labels which disrupt principled inference in the wild. We address this issue by automatic construction of universal taxonomies through iterative dataset integration. Our method detects subset-superset relationships between dataset-specific labels, and supports learning of sub-class logits by treating super-classes as partial labels. We present experiments on collections of standard datasets and demonstrate competitive generalization performance with respect to previous work.

相關內容

分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)學(xue)是(shi)分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)的(de)(de)實踐和(he)(he)科學(xue)。Wikipedia類(lei)(lei)(lei)別說(shuo)明了(le)一(yi)種分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa),可(ke)以(yi)(yi)通過自動方式(shi)提取Wikipedia類(lei)(lei)(lei)別的(de)(de)完整分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)。截(jie)至2009年,已經證明,可(ke)以(yi)(yi)使用(yong)人工構(gou)(gou)(gou)建的(de)(de)分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)(例(li)如(ru)(ru)像WordNet這(zhe)樣(yang)的(de)(de)計算詞典的(de)(de)分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa))來(lai)改(gai)進和(he)(he)重組(zu)Wikipedia類(lei)(lei)(lei)別分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)。 從廣(guang)義上講(jiang),分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)還(huan)適(shi)(shi)(shi)用(yong)于除父子層(ceng)次(ci)結(jie)構(gou)(gou)(gou)以(yi)(yi)外(wai)的(de)(de)關系方案(an),例(li)如(ru)(ru)網絡結(jie)構(gou)(gou)(gou)。然(ran)后分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)可(ke)能包括有多(duo)父母的(de)(de)單身(shen)孩子,例(li)如(ru)(ru),“汽車”可(ke)能與(yu)父母雙(shuang)方一(yi)起出現“車輛”和(he)(he)“鋼結(jie)構(gou)(gou)(gou)”;但(dan)(dan)是(shi)對某些人而(er)言,這(zhe)僅意味著(zhu)“汽車”是(shi)幾種不同分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)的(de)(de)一(yi)部分(fen)(fen)(fen)(fen)(fen)(fen)(fen)。分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)也可(ke)能只是(shi)將事物(wu)組(zu)織成組(zu),或者是(shi)按字母順(shun)序(xu)排列(lie)的(de)(de)列(lie)表(biao);但(dan)(dan)是(shi)在這(zhe)里(li),術語(yu)詞匯更(geng)合適(shi)(shi)(shi)。在知(zhi)識(shi)管(guan)理(li)中的(de)(de)當前用(yong)法(fa)(fa)(fa)(fa)中,分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)被(bei)認為比本體(ti)論(lun)窄,因為本體(ti)論(lun)應(ying)用(yong)了(le)各(ge)種各(ge)樣(yang)的(de)(de)關系類(lei)(lei)(lei)型。 在數學(xue)上,分(fen)(fen)(fen)(fen)(fen)(fen)(fen)層(ceng)分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)法(fa)(fa)(fa)(fa)是(shi)給定對象集(ji)的(de)(de)分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)樹結(jie)構(gou)(gou)(gou)。該(gai)結(jie)構(gou)(gou)(gou)的(de)(de)頂部是(shi)適(shi)(shi)(shi)用(yong)于所有對象的(de)(de)單個分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei),即(ji)根節(jie)點(dian)。此根下的(de)(de)節(jie)點(dian)是(shi)更(geng)具(ju)體(ti)的(de)(de)分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei),適(shi)(shi)(shi)用(yong)于總(zong)分(fen)(fen)(fen)(fen)(fen)(fen)(fen)類(lei)(lei)(lei)對象集(ji)的(de)(de)子集(ji)。推理(li)的(de)(de)進展從一(yi)般(ban)到更(geng)具(ju)體(ti)。

知識薈萃

精品入門和進階教程、論文(wen)和代(dai)碼整理等

更多

查看相關VIP內容、論文、資訊(xun)等

Automated medical image segmentation using deep neural networks typically requires substantial supervised training. However, these models fail to generalize well across different imaging modalities. This shortcoming, amplified by the limited availability of annotated data, has been hampering the deployment of such methods at a larger scale across modalities. To address these issues, we propose M-GenSeg, a new semi-supervised training strategy for accurate cross-modality tumor segmentation on unpaired bi-modal datasets. Based on image-level labels, a first unsupervised objective encourages the model to perform diseased to healthy translation by disentangling tumors from the background, which encompasses the segmentation task. Then, teaching the model to translate between image modalities enables the synthesis of target images from a source modality, thus leveraging the pixel-level annotations from the source modality to enforce generalization to the target modality images. We evaluated the performance on a brain tumor segmentation datasets composed of four different contrast sequences from the public BraTS 2020 challenge dataset. We report consistent improvement in Dice scores on both source and unannotated target modalities. On all twelve distinct domain adaptation experiments, the proposed model shows a clear improvement over state-of-the-art domain-adaptive baselines, with absolute Dice gains on the target modality reaching 0.15.

We introduce MegaPose, a method to estimate the 6D pose of novel objects, that is, objects unseen during training. At inference time, the method only assumes knowledge of (i) a region of interest displaying the object in the image and (ii) a CAD model of the observed object. The contributions of this work are threefold. First, we present a 6D pose refiner based on a render&compare strategy which can be applied to novel objects. The shape and coordinate system of the novel object are provided as inputs to the network by rendering multiple synthetic views of the object's CAD model. Second, we introduce a novel approach for coarse pose estimation which leverages a network trained to classify whether the pose error between a synthetic rendering and an observed image of the same object can be corrected by the refiner. Third, we introduce a large-scale synthetic dataset of photorealistic images of thousands of objects with diverse visual and shape properties and show that this diversity is crucial to obtain good generalization performance on novel objects. We train our approach on this large synthetic dataset and apply it without retraining to hundreds of novel objects in real images from several pose estimation benchmarks. Our approach achieves state-of-the-art performance on the ModelNet and YCB-Video datasets. An extensive evaluation on the 7 core datasets of the BOP challenge demonstrates that our approach achieves performance competitive with existing approaches that require access to the target objects during training. Code, dataset and trained models are available on the project page: //megapose6d.github.io/.

Modern deep learning requires large-scale extensively labelled datasets for training. Few-shot learning aims to alleviate this issue by learning effectively from few labelled examples. In previously proposed few-shot visual classifiers, it is assumed that the feature manifold, where classifier decisions are made, has uncorrelated feature dimensions and uniform feature variance. In this work, we focus on addressing the limitations arising from this assumption by proposing a variance-sensitive class of models that operates in a low-label regime. The first method, Simple CNAPS, employs a hierarchically regularized Mahalanobis-distance based classifier combined with a state of the art neural adaptive feature extractor to achieve strong performance on Meta-Dataset, mini-ImageNet and tiered-ImageNet benchmarks. We further extend this approach to a transductive learning setting, proposing Transductive CNAPS. This transductive method combines a soft k-means parameter refinement procedure with a two-step task encoder to achieve improved test-time classification accuracy using unlabelled data. Transductive CNAPS achieves state of the art performance on Meta-Dataset. Finally, we explore the use of our methods (Simple and Transductive) for "out of the box" continual and active learning. Extensive experiments on large scale benchmarks illustrate robustness and versatility of this, relatively speaking, simple class of models. All trained model checkpoints and corresponding source codes have been made publicly available.

Semantic segmentation in 3D indoor scenes has achieved remarkable performance under the supervision of large-scale annotated data. However, previous works rely on the assumption that the training and testing data are of the same distribution, which may suffer from performance degradation when evaluated on the out-of-distribution scenes. To alleviate the annotation cost and the performance degradation, this paper introduces the synthetic-to-real domain generalization setting to this task. Specifically, the domain gap between synthetic and real-world point cloud data mainly lies in the different layouts and point patterns. To address these problems, we first propose a clustering instance mix (CINMix) augmentation technique to diversify the layouts of the source data. In addition, we augment the point patterns of the source data and introduce non-parametric multi-prototypes to ameliorate the intra-class variance enlarged by the augmented point patterns. The multi-prototypes can model the intra-class variance and rectify the global classifier in both training and inference stages. Experiments on the synthetic-to-real benchmark demonstrate that both CINMix and multi-prototypes can narrow the distribution gap and thus improve the generalization ability on real-world datasets.

Rehearsal, seeking to remind the model by storing old knowledge in lifelong learning, is one of the most effective ways to mitigate catastrophic forgetting, i.e., biased forgetting of previous knowledge when moving to new tasks. However, the old tasks of the most previous rehearsal-based methods suffer from the unpredictable domain shift when training the new task. This is because these methods always ignore two significant factors. First, the Data Imbalance between the new task and old tasks that makes the domain of old tasks prone to shift. Second, the Task Isolation among all tasks will make the domain shift toward unpredictable directions; To address the unpredictable domain shift, in this paper, we propose Multi-Domain Multi-Task (MDMT) rehearsal to train the old tasks and new task parallelly and equally to break the isolation among tasks. Specifically, a two-level angular margin loss is proposed to encourage the intra-class/task compactness and inter-class/task discrepancy, which keeps the model from domain chaos. In addition, to further address domain shift of the old tasks, we propose an optional episodic distillation loss on the memory to anchor the knowledge for each old task. Experiments on benchmark datasets validate the proposed approach can effectively mitigate the unpredictable domain shift.

Deep learning-based semi-supervised learning (SSL) algorithms have led to promising results in medical images segmentation and can alleviate doctors' expensive annotations by leveraging unlabeled data. However, most of the existing SSL algorithms in literature tend to regularize the model training by perturbing networks and/or data. Observing that multi/dual-task learning attends to various levels of information which have inherent prediction perturbation, we ask the question in this work: can we explicitly build task-level regularization rather than implicitly constructing networks- and/or data-level perturbation-and-transformation for SSL? To answer this question, we propose a novel dual-task-consistency semi-supervised framework for the first time. Concretely, we use a dual-task deep network that jointly predicts a pixel-wise segmentation map and a geometry-aware level set representation of the target. The level set representation is converted to an approximated segmentation map through a differentiable task transform layer. Simultaneously, we introduce a dual-task consistency regularization between the level set-derived segmentation maps and directly predicted segmentation maps for both labeled and unlabeled data. Extensive experiments on two public datasets show that our method can largely improve the performance by incorporating the unlabeled data. Meanwhile, our framework outperforms the state-of-the-art semi-supervised medical image segmentation methods. Code is available at: //github.com/Luoxd1996/DTC

This paper aims at revisiting Graph Convolutional Neural Networks by bridging the gap between spectral and spatial design of graph convolutions. We theoretically demonstrate some equivalence of the graph convolution process regardless it is designed in the spatial or the spectral domain. The obtained general framework allows to lead a spectral analysis of the most popular ConvGNNs, explaining their performance and showing their limits. Moreover, the proposed framework is used to design new convolutions in spectral domain with a custom frequency profile while applying them in the spatial domain. We also propose a generalization of the depthwise separable convolution framework for graph convolutional networks, what allows to decrease the total number of trainable parameters by keeping the capacity of the model. To the best of our knowledge, such a framework has never been used in the GNNs literature. Our proposals are evaluated on both transductive and inductive graph learning problems. Obtained results show the relevance of the proposed method and provide one of the first experimental evidence of transferability of spectral filter coefficients from one graph to another. Our source codes are publicly available at: //github.com/balcilar/Spectral-Designed-Graph-Convolutions

Embedding entities and relations into a continuous multi-dimensional vector space have become the dominant method for knowledge graph embedding in representation learning. However, most existing models ignore to represent hierarchical knowledge, such as the similarities and dissimilarities of entities in one domain. We proposed to learn a Domain Representations over existing knowledge graph embedding models, such that entities that have similar attributes are organized into the same domain. Such hierarchical knowledge of domains can give further evidence in link prediction. Experimental results show that domain embeddings give a significant improvement over the most recent state-of-art baseline knowledge graph embedding models.

Deep Convolutional Neural Networks have pushed the state-of-the art for semantic segmentation provided that a large amount of images together with pixel-wise annotations is available. Data collection is expensive and a solution to alleviate it is to use transfer learning. This reduces the amount of annotated data required for the network training but it does not get rid of this heavy processing step. We propose a method of transfer learning without annotations on the target task for datasets with redundant content and distinct pixel distributions. Our method takes advantage of the approximate content alignment of the images between two datasets when the approximation error prevents the reuse of annotation from one dataset to another. Given the annotations for only one dataset, we train a first network in a supervised manner. This network autonomously learns to generate deep data representations relevant to the semantic segmentation. Then the images in the new dataset, we train a new network to generate a deep data representation that matches the one from the first network on the previous dataset. The training consists in a regression between feature maps and does not require any annotations on the new dataset. We show that this method reaches performances similar to a classic transfer learning on the PASCAL VOC dataset with synthetic transformations.

Image segmentation is considered to be one of the critical tasks in hyperspectral remote sensing image processing. Recently, convolutional neural network (CNN) has established itself as a powerful model in segmentation and classification by demonstrating excellent performances. The use of a graphical model such as a conditional random field (CRF) contributes further in capturing contextual information and thus improving the segmentation performance. In this paper, we propose a method to segment hyperspectral images by considering both spectral and spatial information via a combined framework consisting of CNN and CRF. We use multiple spectral cubes to learn deep features using CNN, and then formulate deep CRF with CNN-based unary and pairwise potential functions to effectively extract the semantic correlations between patches consisting of three-dimensional data cubes. Effective piecewise training is applied in order to avoid the computationally expensive iterative CRF inference. Furthermore, we introduce a deep deconvolution network that improves the segmentation masks. We also introduce a new dataset and experimented our proposed method on it along with several widely adopted benchmark datasets to evaluate the effectiveness of our method. By comparing our results with those from several state-of-the-art models, we show the promising potential of our method.

北京阿比特科技有限公司