亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

We study the impact of neural networks in text classification. Our focus is on training deep neural networks with proper weight initialization and greedy layer-wise pretraining. Results are compared with 1-layer neural networks and Support Vector Machines. We work with a dataset of labeled messages from the Twitter microblogging service and aim to predict weather conditions. A feature extraction procedure specific for the task is proposed, which applies dimensionality reduction using Latent Semantic Analysis. Our results show that neural networks outperform Support Vector Machines with Gaussian kernels, noticing performance gains from introducing additional hidden layers with nonlinearities. The impact of using Nesterov's Accelerated Gradient in backpropagation is also studied. We conclude that deep neural networks are a reasonable approach for text classification and propose further ideas to improve performance.

相關內容

神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)(Neural Networks)是世(shi)界上三個(ge)最古老的(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)建(jian)模學(xue)(xue)(xue)(xue)會的(de)(de)(de)檔案期刊:國際神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)學(xue)(xue)(xue)(xue)會(INNS)、歐洲神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)學(xue)(xue)(xue)(xue)會(ENNS)和(he)日本神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)學(xue)(xue)(xue)(xue)會(JNNS)。神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)提供了一個(ge)論壇,以發(fa)(fa)展和(he)培育一個(ge)國際社(she)(she)會的(de)(de)(de)學(xue)(xue)(xue)(xue)者和(he)實踐者感興趣(qu)的(de)(de)(de)所有方面的(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)和(he)相(xiang)關方法的(de)(de)(de)計算智能。神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)歡迎(ying)高質量(liang)論文(wen)的(de)(de)(de)提交(jiao)(jiao),有助于全面的(de)(de)(de)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)研究,從行(xing)為(wei)和(he)大(da)腦(nao)建(jian)模,學(xue)(xue)(xue)(xue)習(xi)算法,通過數(shu)學(xue)(xue)(xue)(xue)和(he)計算分(fen)析,系統的(de)(de)(de)工程和(he)技(ji)術應用(yong),大(da)量(liang)使(shi)用(yong)神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)的(de)(de)(de)概念和(he)技(ji)術。這一獨特而廣泛的(de)(de)(de)范圍(wei)促(cu)進(jin)了生物和(he)技(ji)術研究之間(jian)的(de)(de)(de)思(si)想交(jiao)(jiao)流,并(bing)有助于促(cu)進(jin)對生物啟(qi)發(fa)(fa)的(de)(de)(de)計算智能感興趣(qu)的(de)(de)(de)跨學(xue)(xue)(xue)(xue)科(ke)社(she)(she)區的(de)(de)(de)發(fa)(fa)展。因此(ci),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)網(wang)(wang)(wang)絡(luo)(luo)(luo)(luo)(luo)編委會代表的(de)(de)(de)專(zhuan)家領域包括心理學(xue)(xue)(xue)(xue),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)生物學(xue)(xue)(xue)(xue),計算機科(ke)學(xue)(xue)(xue)(xue),工程,數(shu)學(xue)(xue)(xue)(xue),物理。該雜(za)志(zhi)發(fa)(fa)表文(wen)章、信(xin)件(jian)和(he)評論以及給編輯的(de)(de)(de)信(xin)件(jian)、社(she)(she)論、時(shi)事、軟件(jian)調查(cha)和(he)專(zhuan)利信(xin)息(xi)。文(wen)章發(fa)(fa)表在五個(ge)部分(fen)之一:認知科(ke)學(xue)(xue)(xue)(xue),神(shen)(shen)(shen)(shen)經(jing)(jing)(jing)(jing)科(ke)學(xue)(xue)(xue)(xue),學(xue)(xue)(xue)(xue)習(xi)系統,數(shu)學(xue)(xue)(xue)(xue)和(he)計算分(fen)析、工程和(he)應用(yong)。 官網(wang)(wang)(wang)地址:

In Multi-Label Text Classification (MLTC), one sample can belong to more than one class. It is observed that most MLTC tasks, there are dependencies or correlations among labels. Existing methods tend to ignore the relationship among labels. In this paper, a graph attention network-based model is proposed to capture the attentive dependency structure among the labels. The graph attention network uses a feature matrix and a correlation matrix to capture and explore the crucial dependencies between the labels and generate classifiers for the task. The generated classifiers are applied to sentence feature vectors obtained from the text feature extraction network (BiLSTM) to enable end-to-end training. Attention allows the system to assign different weights to neighbor nodes per label, thus allowing it to learn the dependencies among labels implicitly. The results of the proposed model are validated on five real-world MLTC datasets. The proposed model achieves similar or better performance compared to the previous state-of-the-art models.

Natural Language Processing (NLP) and especially natural language text analysis have seen great advances in recent times. Usage of deep learning in text processing has revolutionized the techniques for text processing and achieved remarkable results. Different deep learning architectures like CNN, LSTM, and very recent Transformer have been used to achieve state of the art results variety on NLP tasks. In this work, we survey a host of deep learning architectures for text classification tasks. The work is specifically concerned with the classification of Hindi text. The research in the classification of morphologically rich and low resource Hindi language written in Devanagari script has been limited due to the absence of large labeled corpus. In this work, we used translated versions of English data-sets to evaluate models based on CNN, LSTM and Attention. Multilingual pre-trained sentence embeddings based on BERT and LASER are also compared to evaluate their effectiveness for the Hindi language. The paper also serves as a tutorial for popular text classification techniques.

Recently, researches have explored the graph neural network (GNN) techniques on text classification, since GNN does well in handling complex structures and preserving global information. However, previous methods based on GNN are mainly faced with the practical problems of fixed corpus level graph structure which do not support online testing and high memory consumption. To tackle the problems, we propose a new GNN based model that builds graphs for each input text with global parameters sharing instead of a single graph for the whole corpus. This method removes the burden of dependence between an individual text and entire corpus which support online testing, but still preserve global information. Besides, we build graphs by much smaller windows in the text, which not only extract more local features but also significantly reduce the edge numbers as well as memory consumption. Experiments show that our model outperforms existing models on several text classification datasets even with consuming less memory.

Recently, a variety of model designs and methods have blossomed in the context of the sentiment analysis domain. However, there is still a lack of wide and comprehensive studies of aspect-based sentiment analysis (ABSA). We want to fill this gap and propose a comparison with ablation analysis of aspect term extraction using various text embedding methods. We particularly focused on architectures based on long short-term memory (LSTM) with optional conditional random field (CRF) enhancement using different pre-trained word embeddings. Moreover, we analyzed the influence on the performance of extending the word vectorization step with character embedding. The experimental results on SemEval datasets revealed that not only does bi-directional long short-term memory (BiLSTM) outperform regular LSTM, but also word embedding coverage and its source highly affect aspect detection performance. An additional CRF layer consistently improves the results as well.

Graph convolutional network (GCN) has been successfully applied to many graph-based applications; however, training a large-scale GCN remains challenging. Current SGD-based algorithms suffer from either a high computational cost that exponentially grows with number of GCN layers, or a large space requirement for keeping the entire graph and the embedding of each node in memory. In this paper, we propose Cluster-GCN, a novel GCN algorithm that is suitable for SGD-based training by exploiting the graph clustering structure. Cluster-GCN works as the following: at each step, it samples a block of nodes that associate with a dense subgraph identified by a graph clustering algorithm, and restricts the neighborhood search within this subgraph. This simple but effective strategy leads to significantly improved memory and computational efficiency while being able to achieve comparable test accuracy with previous algorithms. To test the scalability of our algorithm, we create a new Amazon2M data with 2 million nodes and 61 million edges which is more than 5 times larger than the previous largest publicly available dataset (Reddit). For training a 3-layer GCN on this data, Cluster-GCN is faster than the previous state-of-the-art VR-GCN (1523 seconds vs 1961 seconds) and using much less memory (2.2GB vs 11.2GB). Furthermore, for training 4 layer GCN on this data, our algorithm can finish in around 36 minutes while all the existing GCN training algorithms fail to train due to the out-of-memory issue. Furthermore, Cluster-GCN allows us to train much deeper GCN without much time and memory overhead, which leads to improved prediction accuracy---using a 5-layer Cluster-GCN, we achieve state-of-the-art test F1 score 99.36 on the PPI dataset, while the previous best result was 98.71 by [16].

In this paper, we address the hyperspectral image (HSI) classification task with a generative adversarial network and conditional random field (GAN-CRF) -based framework, which integrates a semi-supervised deep learning and a probabilistic graphical model, and make three contributions. First, we design four types of convolutional and transposed convolutional layers that consider the characteristics of HSIs to help with extracting discriminative features from limited numbers of labeled HSI samples. Second, we construct semi-supervised GANs to alleviate the shortage of training samples by adding labels to them and implicitly reconstructing real HSI data distribution through adversarial training. Third, we build dense conditional random fields (CRFs) on top of the random variables that are initialized to the softmax predictions of the trained GANs and are conditioned on HSIs to refine classification maps. This semi-supervised framework leverages the merits of discriminative and generative models through a game-theoretical approach. Moreover, even though we used very small numbers of labeled training HSI samples from the two most challenging and extensively studied datasets, the experimental results demonstrated that spectral-spatial GAN-CRF (SS-GAN-CRF) models achieved top-ranking accuracy for semi-supervised HSI classification.

Text Classification is an important and classical problem in natural language processing. There have been a number of studies that applied convolutional neural networks (convolution on regular grid, e.g., sequence) to classification. However, only a limited number of studies have explored the more flexible graph convolutional neural networks (e.g., convolution on non-grid, e.g., arbitrary graph) for the task. In this work, we propose to use graph convolutional networks for text classification. We build a single text graph for a corpus based on word co-occurrence and document word relations, then learn a Text Graph Convolutional Network (Text GCN) for the corpus. Our Text GCN is initialized with one-hot representation for word and document, it then jointly learns the embeddings for both words and documents, as supervised by the known class labels for documents. Our experimental results on multiple benchmark datasets demonstrate that a vanilla Text GCN without any external word embeddings or knowledge outperforms state-of-the-art methods for text classification. On the other hand, Text GCN also learns predictive word and document embeddings. In addition, experimental results show that the improvement of Text GCN over state-of-the-art comparison methods become more prominent as we lower the percentage of training data, suggesting the robustness of Text GCN to less training data in text classification.

Classifying large scale networks into several categories and distinguishing them according to their fine structures is of great importance with several applications in real life. However, most studies of complex networks focus on properties of a single network but seldom on classification, clustering, and comparison between different networks, in which the network is treated as a whole. Due to the non-Euclidean properties of the data, conventional methods can hardly be applied on networks directly. In this paper, we propose a novel framework of complex network classifier (CNC) by integrating network embedding and convolutional neural network to tackle the problem of network classification. By training the classifiers on synthetic complex network data and real international trade network data, we show CNC can not only classify networks in a high accuracy and robustness, it can also extract the features of the networks automatically.

Cross-modal information retrieval aims to find heterogeneous data of various modalities from a given query of one modality. The main challenge is to map different modalities into a common semantic space, in which distance between concepts in different modalities can be well modeled. For cross-modal information retrieval between images and texts, existing work mostly uses off-the-shelf Convolutional Neural Network (CNN) for image feature extraction. For texts, word-level features such as bag-of-words or word2vec are employed to build deep learning models to represent texts. Besides word-level semantics, the semantic relations between words are also informative but less explored. In this paper, we model texts by graphs using similarity measure based on word2vec. A dual-path neural network model is proposed for couple feature learning in cross-modal information retrieval. One path utilizes Graph Convolutional Network (GCN) for text modeling based on graph representations. The other path uses a neural network with layers of nonlinearities for image modeling based on off-the-shelf features. The model is trained by a pairwise similarity loss function to maximize the similarity of relevant text-image pairs and minimize the similarity of irrelevant pairs. Experimental results show that the proposed model outperforms the state-of-the-art methods significantly, with 17% improvement on accuracy for the best case.

Spectral clustering is a leading and popular technique in unsupervised data analysis. Two of its major limitations are scalability and generalization of the spectral embedding (i.e., out-of-sample-extension). In this paper we introduce a deep learning approach to spectral clustering that overcomes the above shortcomings. Our network, which we call SpectralNet, learns a map that embeds input data points into the eigenspace of their associated graph Laplacian matrix and subsequently clusters them. We train SpectralNet using a procedure that involves constrained stochastic optimization. Stochastic optimization allows it to scale to large datasets, while the constraints, which are implemented using a special-purpose output layer, allow us to keep the network output orthogonal. Moreover, the map learned by SpectralNet naturally generalizes the spectral embedding to unseen data points. To further improve the quality of the clustering, we replace the standard pairwise Gaussian affinities with affinities leaned from unlabeled data using a Siamese network. Additional improvement can be achieved by applying the network to code representations produced, e.g., by standard autoencoders. Our end-to-end learning procedure is fully unsupervised. In addition, we apply VC dimension theory to derive a lower bound on the size of SpectralNet. State-of-the-art clustering results are reported on the Reuters dataset. Our implementation is publicly available at //github.com/kstant0725/SpectralNet .

北京阿比特科技有限公司