亚洲男人的天堂2018av,欧美草比,久久久久久免费视频精选,国色天香在线看免费,久久久久亚洲av成人片仓井空

We introduce the notion of symmetric covariation, which is a new measure of dependence between two components of a symmetric $\alpha$-stable random vector, where the stability parameter $\alpha$ measures the heavy-tailedness of its distribution. Unlike covariation that exists only when $\alpha\in(1,2]$, symmetric covariation is well defined for all $\alpha\in(0,2]$. We show that symmetric covariation can be defined using the proposed generalized fractional derivative, which has broader usages than those involved in this work. Several properties of symmetric covariation have been derived. These are either similar to or more general than those of the covariance functions in the Gaussian case. The main contribution of this framework is the representation of the characteristic function of bivariate symmetric $\alpha$-stable distribution via convergent series based on a sequence of symmetric covariations. This series representation extends the one of bivariate Gaussian.

相關內容

We propose a family of four-parameter distributions that contain the K-distribution as special case. The family is derived as a mixture distribution that uses the three-parameter reflected Gamma distribution as parental and the two-parameter Gamma distribution as prior. Properties of the proposed family are investigated as well; these include probability density function, cumulative distribution function, moments, and cumulants. The family is termed symmetric K-distribution (SKD) based on its resemblance to the K-distribution as well as its symmetric nature. The standard form of the SKD, which often proves to be an adequate model, is also discussed. Moreover, an order statistics analysis is provided as well as the distributions of the product and ratio of two independent and identical SKD random variables are derived. Finally, a generalisation of the proposed family, which enables non-zero skewness values, is investigated, while both the SKD and the skew-SKD are proven capable of describing the complex dynamics of machine learning, Bayesian analysis and other fields through simplified expressions with high accuracy.

The symmetries described by Pin groups are the result of combining a finite number of discrete reflections in (hyper)planes. The current work shows how an analysis using geometric algebra provides a picture complementary to that of the classic matrix Lie algebra approach, while retaining information about the number of reflections in a given transformation. This imposes a graded structure on Lie groups, which is not evident in their matrix representation. By embracing this graded structure, the invariant decomposition theorem was proven: any composition of $k$ linearly independent reflections can be decomposed into $\lceil k/2 \rceil$ commuting factors, each of which is the product of at most two reflections. This generalizes a conjecture by M. Riesz, and has e.g. the Mozzi-Chasles' theorem as its 3D Euclidean special case. To demonstrate its utility, we briefly discuss various examples such as Lorentz transformations, Wigner rotations, and screw transformations. The invariant decomposition also directly leads to closed form formulas for the exponential and logarithmic function for all Spin groups, and identifies element of geometry such as planes, lines, points, as the invariants of $k$-reflections. We conclude by presenting novel matrix/vector representations for geometric algebras $\mathbb{R}_{pqr}$, and use this in E(3) to illustrate the relationship with the classic covariant, contravariant and adjoint representations for the transformation of points, planes and lines.

Effective molecular representation learning is of great importance to facilitate molecular property prediction, which is a fundamental task for the drug and material industry. Recent advances in graph neural networks (GNNs) have shown great promise in applying GNNs for molecular representation learning. Moreover, a few recent studies have also demonstrated successful applications of self-supervised learning methods to pre-train the GNNs to overcome the problem of insufficient labeled molecules. However, existing GNNs and pre-training strategies usually treat molecules as topological graph data without fully utilizing the molecular geometry information. Whereas, the three-dimensional (3D) spatial structure of a molecule, a.k.a molecular geometry, is one of the most critical factors for determining molecular physical, chemical, and biological properties. To this end, we propose a novel Geometry Enhanced Molecular representation learning method (GEM) for Chemical Representation Learning (ChemRL). At first, we design a geometry-based GNN architecture that simultaneously models atoms, bonds, and bond angles in a molecule. To be specific, we devised double graphs for a molecule: The first one encodes the atom-bond relations; The second one encodes bond-angle relations. Moreover, on top of the devised GNN architecture, we propose several novel geometry-level self-supervised learning strategies to learn spatial knowledge by utilizing the local and global molecular 3D structures. We compare ChemRL-GEM with various state-of-the-art (SOTA) baselines on different molecular benchmarks and exhibit that ChemRL-GEM can significantly outperform all baselines in both regression and classification tasks. For example, the experimental results show an overall improvement of 8.8% on average compared to SOTA baselines on the regression tasks, demonstrating the superiority of the proposed method.

We explore links between the thin concurrent games of Castellan, Clairambault and Winskel, and the weighted relational models of linear logic studied by Laird, Manzonetto, McCusker and Pagani. More precisely, we show that there is an interpretationpreserving "collapse" functor from the former to the latter. On objects, the functor defines for each game a set of possible execution states. Defining the action on morphisms is more subtle, and this is the main contribution of the paper. Given a strategy and an execution state, our functor needs to count the witnesses for this state within the strategy. Strategies in thin concurrent games describe non-linear behaviour explicitly, so in general each witness exists in countably many symmetric copies. The challenge is to define the right notion of witnesses, factoring out this infinity while matching the weighted relational model. Understanding how witnesses compose is particularly subtle and requires a delve into the combinatorics of witnesses and their symmetries. In its basic form, this functor connects thin concurrent games and a relational model weighted by N $\cup$ {+$\infty$}. We will additionally consider a generalised setting where both models are weighted by elements of an arbitrary continuous semiring; this covers the probabilistic case, among others. Witnesses now additionally carry a value from the semiring, and our interpretation-preserving collapse functor extends to this setting.

We apply the concept of distance correlation for testing independence of long-range dependent time series. For this, we establish a non-central limit theorem for stochastic processes with values in an $L_2$-Hilbert space. This limit theorem is of a general theoretical interest that goes beyond the context of this article. For the purpose of this article, it provides the basis for deriving the asymptotic distribution of the distance covariance of subordinated Gaussian processes. Depending on the dependence in the data, the standardization and the limit of distance correlation vary. In any case, the limit is not feasible, such that test decisions are based on a subsampling procedure. We prove the validity of the subsampling procedure and assess the finite sample performance of a hypothesis test based on the distance covariance. In particular, we compare its finite sample performance to that of a test based on Pearson's sample correlation coefficient. For this purpose, we additionally establish convergence results for this dependence measure. Different dependencies between the vectors are considered. It turns out that only linear correlation is better detected by Pearson's sample correlation coefficient, while all other dependencies are better detected by distance correlation. An analysis with regard to cross-dependencies between the mean monthly discharges of three different rivers provides an application of the theoretical results established in this article.

The information bottleneck (IB) method is a technique for extracting information that is relevant for predicting the target random variable from the source random variable, which is typically implemented by optimizing the IB Lagrangian that balances the compression and prediction terms. However, the IB Lagrangian is hard to optimize, and multiple trials for tuning values of Lagrangian multiplier are required. Moreover, we show that the prediction performance strictly decreases as the compression gets stronger during optimizing the IB Lagrangian. In this paper, we implement the IB method from the perspective of supervised disentangling. Specifically, we introduce Disentangled Information Bottleneck (DisenIB) that is consistent on compressing source maximally without target prediction performance loss (maximum compression). Theoretical and experimental results demonstrate that our method is consistent on maximum compression, and performs well in terms of generalization, robustness to adversarial attack, out-of-distribution detection, and supervised disentangling.

To learn intrinsic low-dimensional structures from high-dimensional data that most discriminate between classes, we propose the principle of Maximal Coding Rate Reduction ($\text{MCR}^2$), an information-theoretic measure that maximizes the coding rate difference between the whole dataset and the sum of each individual class. We clarify its relationships with most existing frameworks such as cross-entropy, information bottleneck, information gain, contractive and contrastive learning, and provide theoretical guarantees for learning diverse and discriminative features. The coding rate can be accurately computed from finite samples of degenerate subspace-like distributions and can learn intrinsic representations in supervised, self-supervised, and unsupervised settings in a unified manner. Empirically, the representations learned using this principle alone are significantly more robust to label corruptions in classification than those using cross-entropy, and can lead to state-of-the-art results in clustering mixed data from self-learned invariant features.

Network embedding aims to learn a latent, low-dimensional vector representations of network nodes, effective in supporting various network analytic tasks. While prior arts on network embedding focus primarily on preserving network topology structure to learn node representations, recently proposed attributed network embedding algorithms attempt to integrate rich node content information with network topological structure for enhancing the quality of network embedding. In reality, networks often have sparse content, incomplete node attributes, as well as the discrepancy between node attribute feature space and network structure space, which severely deteriorates the performance of existing methods. In this paper, we propose a unified framework for attributed network embedding-attri2vec-that learns node embeddings by discovering a latent node attribute subspace via a network structure guided transformation performed on the original attribute space. The resultant latent subspace can respect network structure in a more consistent way towards learning high-quality node representations. We formulate an optimization problem which is solved by an efficient stochastic gradient descent algorithm, with linear time complexity to the number of nodes. We investigate a series of linear and non-linear transformations performed on node attributes and empirically validate their effectiveness on various types of networks. Another advantage of attri2vec is its ability to solve out-of-sample problems, where embeddings of new coming nodes can be inferred from their node attributes through the learned mapping function. Experiments on various types of networks confirm that attri2vec is superior to state-of-the-art baselines for node classification, node clustering, as well as out-of-sample link prediction tasks. The source code of this paper is available at //github.com/daokunzhang/attri2vec.

We present the problem of selecting relevant premises for a proof of a given statement. When stated as a binary classification task for pairs (conjecture, axiom), it can be efficiently solved using artificial neural networks. The key difference between our advance to solve this problem and previous approaches is the use of just functional signatures of premises. To further improve the performance of the model, we use dimensionality reduction technique, to replace long and sparse signature vectors with their compact and dense embedded versions. These are obtained by firstly defining the concept of a context for each functor symbol, and then training a simple neural network to predict the distribution of other functor symbols in the context of this functor. After training the network, the output of its hidden layer is used to construct a lower dimensional embedding of a functional signature (for each premise) with a distributed representation of features. This allows us to use 512-dimensional embeddings for conjecture-axiom pairs, containing enough information about the original statements to reach the accuracy of 76.45% in premise selection task, only with simple two-layer densely connected neural networks.

Methods that align distributions by minimizing an adversarial distance between them have recently achieved impressive results. However, these approaches are difficult to optimize with gradient descent and they often do not converge well without careful hyperparameter tuning and proper initialization. We investigate whether turning the adversarial min-max problem into an optimization problem by replacing the maximization part with its dual improves the quality of the resulting alignment and explore its connections to Maximum Mean Discrepancy. Our empirical results suggest that using the dual formulation for the restricted family of linear discriminators results in a more stable convergence to a desirable solution when compared with the performance of a primal min-max GAN-like objective and an MMD objective under the same restrictions. We test our hypothesis on the problem of aligning two synthetic point clouds on a plane and on a real-image domain adaptation problem on digits. In both cases, the dual formulation yields an iterative procedure that gives more stable and monotonic improvement over time.

北京阿比特科技有限公司