In this paper motivated from subspace coding we introduce subspace-metric codes and subset-metric codes. These are coordinate-position independent pseudometrics and suitable for the folded codes. The half-Singleton upper bounds for linear subspace-metric codes and linear subset-metric codes are proved. Subspace distances and subset distances of codes are natural lower bounds for insdel distances of codes, and then can be used to lower bound the insertion-deletion error-correcting capabilities of codes. Our subspace-metric codes or subset-metric codes can be used to construct explicit well-structured insertion-deletion codes directly. $k$-deletion correcting codes with rate approaching $1$ can be constructed from subspace codes. By analysing the subset distances of folded codes from evaluation codes of linear mappings, we prove that they have high subset distances and then are explicit good insertion-deletion codes.
In this paper we investigate known Singleton-like bounds in the Lee metric and characterize optimal codes, which turn out to be very few. We then focus on Plotkin-like bounds in the Lee metric and present a new bound that extends and refines a previously known, and out-performs it in the case of non-free codes. We then compute the density of optimal codes with regard to the new bound. Finally we fill a gap in the characterization of Lee-equidistant codes.
A Boolean maximum constraint satisfaction problem, Max-CSP($f$), is specified by a constraint function $f:\{-1,1\}^k\to\{0,1\}$; an instance on $n$ variables is given by a list of constraints applying $f$ on a tuple of "literals" of $k$ distinct variables chosen from the $n$ variables. Chou, Golovnev, and Velusamy [CGV20] obtained explicit constants characterizing the streaming approximability of all symmetric Max-2CSPs. More recently, Chou, Golovnev, Sudan, and Velusamy [CGSV21] proved a general dichotomy theorem tightly characterizing the approximability of Boolean Max-CSPs with respect to sketching algorithms. For every $f$, they showed that there exists an optimal approximation ratio $\alpha(f)\in (0,1]$ such that for every $\epsilon>0$, Max-CSP($f$) is $(\alpha(f)-\epsilon)$-approximable by a linear sketching algorithm in $O(\log n)$ space, but any $(\alpha(f)+\epsilon)$-approximation sketching algorithm for Max-CSP($f$) requires $\Omega(\sqrt{n})$ space. In this work, we build on the [CGSV21] dichotomy theorem and give closed-form expressions for the sketching approximation ratios of multiple families of symmetric Boolean functions. The functions include $k$AND and Th$_k^{k-1}$ (the ``weight-at-least-$(k-1)$'' threshold function on $k$ variables). In particular, letting $\alpha'_k = 2^{-(k-1)} (1-k^{-2})^{(k-1)/2}$, we show that for odd $k \geq 3$, $\alpha(k$AND$ = \alpha'_k$; for even $k \geq 2$, $\alpha(k$AND$) = 2\alpha'_{k+1}$; and for even $k \geq 2$, $\alpha($Th$_k^{k-1}) = \frac{k}2\alpha'_{k-1}$. We also resolve the ratio for the ``weight-exactly-$\frac{k+1}2$'' function for odd $k \in \{3,\ldots,51\}$ as well as fifteen other functions. These closed-form expressions need not have existed just given the [CGSV21] dichotomy. For arbitrary threshold functions, we also give optimal "bias-based" approximation algorithms generalizing [CGV20] and simplifying [CGSV21].
We provide a geometric characterization of $k$-dimensional $\mathbb{F}_{q^m}$-linear sum-rank metric codes as tuples of $\mathbb{F}_q$-subspaces of $\mathbb{F}_{q^m}^k$. We then use this characterization to study one-weight codes in the sum-rank metric. This leads us to extend the family of linearized Reed-Solomon codes in order to obtain a doubly-extended version of them. We prove that these codes are still maximum sum-rank distance (MSRD) codes and, when $k=2$, they are one-weight, as in the Hamming-metric case. We then focus on constant rank-profile codes in the sum-rank metric, which are a special family of one weight-codes, and derive constraints on their parameters with the aid of an associated Hamming-metric code. Furthermore, we introduce the $n$-simplex codes in the sum-rank metric, which are obtained as the orbit of a Singer subgroup of $\mathrm{GL}(k,q^m)$. They turn out to be constant rank-profile - and hence one-weight - and generalize the simplex codes in both the Hamming and the rank metric. Finally, we focus on $2$-dimensional one-weight codes, deriving constraints on the parameters of those which are also MSRD, and we find a new construction of one-weight MSRD codes when $q=2$.
The Skinned Multi-Person Linear (SMPL) model can represent a human body by mapping pose and shape parameters to body meshes. This has been shown to facilitate inferring 3D human pose and shape from images via different learning models. However, not all pose and shape parameter values yield physically-plausible or even realistic body meshes. In other words, SMPL is under-constrained and may thus lead to invalid results when used to reconstruct humans from images, either by directly optimizing its parameters, or by learning a mapping from the image to these parameters. In this paper, we therefore learn a prior that restricts the SMPL parameters to values that produce realistic poses via adversarial training. We show that our learned prior covers the diversity of the real-data distribution, facilitates optimization for 3D reconstruction from 2D keypoints, and yields better pose estimates when used for regression from images. We found that the prior based on spherical distribution gets the best results. Furthermore, in all these tasks, it outperforms the state-of-the-art VAE-based approach to constraining the SMPL parameters.
In this article, we define a new non-archimedean metric structure, called cophenetic metric, on persistent homology classes of all degrees. We then show that zeroth persistent homology together with the cophenetic metric and hierarchical clustering algorithms with a number of different metrics do deliver statistically verifiable commensurate topological information based on experimental results we obtained on different datasets. We also observe that the resulting clusters coming from cophenetic distance do shine in terms of different evaluation measures such as silhouette score and the Rand index. Moreover, since the cophenetic metric is defined for all homology degrees, one can now display the inter-relations of persistent homology classes in all degrees via rooted trees.
For Kolmogorov test we find necessary and sufficient conditions of uniform consistency of sets of alternatives approaching to hypothesis. Sets of alternatives can be defined both in terms of distribution functions and in terms of densities.
Estimation of linear functionals from observed data is an important task in many subjects. Juditsky & Nemirovski [The Annals of Statistics 37.5A (2009): 2278-2300] propose a framework for non-parametric estimation of linear functionals in a very general setting, with nearly minimax optimal confidence intervals. They compute this estimator and the associated confidence interval by approximating the saddle-point of a function. While this optimization problem is convex, it is rather difficult to solve using existing off-the-shelf optimization software. Furthermore, this computation can be expensive when the estimators live in a high-dimensional space. We propose a different algorithm to construct this estimator. Our algorithm can be used with existing optimization software and is much cheaper to implement even when the estimators are in a high-dimensional space, as long as the Hellinger affinity (or the Bhattacharyya coefficient) for the chosen parametric distribution can be efficiently computed given the parameters. We hope that our algorithm will foster the adoption of this estimation technique to a wider variety of problems with relative ease.
The problem of Approximate Nearest Neighbor (ANN) search is fundamental in computer science and has benefited from significant progress in the past couple of decades. However, most work has been devoted to pointsets whereas complex shapes have not been sufficiently treated. Here, we focus on distance functions between discretized curves in Euclidean space: they appear in a wide range of applications, from road segments to time-series in general dimension. For $\ell_p$-products of Euclidean metrics, for any $p$, we design simple and efficient data structures for ANN, based on randomized projections, which are of independent interest. They serve to solve proximity problems under a notion of distance between discretized curves, which generalizes both discrete Fr\'echet and Dynamic Time Warping distances. These are the most popular and practical approaches to comparing such curves. We offer the first data structures and query algorithms for ANN with arbitrarily good approximation factor, at the expense of increasing space usage and preprocessing time over existing methods. Query time complexity is comparable or significantly improved by our algorithms, our algorithm is especially efficient when the length of the curves is bounded.
Transformer architectures show significant promise for natural language processing. Given that a single pretrained model can be fine-tuned to perform well on many different tasks, these networks appear to extract generally useful linguistic features. A natural question is how such networks represent this information internally. This paper describes qualitative and quantitative investigations of one particularly effective model, BERT. At a high level, linguistic features seem to be represented in separate semantic and syntactic subspaces. We find evidence of a fine-grained geometric representation of word senses. We also present empirical descriptions of syntactic representations in both attention matrices and individual word embeddings, as well as a mathematical argument to explain the geometry of these representations.
Deep learning is the mainstream technique for many machine learning tasks, including image recognition, machine translation, speech recognition, and so on. It has outperformed conventional methods in various fields and achieved great successes. Unfortunately, the understanding on how it works remains unclear. It has the central importance to lay down the theoretic foundation for deep learning. In this work, we give a geometric view to understand deep learning: we show that the fundamental principle attributing to the success is the manifold structure in data, namely natural high dimensional data concentrates close to a low-dimensional manifold, deep learning learns the manifold and the probability distribution on it. We further introduce the concepts of rectified linear complexity for deep neural network measuring its learning capability, rectified linear complexity of an embedding manifold describing the difficulty to be learned. Then we show for any deep neural network with fixed architecture, there exists a manifold that cannot be learned by the network. Finally, we propose to apply optimal mass transportation theory to control the probability distribution in the latent space.