Ridges play a vital role in accurately approximating the underlying structure of manifolds. In this paper, we explore the ridge's variation by applying a concave nonlinear transformation to the density function. Through the derivation of the Hessian matrix, we observe that nonlinear transformations yield a rank-one modification of the Hessian matrix. Leveraging the variational properties of eigenvalue problems, we establish a partial order inclusion relationship among the corresponding ridges. We intuitively discover that the transformation can lead to improved estimation of the tangent space via rank-one modification of the Hessian matrix. To validate our theories, we conduct extensive numerical experiments on synthetic and real-world datasets that demonstrate the superiority of the ridges obtained from our transformed approach in approximating the underlying truth manifold compared to other manifold fitting algorithms.
Improving ASR systems is necessary to make new LLM-based use-cases accessible to people across the globe. In this paper, we focus on Indian languages, and make the case that diverse benchmarks are required to evaluate and improve ASR systems for Indian languages. To address this, we collate Vistaar as a set of 59 benchmarks across various language and domain combinations, on which we evaluate 3 publicly available ASR systems and 2 commercial systems. We also train IndicWhisper models by fine-tuning the Whisper models on publicly available training datasets across 12 Indian languages totalling to 10.7K hours. We show that IndicWhisper significantly improves on considered ASR systems on the Vistaar benchmark. Indeed, IndicWhisper has the lowest WER in 39 out of the 59 benchmarks, with an average reduction of 4.1 WER. We open-source all datasets, code and models.
Many neurodegenerative diseases are connected to the spreading of misfolded prionic proteins. In this paper, we analyse the process of misfolding and spreading of both $\alpha$-synuclein and Amyloid-$\beta$, related to Parkinson's and Alzheimer's diseases, respectively. We introduce and analyze a positivity-preserving numerical method for the discretization of the Fisher-Kolmogorov equation, modelling accumulation and spreading of prionic proteins. The proposed approximation method is based on the discontinuous Galerkin method on polygonal and polyhedral grids for space discretization and on $\vartheta-$method time integration scheme. We prove the existence of the discrete solution and a convergence result where the Implicit Euler scheme is employed for time integration. We show that the proposed approach is structure-preserving, in the sense that it guaranteed that the discrete solution is non-negative, a feature that is of paramount importance in practical application. The numerical verification of our numerical model is performed both using a manufactured solution and considering wavefront propagation in two-dimensional polygonal grids. Next, we present a simulation of $\alpha$-synuclein spreading in a two-dimensional brain slice in the sagittal plane. The polygonal mesh for this simulation is agglomerated maintaining the distinction of white and grey matter, taking advantage of the flexibility of PolyDG methods in the mesh construction. Finally, we simulate the spreading of Amyloid-$\beta$ in a patient-specific setting by using a three-dimensional geometry reconstructed from magnetic resonance images and an initial condition reconstructed from positron emission tomography. Our numerical simulations confirm that the proposed method is able to capture the evolution of Parkinson's and Alzheimer's diseases.
In this paper we analyze a homogeneous parabolic problem with initial data in the space of regular Borel measures. The problem is discretized in time with a discontinuous Galerkin scheme of arbitrary degree and in space with continuous finite elements of orders one or two. We show parabolic smoothing results for the continuous, semidiscrete and fully discrete problems. Our main results are interior $L^\infty$ error estimates for the evaluation at the endtime, in cases where the initial data is supported in a subdomain. In order to obtain these, we additionally show interior $L^\infty$ error estimates for $L^2$ initial data and quadratic finite elements, which extends the corresponding result previously established by the authors for linear finite elements.
In this paper, we propose a novel method of formulating an NP-hard wireless channel assignment problem as a higher-order unconstrained binary optimization (HUBO), where the Grover adaptive search (GAS) is used to provide a quadratic speedup for solving the problem. The conventional method relies on a one-hot encoding of the channel indices, resulting in a quadratic formulation. By contrast, we conceive ascending and descending binary encodings of the channel indices, construct a specific quantum circuit, and derive the exact numbers of qubits and gates required by GAS. Our analysis clarifies that the proposed HUBO formulation significantly reduces the number of qubits and the query complexity compared with the conventional quadratic formulation. This advantage is achieved at the cost of an increased number of quantum gates, which we demonstrate can be reduced by our proposed descending binary encoding.
Complex multibody legged robots can have complex rotational control challenges. In this paper, we propose a concise way to understand and formulate a \emph{whole-body orientation} that (i) depends on system configuration only and not a history of motion, (ii) can be representative of the orientation of the entire system while not being attached to any specific link, and (iii) has a rate of change that approximates total system angular momentum. We relate this orientation coordinate to past work, and discuss and demonstrate, including on hardware, several different uses for it.
In this paper we investigate the Curry-Howard correspondence for constructive modal logic in light of the gap between the proof equivalences enforced by the lambda calculi from the literature and by the recently defined winning strategies for this logic. We define a new lambda-calculus for a minimal constructive modal logic by enriching the calculus from the literature with additional reduction rules and we prove normalization and confluence for our calculus. We then provide a typing system in the style of focused proof systems allowing us to provide a unique proof for each term in normal form, and we use this result to show a one-to-one correspondence between terms in normal form and winning innocent strategies.
In this paper, we consider the coupled N/TH problem, in which the termination criterion for the neutronics iteration adopts an adaptive tolerance with respect to the fuel temperature residual at each Picard iteration. We refer to this coupling scheme as the inexact Picard iteration method. Fourier analysis is performed to investigate how the convergence behavior of Picard iteration is influenced by the inexact neutronics solution. It is found that if the convergence of the inner neutronics iteration is slow, Picard coupling may become unstable unless a tighter tolerance is used for the neutronics iteration. Nevertheless, our analysis indicates that a certain amount of over-solving is necessary for maintaining the stability of Picard iteration if the iterative solution of the subproblem is not fast enough. However, this issue has not been addressed in the previous studies.
In this paper, we propose a method for resume rating using Latent Dirichlet Allocation (LDA) and entity detection with SpaCy. The proposed method first extracts relevant entities such as education, experience, and skills from the resume using SpaCy's Named Entity Recognition (NER). The LDA model then uses these entities to rate the resume by assigning topic probabilities to each entity. Furthermore, we conduct a detailed analysis of the entity detection using SpaCy's NER and report its evaluation metrics. Using LDA, our proposed system breaks down resumes into latent topics and extracts meaningful semantic representations. With a vision to define our resume score to be more content-driven rather than a structure and keyword match driven, our model has achieved 77% accuracy with respect to only skills in consideration and an overall 82% accuracy with all attributes in consideration. (like college name, work experience, degree and skills)
In this paper, we proposed to apply meta learning approach for low-resource automatic speech recognition (ASR). We formulated ASR for different languages as different tasks, and meta-learned the initialization parameters from many pretraining languages to achieve fast adaptation on unseen target language, via recently proposed model-agnostic meta learning algorithm (MAML). We evaluated the proposed approach using six languages as pretraining tasks and four languages as target tasks. Preliminary results showed that the proposed method, MetaASR, significantly outperforms the state-of-the-art multitask pretraining approach on all target languages with different combinations of pretraining languages. In addition, since MAML's model-agnostic property, this paper also opens new research direction of applying meta learning to more speech-related applications.
In this paper, we present an accurate and scalable approach to the face clustering task. We aim at grouping a set of faces by their potential identities. We formulate this task as a link prediction problem: a link exists between two faces if they are of the same identity. The key idea is that we find the local context in the feature space around an instance (face) contains rich information about the linkage relationship between this instance and its neighbors. By constructing sub-graphs around each instance as input data, which depict the local context, we utilize the graph convolution network (GCN) to perform reasoning and infer the likelihood of linkage between pairs in the sub-graphs. Experiments show that our method is more robust to the complex distribution of faces than conventional methods, yielding favorably comparable results to state-of-the-art methods on standard face clustering benchmarks, and is scalable to large datasets. Furthermore, we show that the proposed method does not need the number of clusters as prior, is aware of noises and outliers, and can be extended to a multi-view version for more accurate clustering accuracy.