Casimir preserving integrators for stochastic Lie-Poisson equations with Stratonovich noise are developed extending Runge-Kutta Munthe-Kaas methods. The underlying Lie-Poisson structure is preserved along stochastic trajectories. A related stochastic differential equation on the Lie algebra is derived. The solution of this differential equation updates the evolution of the Lie-Poisson dynamics by means of the exponential map. The constructed numerical method conserves Casimir-invariants exactly, which is important for long time integration. This is illustrated numerically for the case of the stochastic heavy top.
Particle smoothers are SMC (Sequential Monte Carlo) algorithms designed to approximate the joint distribution of the states given observations from a state-space model. We propose dSMC (de-Sequentialized Monte Carlo), a new particle smoother that is able to process $T$ observations in $\mathcal{O}(\log T)$ time on parallel architecture. This compares favourably with standard particle smoothers, the complexity of which is linear in $T$. We derive $\mathcal{L}_p$ convergence results for dSMC, with an explicit upper bound, polynomial in $T$. We then discuss how to reduce the variance of the smoothing estimates computed by dSMC by (i) designing good proposal distributions for sampling the particles at the initialization of the algorithm, as well as by (ii) using lazy resampling to increase the number of particles used in dSMC. Finally, we design a particle Gibbs sampler based on dSMC, which is able to perform parameter inference in a state-space model at a $\mathcal{O}(\log(T))$ cost on parallel hardware.
A new integer-valued autoregressive process (INAR) with Generalised Lagrangian Katz (GLK) innovations is defined. We show that our GLK-INAR process is stationary, discrete semi-self-decomposable, infinite divisible, and provides a flexible modelling framework for count data allowing for under- and over-dispersion, asymmetry, and excess of kurtosis. A Bayesian inference framework and an efficient posterior approximation procedure based on Markov Chain Monte Carlo are provided. The proposed model family is applied to a Google Trend dataset which proxies the public concern about climate change around the world. The empirical results provide new evidence of heterogeneity across countries and keywords in the persistence, uncertainty, and long-run public awareness level.
This work presents the first study of using the popular Monte Carlo Tree Search (MCTS) method combined with dedicated heuristics for solving the Weighted Vertex Coloring Problem. Starting with the basic MCTS algorithm, we gradually introduce a number of algorithmic variants where MCTS is extended by various simulation strategies including greedy and local search heuristics. We conduct experiments on well-known benchmark instances to assess the value of each studied combination. We also provide empirical evidence to shed light on the advantages and limits of each strategy.
Numerical methods that preserve geometric invariants of the system, such as energy, momentum or the symplectic form, are called geometric integrators. Variational integrators are an important class of geometric integrators. The general idea for those variational integrators is to discretize Hamilton's principle rather than the equations of motion in a way that preserves some of the invariants of the original system. In this paper we construct variational integrators with fixed time step for time-dependent Lagrangian systems modelling an important class of autonomous dissipative systems. These integrators are derived via a family of discrete Lagrangian functions each one for a fixed time-step. This allows to recover at each step on the set of discrete sequences the preservation properties of variational integrators for autonomous Lagrangian systems, such as symplecticity or backward error analysis for these systems. We also present a discrete Noether theorem for this class of systems. Applications of the results are shown for the problem of formation stabilization of multi-agent systems.
In this paper, we study the probabilistic stability analysis of a subclass of stochastic hybrid systems, called the Planar Probabilistic Piecewise Constant Derivative Systems (Planar PPCD), where the continuous dynamics is deterministic, constant rate and planar, the discrete switching between the modes is probabilistic and happens at boundary of the invariant regions, and the continuous states are not reset during switching. These aptly model piecewise linear behaviors of planar robots. Our main result is an exact algorithm for deciding absolute and almost sure stability of Planar PPCD under some mild assumptions on mutual reachability between the states and the presence of non-zero probability self-loops. Our main idea is to reduce the stability problems on planar PPCD into corresponding problems on Discrete Time Markov Chains with edge weights. Our experimental results on planar robots with faulty angle actuator demonstrate the practical feasibility of this approach.
We introduce a general framework of low regularity integrators which allows us to approximate the time dynamics of a large class of equations, including parabolic and hyperbolic problems, as well as dispersive equations, up to arbitrary high order on general domains. The structure of the local error of the new schemes is driven by nested commutators which in general require (much) lower regularity assumptions than classical methods do. Our main idea lies in embedding the central oscillations of the nonlinear PDE into the numerical discretisation. The latter is achieved by a novel decorated tree formalism inspired by singular SPDEs with Regularity Structures and allows us to control the nonlinear interactions in the system up to arbitrary high order on the infinite dimensional (continuous) as well as finite dimensional (discrete) level.
Since almost twenty years, modified Patankar--Runge--Kutta (MPRK) methods have proven to be efficient and robust numerical schemes that preserve positivity and conservativity of the production-destruction system irrespectively of the time step size chosen. Due to these advantageous properties they are used for a wide variety of applications. Nevertheless, until now, an analytic investigation of the stability of MPRK schemes is still missing, since the usual approach by means of Dahlquist's equation is not feasible. Therefore, we consider a positive and conservative 2D test problem and provide statements usable for a stability analysis of general positive and conservative time integrator schemes based on the center manifold theory. We use this approach to investigate the Lyapunov stability of the second order MPRK22($\alpha$) and MPRK22ncs($\alpha$) schemes. We prove that MPRK22($\alpha$) schemes are unconditionally stable and derive the stability regions of MPRK22ncs($\alpha$) schemes. Finally, numerical experiments are presented, which confirm the theoretical results.
This paper deals with the control of parasitism in variational integrators for degenerate Lagrangian systems by writing them as general linear methods. This enables us to calculate their parasitic growth parameters which are responsible for the loss of long-time energy conservation properties of these algorithms. As a remedy and to offset the effects of parasitism, the standard projection technique is then applied to the general linear methods to numerically preserve the invariants of the degenerate Lagrangian systems by projecting the solution onto the desired manifold.
Interpretation of Deep Neural Networks (DNNs) training as an optimal control problem with nonlinear dynamical systems has received considerable attention recently, yet the algorithmic development remains relatively limited. In this work, we make an attempt along this line by reformulating the training procedure from the trajectory optimization perspective. We first show that most widely-used algorithms for training DNNs can be linked to the Differential Dynamic Programming (DDP), a celebrated second-order trajectory optimization algorithm rooted in the Approximate Dynamic Programming. In this vein, we propose a new variant of DDP that can accept batch optimization for training feedforward networks, while integrating naturally with the recent progress in curvature approximation. The resulting algorithm features layer-wise feedback policies which improve convergence rate and reduce sensitivity to hyper-parameter over existing methods. We show that the algorithm is competitive against state-ofthe-art first and second order methods. Our work opens up new avenues for principled algorithmic design built upon the optimal control theory.
We study the problem of training deep neural networks with Rectified Linear Unit (ReLU) activiation function using gradient descent and stochastic gradient descent. In particular, we study the binary classification problem and show that for a broad family of loss functions, with proper random weight initialization, both gradient descent and stochastic gradient descent can find the global minima of the training loss for an over-parameterized deep ReLU network, under mild assumption on the training data. The key idea of our proof is that Gaussian random initialization followed by (stochastic) gradient descent produces a sequence of iterates that stay inside a small perturbation region centering around the initial weights, in which the empirical loss function of deep ReLU networks enjoys nice local curvature properties that ensure the global convergence of (stochastic) gradient descent. Our theoretical results shed light on understanding the optimization of deep learning, and pave the way to study the optimization dynamics of training modern deep neural networks.