In this paper we consider the age of information (AoI) of a status updating system with a relay, where the updates are delivered to destination either from the direct line or the two-hop link via the relay. An updating packet generated at source is sent to receiver and the relay simultaneously. When the direct packet transmission fails, the relay replaces the source and retransmits the packet until it is eventually obtained at the receiver side. Assume that the propagation delay on each link is one time slot, we determine the stationary distribution of the AoI for three cases: (a) relay has no buffer and the packet delivery from relay cannot be preempted by fresher updates from source; (b) relay has no buffer but the packet substitution is allowable; (c) relay has size 1 buffer and the packet in buffer is refreshed when a newer packet is obtained. The idea is invoking a multiple-dimensional state vector which contains the AoI as a part and constituting the multiple-dimensional AoI stochastic process. We find the steady state of each multiple-dimensional AoI process by solving the system of stationary equations. Once the steady-state distribution of larger-dimensional AoI process is known, the stationary AoI distribution is also obtained as it is one of the marginal distributions of that process's steady-state distribution. For all the situations, we derive the explicit expression of AoI distribution, and calculate the mean and the variance of the stationary AoI. All the results are compared numerically, including the AoI performance of the non-relay state updating system. Numerical results show that adding the relay improves the system's timeliness dramatically, and no-buffer-and-preemption setting in relay achieves both minimal average AoI and AoI's variance. Thus, for the system model discussed in this paper, to reduce the AoI at receiver there is no need to add the buffer in relay.
In this paper we consider the problem of best-arm identification in multi-armed bandits in the fixed confidence setting, where the goal is to identify, with probability $1-\delta$ for some $\delta>0$, the arm with the highest mean reward in minimum possible samples from the set of arms $\mathcal{K}$. Most existing best-arm identification algorithms and analyses operate under the assumption that the rewards corresponding to different arms are independent of each other. We propose a novel correlated bandit framework that captures domain knowledge about correlation between arms in the form of upper bounds on expected conditional reward of an arm, given a reward realization from another arm. Our proposed algorithm C-LUCB, which generalizes the LUCB algorithm utilizes this partial knowledge of correlations to sharply reduce the sample complexity of best-arm identification. More interestingly, we show that the total samples obtained by C-LUCB are of the form $\mathcal{O}\left(\sum_{k \in \mathcal{C}} \log\left(\frac{1}{\delta}\right)\right)$ as opposed to the typical $\mathcal{O}\left(\sum_{k \in \mathcal{K}} \log\left(\frac{1}{\delta}\right)\right)$ samples required in the independent reward setting. The improvement comes, as the $\mathcal{O}(\log(1/\delta))$ term is summed only for the set of competitive arms $\mathcal{C}$, which is a subset of the original set of arms $\mathcal{K}$. The size of the set $\mathcal{C}$, depending on the problem setting, can be as small as $2$, and hence using C-LUCB in the correlated bandits setting can lead to significant performance improvements. Our theoretical findings are supported by experiments on the Movielens and Goodreads recommendation datasets.
Deciding on the unimodality of a dataset is an important problem in data analysis and statistical modeling. It allows to obtain knowledge about the structure of the dataset, ie. whether data points have been generated by a probability distribution with a single or more than one peaks. Such knowledge is very useful for several data analysis problems, such as for deciding on the number of clusters and determining unimodal projections. We propose a technique called UU-test (Unimodal Uniform test) to decide on the unimodality of a one-dimensional dataset. The method operates on the empirical cumulative density function (ecdf) of the dataset. It attempts to build a piecewise linear approximation of the ecdf that is unimodal and models the data sufficiently in the sense that the data corresponding to each linear segment follows the uniform distribution. A unique feature of this approach is that in the case of unimodality, it also provides a statistical model of the data in the form of a Uniform Mixture Model. We present experimental results in order to assess the ability of the method to decide on unimodality and perform comparisons with the well-known dip-test approach. In addition, in the case of unimodal datasets we evaluate the Uniform Mixture Models provided by the proposed method using the test set log-likelihood and the two-sample Kolmogorov-Smirnov (KS) test.
We derive fundamental performance limitations for intrinsic average consensus problems in open multi-agent systems, which are systems subject to frequent arrivals and departures of agents. Each agent holds a value, and the objective of the agents is to collaboratively estimate the average of the values of the agents presently in the system. Algorithms solving such problems in open systems are poised to never converge because of the permanent variations in the composition, size and objective pursued by the agents of the system. We provide lower bounds on the expected Mean Squared Error achievable by any averaging algorithms in open systems of fixed size. Our derivation is based on the analysis of a conceptual algorithm that would achieve optimal performance for a given model of replacements. We obtain a general bound that depends on the properties of the model defining the interactions between the agents, and instantiate that result for all-to-one and one-to-one interaction models. A comparison between those bounds and algorithms implementable with those models is then provided to highlight their validity.
Ultra-reliability and low latency communication has long been an important but challenging task in the fifth and sixth generation wireless communication systems. Scheduling as many users as possible to serve on the limited time-frequency resource is one of a crucial topic, subjecting to the maximum allowable transmission power and the minimum rate requirement of each user. We address it by proposing a mixed integer programming model, with the goal of maximizing the set cardinality of users instead of maximizing the system sum rate or energy efficiency. Mathematical transformations and successive convex approximation are combined to solve the complex optimization problem. Numerical results show that the proposed method achieves a considerable performance compared with exhaustive search method, but with lower computational complexity.
This paper presents a way to define, classify and accelerate the order of convergence of an uncountable family of fractional fixed point methods, which may be useful to continue expanding the applications of fractional operators. The proposed method to accelerate convergence is used in a fractional iterative method, and with the obtained method are solved simultaneously two nonlinear algebraic systems that depend on time-dependent parameters, and that allow obtaining the temperatures and efficiencies of a hybrid solar receiver. Finally, two uncountable families of fractional fixed point methods are presented, in which the proposed method to accelerate convergence can be implemented.
In this paper, we analyze status update systems modeled through the Stochastic Hybrid Systems (SHSs) tool. Contrary to previous works, we allow the system's transition dynamics to be functions of the Age of Information (AoI). This dependence allows us to encapsulate many applications and opens the door for more sophisticated systems to be studied. However, this same dependence on the AoI engenders technical and analytical difficulties that we address in this paper. Specifically, we first showcase several characteristics of the age processes modeled through the SHSs tool. Then, we provide a framework to establish the Lagrange stability and positive recurrence of these processes. Building on this, we provide an approach to compute the m-th moment of the age processes. Interestingly, this technique allows us to approximate the average age by solving a simple set of linear equations. Equipped with this approach, we also provide a sequential convex approximation method to optimize the average age by calibrating the parameters of the system. Finally, we consider an age-dependent CSMA environment where the backoff duration depends on the instantaneous age. By leveraging our analysis, we contrast its performance to the age-blind CSMA and showcase the age performance gain provided by the former.
In this paper, the problem of minimizing the weighted sum of age of information (AoI) and total energy consumption of Internet of Things (IoT) devices is studied. In the considered model, each IoT device monitors a physical process that follows nonlinear dynamics. As the dynamics of the physical process vary over time, each device must find an optimal sampling frequency to sample the real-time dynamics of the physical system and send sampled information to a base station (BS). Due to limited wireless resources, the BS can only select a subset of devices to transmit their sampled information. Thus, edge devices must cooperatively sample their monitored dynamics based on the local observations and the BS must collect the sampled information from the devices immediately, hence avoiding the additional time and energy used for sampling and information transmission. To this end, it is necessary to jointly optimize the sampling policy of each device and the device selection scheme of the BS so as to accurately monitor the dynamics of the physical process using minimum energy. This problem is formulated as an optimization problem whose goal is to minimize the weighted sum of AoI cost and energy consumption. To solve this problem, we propose a novel distributed reinforcement learning (RL) approach for the sampling policy optimization. The proposed algorithm enables edge devices to cooperatively find the global optimal sampling policy using their own local observations. Given the sampling policy, the device selection scheme can be optimized thus minimizing the weighted sum of AoI and energy consumption of all devices. Simulations with real data of PM 2.5 pollution show that the proposed algorithm can reduce the sum of AoI by up to 17.8% and 33.9% and the total energy consumption by up to 13.2% and 35.1%, compared to a conventional deep Q network method and a uniform sampling policy.
In applications of remote sensing, estimation, and control, timely communication is not always ensured by high-rate communication. This work proposes distributed age-efficient transmission policies for random access channels with $M$ transmitters. In the first part of this work, we analyze the age performance of stationary randomized policies by relating the problem of finding age to the absorption time of a related Markov chain. In the second part of this work, we propose the notion of \emph{age-gain} of a packet to quantify how much the packet will reduce the instantaneous age of information at the receiver side upon successful delivery. We then utilize this notion to propose a transmission policy in which transmitters act in a distributed manner based on the age-gain of their available packets. In particular, each transmitter sends its latest packet only if its corresponding age-gain is beyond a certain threshold which could be computed adaptively using the collision feedback or found as a fixed value analytically in advance. Both methods improve age of information significantly compared to the state of the art. In the limit of large $M$, we prove that when the arrival rate is small (below $\frac{1}{eM}$), slotted ALOHA-type algorithms are asymptotically optimal. As the arrival rate increases beyond $\frac{1}{eM}$, while age increases under slotted ALOHA, it decreases significantly under the proposed age-based policies. For arrival rates $\theta$, $\theta=\frac{1}{o(M)}$, the proposed algorithms provide a multiplicative factor of at least two compared to the minimum age under slotted ALOHA (minimum over all arrival rates). We conclude that, as opposed to the common practice, it is beneficial to increase the sampling rate (and hence the arrival rate) and transmit packets selectively based on their age-gain.
Efficient sampling and remote estimation are critical for a plethora of wireless-empowered applications in the Internet of Things and cyber-physical systems. Motivated by such applications, this work proposes decentralized policies for the real-time monitoring and estimation of autoregressive processes over random access channels. Two classes of policies are investigated: (i) oblivious schemes in which sampling and transmission policies are independent of the processes that are monitored, and (ii) non-oblivious schemes in which transmitters causally observe their corresponding processes for decision making. In the class of oblivious policies, we show that minimizing the expected time-average estimation error is equivalent to minimizing the expected age of information. Consequently, we prove lower and upper bounds on the minimum achievable estimation error in this class. Next, we consider non-oblivious policies and design a threshold policy, called error-based thinning, in which each source node becomes active if its instantaneous error has crossed a fixed threshold (which we optimize). Active nodes then transmit stochastically following a slotted ALOHA policy. A closed-form, approximately optimal, solution is found for the threshold as well as the resulting estimation error. It is shown that non-oblivious policies offer a multiplicative gain close to $3$ compared to oblivious policies. Moreover, it is shown that oblivious policies that use the age of information for decision making improve the state-of-the-art at least by the multiplicative factor $2$. The performance of all discussed policies is compared using simulations. The numerical comparison shows that the performance of the proposed decentralized policy is very close to that of centralized greedy scheduling.
This paper addresses the problem of formally verifying desirable properties of neural networks, i.e., obtaining provable guarantees that neural networks satisfy specifications relating their inputs and outputs (robustness to bounded norm adversarial perturbations, for example). Most previous work on this topic was limited in its applicability by the size of the network, network architecture and the complexity of properties to be verified. In contrast, our framework applies to a general class of activation functions and specifications on neural network inputs and outputs. We formulate verification as an optimization problem (seeking to find the largest violation of the specification) and solve a Lagrangian relaxation of the optimization problem to obtain an upper bound on the worst case violation of the specification being verified. Our approach is anytime i.e. it can be stopped at any time and a valid bound on the maximum violation can be obtained. We develop specialized verification algorithms with provable tightness guarantees under special assumptions and demonstrate the practical significance of our general verification approach on a variety of verification tasks.