In the Internet of Things, learning is one of most prominent tasks. In this paper, we consider an Internet of Things scenario where federated learning is used with simultaneous transmission of model data and wireless power. We investigate the trade-off between the number of communication rounds and communication round time while harvesting energy to compensate the energy expenditure. We formulate and solve an optimization problem by considering the number of local iterations on devices, the time to transmit-receive the model updates, and to harvest sufficient energy. Numerical results indicate that maximum ratio transmission and zero-forcing beamforming for the optimization of the local iterations on devices substantially boost the test accuracy of the learning task. Moreover, maximum ratio transmission instead of zero-forcing provides the best test accuracy and communication round time trade-off for various energy harvesting percentages. Thus, it is possible to learn a model quickly with few communication rounds without depleting the battery.
This paper studies decentralized federated learning algorithms in wireless IoT networks. The traditional parameter server architecture for federated learning faces some problems such as low fault tolerance, large communication overhead and inaccessibility of private data. To solve these problems, we propose a Decentralized-Wireless-Federated-Learning algorithm called DWFL. The algorithm works in a system where the workers are organized in a peer-to-peer and server-less manner, and the workers exchange their privacy preserving data with the anolog transmission scheme over wireless channels in parallel. With rigorous analysis, we show that DWFL satisfies $(\epsilon,\delta)$-differential privacy and the privacy budget per worker scale as $\mathcal{O}(\frac{1}{\sqrt{N}})$, in contrast with the constant budget in the orthogonal transmission approach. Furthermore, DWFL converges at the same rate of $\sqrt{\frac{1}{TN}}$ as the best known centralized algorithm with a central parameter server. Extensive experiments demonstrate that our algorithm DWFL also performs well in real settings.
Major bottlenecks of large-scale Federated Learning(FL) networks are the high costs for communication and computation. This is due to the fact that most of current FL frameworks only consider a star network topology where all local trained models are aggregated at a single server (e.g., a cloud server). This causes significant overhead at the server when the number of users are huge and local models' sizes are large. This paper proposes a novel edge network architecture which decentralizes the model aggregation process at the server, thereby significantly reducing the aggregation latency of the whole network. In this architecture, we propose a highly-effective in-network computation protocol consisting of two components. First, an in-network aggregation process is designed so that the majority of aggregation computations can be offloaded from cloud server to edge nodes. Second, a joint routing and resource allocation optimization problem is formulated to minimize the aggregation latency for the whole system at every learning round. The problem turns out to be NP-hard, and thus we propose a polynomial time routing algorithm which can achieve near optimal performance with a theoretical bound. Numerical results show that our proposed framework can dramatically reduce the network latency, up to 4.6 times. Furthermore, this framework can significantly decrease cloud's traffic and computing overhead by a factor of K/M, where K is the number of users and M is the number of edge nodes, in comparison with conventional baselines.
In this work we treat the unsourced random access problem on a Rayleigh block-fading AWGN channel with multiple receive antennas. Specifically, we consider the slowly fading scenario where the coherence block-length is large compared to the number of active users and the message can be transmitted in one coherence block. Unsourced random access refers to a form of grant-free random access where users are considered to be a-priori indistinguishable and the receiver recovers a list of transmitted messages up to permutation. In this work we show that, when the coherence block length is large enough, a conventional approach based on the transmission of non-orthogonal pilot sequences with subsequent channel estimation and Maximum-Ratio-Combining (MRC) provides a simple energy-efficient solution whose performance can be well approximated in closed form. Furthermore, we analyse the MRC step when successive interference cancellation (SIC) is done in groups, which allows to strike a balance between receiver complexity and reduced transmit powers. Finally, we investigate the impact of power control policies taking into account the unique nature of massive random access, including short message lengths, uncoordinated transmission, a very large amount of concurrent transmitters with unknown identities, channel estimation errors and decoding errors. As a byproduct we also present an extension of the MMV-AMP algorithm which allows to treat pathloss coefficients as deterministic unknowns by performing maximum likelihood estimation in each step of the MMV-AMP algorithm.
In this paper, a novel intelligent reflecting surface (IRS)-assisted wireless powered communication network (WPCN) architecture is proposed for power-constrained Internet-of-Things (IoT) smart devices, where IRS is exploited to improve the performance of WPCN under imperfect channel state information (CSI). We formulate a hybrid access point (HAP) transmit energy minimization problem by jointly optimizing time allocation, HAP energy beamforming, receiving beamforming, user transmit power allocation, IRS energy reflection coefficient and information reflection coefficient under the imperfect CSI and non-linear energy harvesting model. On account of the high coupling of optimization variables, the formulated problem is a non-convex optimization problem that is difficult to solve directly. To address the above-mentioned challenging problem, alternating optimization (AO) technique is applied to decouple the optimization variables to solve the problem. Specifically, through AO, time allocation, HAP energy beamforming, receiving beamforming, user transmit power allocation, IRS energy reflection coefficient and information reflection coefficient are divided into three sub-problems to be solved alternately. The difference-of-convex (DC) programming is uesd to solve the non-convex rank-one constraint in solving IRS energy reflection coefficient and information reflection coefficient. Numerical simulations verify the superiority of the proposed optimization algorithm in decreasing HAP transmit energy compared with other benchmark schemes.
In this paper, we study the problem of minimizing the uplink aggregate transmit power subject to the users' minimum data rate and peak power constraint on each sub-channel for multi-cell wireless networks. To address this problem, a distributed sub-optimal joint power and rate control algorithm called JPRC is proposed, which is applicable to both non-orthogonal frequency-division multiple access (NOMA) and orthogonal frequency-division multiple access (OFDMA) schemes. Employing JPRC, each user updates its transmit power using only local information. Simulation results illustrate that the JPRC algorithm can reach a performance close to that obtained by the optimal solution via exhaustive search, with the NOMA scheme achieving a 59\% improvement on the aggregate transmit power over the OFDMA counterpart. It is also shown that the JPRC algorithm can outperform existing distributed power control algorithms.
We investigate the age-of-information (AoI) in the context of random access networks, in which transmitters need to send a sequence of information packets to the intended receivers over a shared spectrum. Due to interference, the dynamics at the link pairs will interact with each other over both space and time, and the effects of these spatiotemporal interactions on the AoI are not well understood. In this paper, we straddle queueing theory and stochastic geometry to establish an analytical framework, that accounts for the interplay between the temporal traffic attributes and spatial network topology, for such a study. Specifically, we derive accurate and tractable expressions to quantify the network average AoI as well as the outage probability of peak AoI. Besides, we develop a decentralized channel access policy that exploits the local observation at each node to make transmission decisions that minimize the AoI. Our analysis reveals that when the packet transmissions are scheduled in a last-come first-serve (LCFS) order, whereas the newly incoming packets can replace the undelivered ones, depending on the deployment density, there may or may not exist a tradeoff on the packet arrival rate that minimizes the network average AoI. Moreover, the slotted ALOHA protocol is shown to be instrumental in reducing the AoI when the packet arrival rates are high, yet it cannot contribute to decreasing the AoI in the regime of infrequent packet arrivals. The numerical results also confirm the efficacy of the proposed scheme, where the gain is particularly pronounced when the network grows in size because our method is able to adapt the channel access probabilities with the change of ambient environment.
The combination of energy harvesting (EH), cognitive radio (CR), and non-orthogonal multiple access (NOMA) is a promising solution to improve energy efficiency and spectral efficiency of the upcoming beyond fifth generation network (B5G), especially for support the wireless sensor communications in Internet of things (IoT) system. However, how to realize intelligent frequency, time, and energy resource allocation to support better performances is an important problem to be solved. In this paper, we study joint spectrum, energy, and time resource management for the EH-CR-NOMA IoT systems. Our goal is to minimize the number of data packets losses for all secondary sensing users (SSU), while satisfying the constraints on the maximum charging battery capacity, maximum transmitting power, maximum buffer capacity, and minimum data rate of primary users (PU) and SSUs. Due to the non-convexity of this optimization problem and the stochastic nature of the wireless environment, we propose a distributed multidimensional resource management algorithm based on deep reinforcement learning (DRL). Considering the continuity of the resources to be managed, the deep deterministic policy gradient (DDPG) algorithm is adopted, based on which each agent (SSU) can manage its own multidimensional resources without collaboration. In addition, a simplified but practical action adjuster (AA) is introduced for improving the training efficiency and battery performance protection. The provided results show that the convergence speed of the proposed algorithm is about 4 times faster than that of DDPG, and the average number of packet losses (ANPL) is about 8 times lower than that of the greedy algorithm.
Federated learning enables multiple parties to collaboratively train a machine learning model without communicating their local data. A key challenge in federated learning is to handle the heterogeneity of local data distribution across parties. Although many studies have been proposed to address this challenge, we find that they fail to achieve high performance in image datasets with deep learning models. In this paper, we propose MOON: model-contrastive federated learning. MOON is a simple and effective federated learning framework. The key idea of MOON is to utilize the similarity between model representations to correct the local training of individual parties, i.e., conducting contrastive learning in model-level. Our extensive experiments show that MOON significantly outperforms the other state-of-the-art federated learning algorithms on various image classification tasks.
Federated learning (FL) is a machine learning setting where many clients (e.g. mobile devices or whole organizations) collaboratively train a model under the orchestration of a central server (e.g. service provider), while keeping the training data decentralized. FL embodies the principles of focused data collection and minimization, and can mitigate many of the systemic privacy risks and costs resulting from traditional, centralized machine learning and data science approaches. Motivated by the explosive growth in FL research, this paper discusses recent advances and presents an extensive collection of open problems and challenges.
In this paper, an interference-aware path planning scheme for a network of cellular-connected unmanned aerial vehicles (UAVs) is proposed. In particular, each UAV aims at achieving a tradeoff between maximizing energy efficiency and minimizing both wireless latency and the interference level caused on the ground network along its path. The problem is cast as a dynamic game among UAVs. To solve this game, a deep reinforcement learning algorithm, based on echo state network (ESN) cells, is proposed. The introduced deep ESN architecture is trained to allow each UAV to map each observation of the network state to an action, with the goal of minimizing a sequence of time-dependent utility functions. Each UAV uses ESN to learn its optimal path, transmission power level, and cell association vector at different locations along its path. The proposed algorithm is shown to reach a subgame perfect Nash equilibrium (SPNE) upon convergence. Moreover, an upper and lower bound for the altitude of the UAVs is derived thus reducing the computational complexity of the proposed algorithm. Simulation results show that the proposed scheme achieves better wireless latency per UAV and rate per ground user (UE) while requiring a number of steps that is comparable to a heuristic baseline that considers moving via the shortest distance towards the corresponding destinations. The results also show that the optimal altitude of the UAVs varies based on the ground network density and the UE data rate requirements and plays a vital role in minimizing the interference level on the ground UEs as well as the wireless transmission delay of the UAV.