The integration of advanced localization techniques in the upcoming next generation networks (B5G/6G) is becoming increasingly important for many use cases comprising contact tracing, natural disasters, terrorist attacks, etc. Therefore, emerging lightweight and passive technologies that allow accurately controlling the propagation environment, such as reconfigurable intelligent surfaces (RISs), may help to develop advance positioning solutions relying on channel statistics and beamforming. In this paper, we devise PAPIR, a practical localization system leveraging on RISs by designing a two-stage solution building upon prior statistical information on the target user equipment (UE) position. PAPIR aims at finely estimating the UE position by performing statistical beamforming, direction-of-arrival (DoA) and time-of-arrival (ToA) estimation on a given three-dimensional search space, which is iteratively updated by exploiting the likelihood of the UE position.
Despite extensive studies on motion stabilization of bipeds, they still suffer from the lack of disturbance coping capability on slippery surfaces. In this paper, a novel controller for stabilizing a bipedal motion in its sagittal plane is developed with regard to the surface friction limitations. By taking into account the physical limitation of the surface in the stabilization trend, a more advanced level of reliability is achieved that provides higher functionalities such as push recovery on low-friction surfaces and prevents the stabilizer from overreacting. The discrete event-based strategy consists of modifying the step length and time period at the beginning of each footstep in order to reestablish stability necessary conditions while taking into account the surface friction limitation as a constraint to prevent slippage. Adjusting footsteps to prevent slippage in confronting external disturbances is perceived as a novel strategy for keeping stability, quite similar to human reaction. The developed methodology consists of rough closed-form solutions utilizing elementary math operations for obtaining the control inputs, allowing to reach a balance between convergence and computational cost, which is quite suitable for real-time operations even with modest computational hardware. Several numerical simulations, including push recovery and switching between different gates on low-friction surfaces, are performed to demonstrate the effectiveness of the proposed controller. In correlation with human-gait experience, the results also reveal some physical aspects favoring stability and the fact of switching between gaits to reduce the risk of falling in confronting different conditions.
The computational capabilities of recent mobile devices enable the processing of natural features for Augmented Reality (AR), but the scalability is still limited by the devices' computation power and available resources. In this paper, we propose EdgeXAR, a mobile AR framework that utilizes the advantages of edge computing through task offloading to support flexible camera-based AR interaction. We propose a hybrid tracking system for mobile devices that provides lightweight tracking with 6 Degrees of Freedom and hides the offloading latency from users' perception. A practical, reliable and unreliable communication mechanism is used to achieve fast response and consistency of crucial information. We also propose a multi-object image retrieval pipeline that executes fast and accurate image recognition tasks on the cloud and edge servers. Extensive experiments are carried out to evaluate the performance of EdgeXAR by building mobile AR Apps upon it. Regarding the Quality of Experience (QoE), the mobile AR Apps powered by EdgeXAR framework run on average at the speed of 30 frames per second with precise tracking of only 1~2 pixel errors and accurate image recognition of at least 97% accuracy. As compared to Vuforia, one of the leading commercial AR frameworks, EdgeXAR transmits 87% less data while providing a stable 30 FPS performance and reducing the offloading latency by 50 to 70% depending on the transmission medium. Our work facilitates the large-scale deployment of AR as the next generation of ubiquitous interfaces.
To mitigate the effects of shadow fading and obstacle blocking, reconfigurable intelligent surface (RIS) has become a promising technology to improve the signal transmission quality of wireless communications by controlling the reconfigurable passive elements with less hardware cost and lower power consumption. However, accurate, low-latency and low-pilot-overhead channel state information (CSI) acquisition remains a considerable challenge in RIS-assisted systems due to the large number of RIS passive elements. In this paper, we propose a three-stage joint channel decomposition and prediction framework to require CSI. The proposed framework exploits the two-timescale property that the base station (BS)-RIS channel is quasi-static and the RIS-user equipment (UE) channel is fast time-varying. Specifically, in the first stage, we use the full-duplex technique to estimate the channel between a BS's specific antenna and the RIS, addressing the critical scaling ambiguity problem in the channel decomposition. We then design a novel deep neural network, namely, the sparse-connected long short-term memory (SCLSTM), and propose a SCLSTM-based algorithm in the second and third stages, respectively. The algorithm can simultaneously decompose the BS-RIS channel and RIS-UE channel from the cascaded channel and capture the temporal relationship of the RIS-UE channel for prediction. Simulation results show that our proposed framework has lower pilot overhead than the traditional channel estimation algorithms, and the proposed SCLSTM-based algorithm can also achieve more accurate CSI acquisition robustly and effectively.
Reconfigurable intelligent surface has attracted the attention of academia and industry as soon as it appears because it can flexibly manipulate the electromagnetic characteristics of wireless channel. Especially in the past one or two years, RIS has been developing rapidly in academic research and industry promotion and is one of the key candidate technologies for 5G-Advanced and 6G networks. RIS can build a smart radio environment through its ability to regulate radio wave transmission in a flexible way. The introduction of RIS may create a new network paradigm, which brings new possibilities to the future network, but also leads to many new challenges in the technological and engineering applications. This paper first introduces the main aspects of RIS enabled wireless communication network from a new perspective, and then focuses on the key challenges faced by the introduction of RIS. This paper briefly summarizes the main engineering application challenges faced by RIS networks, and further analyzes and discusses several key technical challenges among of them in depth, such as channel degradation, network coexistence, network coexistence and network deployment, and proposes possible solutions.
We present IOHexperimenter, the experimentation module of the IOHprofiler project, which aims at providing an easy-to-use and highly customizable toolbox for benchmarking iterative optimization heuristics such as evolutionary and genetic algorithms, local search algorithms, Bayesian optimization techniques, etc. IOHexperimenter can be used as a stand-alone tool or as part of a benchmarking pipeline that uses other components of IOHprofiler such as IOHanalyzer, the module for interactive performance analysis and visualization. IOHexperimenter provides an efficient interface between optimization problems and their solvers while allowing for granular logging of the optimization process. These logs are fully compatible with existing tools for interactive data analysis, which significantly speeds up the deployment of a benchmarking pipeline. The main components of IOHexperimenter are the environment to build customized problem suites and the various logging options that allow users to steer the granularity of the data records.
Bistatic backscatter communication (BackCom) allows passive tags to transmit over extended ranges, but at the cost of having carrier emitters either transmitting at high powers or being deployed very close to tags. In this paper, we examine how the presence of an intelligent reflecting surface (IRS) could benefit the bistatic BackCom system. We study the transmit power minimization problem at the carrier emitter, where its transmit beamforming vector is jointly optimized with the IRS phase shifts, whilst guaranteeing a required BackCom performance. A unique feature in this system setup is the multiple IRS reflections experienced by signals traveling from the carrier emitter to the reader, which renders the optimization problem highly nonconvex. Therefore, we propose algorithms based on the minorization-maximization and alternating optimization techniques to obtain approximate solutions for the joint design. We also propose low-complexity algorithms based on successive optimization of individual phase shifts. Our results reveal considerable transmit power savings in both single-tag and multi-tag systems, even with moderate IRS sizes, which may be translated to significant range improvements using the original transmit power or a reduction of the reliance of tags on carrier emitters located at close range.
Simultaneously transmitting and reflecting reconfigurable intelligent surface (STAR-RIS) is a promising technology to achieve full-space coverage. This paper investigates the resource allocation problem in a STAR-RIS-assisted multi-carrier communication network. To maximize the system sum-rate, a joint optimization problem for orthogonal multiple access (OMA) is first formulated, which is a mixed-integer non-linear programming problem. To solve this challenging problem, we first propose a channel assignment scheme utilizing matching theory and then invoke the alternating optimization-based method to optimize the resource allocation policy and beamforming vectors iteratively. Furthermore, the sum-rate maximization problem for non-orthogonal multiple access (NOMA) is investigated. To efficiently solve it, we first propose a location-based matching algorithm to determine the sub-channel assignment, where a transmitted user and a reflected user are grouped on a sub-channel. Then, a three-step approach is proposed, where the decoding orders, beamforming-coefficient vectors, and power allocation are optimized by employing semidefinite programming, convex upper bound approximation, and geometry programming, respectively. Numerical results unveil that: 1) For OMA, a general design that includes same-side user-pairing for channel assignment is preferable, while for NOMA, the proposed transmission-and-reflection scheme can achieve near-optimal performance. 2) The STAR-RIS-aided NOMA network significantly outperforms the networks employing conventional RISs and OMA.
Visual information extraction (VIE) has attracted considerable attention recently owing to its various advanced applications such as document understanding, automatic marking and intelligent education. Most existing works decoupled this problem into several independent sub-tasks of text spotting (text detection and recognition) and information extraction, which completely ignored the high correlation among them during optimization. In this paper, we propose a robust visual information extraction system (VIES) towards real-world scenarios, which is a unified end-to-end trainable framework for simultaneous text detection, recognition and information extraction by taking a single document image as input and outputting the structured information. Specifically, the information extraction branch collects abundant visual and semantic representations from text spotting for multimodal feature fusion and conversely, provides higher-level semantic clues to contribute to the optimization of text spotting. Moreover, regarding the shortage of public benchmarks, we construct a fully-annotated dataset called EPHOIE (//github.com/HCIILAB/EPHOIE), which is the first Chinese benchmark for both text spotting and visual information extraction. EPHOIE consists of 1,494 images of examination paper head with complex layouts and background, including a total of 15,771 Chinese handwritten or printed text instances. Compared with the state-of-the-art methods, our VIES shows significant superior performance on the EPHOIE dataset and achieves a 9.01% F-score gain on the widely used SROIE dataset under the end-to-end scenario.
Deep learning-based semi-supervised learning (SSL) algorithms have led to promising results in medical images segmentation and can alleviate doctors' expensive annotations by leveraging unlabeled data. However, most of the existing SSL algorithms in literature tend to regularize the model training by perturbing networks and/or data. Observing that multi/dual-task learning attends to various levels of information which have inherent prediction perturbation, we ask the question in this work: can we explicitly build task-level regularization rather than implicitly constructing networks- and/or data-level perturbation-and-transformation for SSL? To answer this question, we propose a novel dual-task-consistency semi-supervised framework for the first time. Concretely, we use a dual-task deep network that jointly predicts a pixel-wise segmentation map and a geometry-aware level set representation of the target. The level set representation is converted to an approximated segmentation map through a differentiable task transform layer. Simultaneously, we introduce a dual-task consistency regularization between the level set-derived segmentation maps and directly predicted segmentation maps for both labeled and unlabeled data. Extensive experiments on two public datasets show that our method can largely improve the performance by incorporating the unlabeled data. Meanwhile, our framework outperforms the state-of-the-art semi-supervised medical image segmentation methods. Code is available at: //github.com/Luoxd1996/DTC
In recent years with the rise of Cloud Computing (CC), many companies providing services in the cloud, are empowered a new series of services to their catalog, such as data mining (DM) and data processing, taking advantage of the vast computing resources available to them. Different service definition proposals have been proposed to address the problem of describing services in CC in a comprehensive way. Bearing in mind that each provider has its own definition of the logic of its services, and specifically of DM services, it should be pointed out that the possibility of describing services in a flexible way between providers is fundamental in order to maintain the usability and portability of this type of CC services. The use of semantic technologies based on the proposal offered by Linked Data (LD) for the definition of services, allows the design and modelling of DM services, achieving a high degree of interoperability. In this article a schema for the definition of DM services on CC is presented, in addition are considered all key aspects of service in CC, such as prices, interfaces, Software Level Agreement, instances or workflow of experimentation, among others. The proposal presented is based on LD, so that it reuses other schemata obtaining a best definition of the service. For the validation of the schema, a series of DM services have been created where some of the best known algorithms such as \textit{Random Forest} or \textit{KMeans} are modeled as services.