Next Article in Journal
SwimmerNET: Underwater 2D Swimmer Pose Estimation Exploiting Fully Convolutional Neural Networks
Next Article in Special Issue
High-Precision AI-Enabled Flood Prediction Integrating Local Sensor Data and 3rd Party Weather Forecast
Previous Article in Journal
Unsupervised Learning of Disentangled Representation via Auto-Encoding: A Survey
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LP-MAB: Improving the Energy Efficiency of LoRaWAN Using a Reinforcement-Learning-Based Adaptive Configuration Algorithm

by
Benyamin Teymuri
1,*,
Reza Serati
1,
Nikolaos Athanasios Anagnostopoulos
2 and
Mehdi Rasti
1,3
1
Department of Computer Engineering, Amirkabir University of Technology, Tehran P.O. Box 15875-4413, Iran
2
Faculty of Computer Science and Mathematics, University of Passau, 94032 Passau, Germany
3
Centre for Wireless Communications, University of Oulu, 90570 Oulu, Finland
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 2363; https://doi.org/10.3390/s23042363
Submission received: 12 January 2023 / Revised: 13 February 2023 / Accepted: 15 February 2023 / Published: 20 February 2023
(This article belongs to the Special Issue Intelligent IoT and Wireless Communications)

Abstract

:
In the Internet of Things (IoT), Low-Power Wide-Area Networks (LPWANs) are designed to provide low energy consumption while maintaining a long communications’ range for End Devices (EDs). LoRa is a communication protocol that can cover a wide range with low energy consumption. To evaluate the efficiency of the LoRa Wide-Area Network (LoRaWAN), three criteria can be considered, namely, the Packet Delivery Rate (PDR), Energy Consumption (EC), and coverage area. A set of transmission parameters have to be configured to establish a communication link. These parameters can affect the data rate, noise resistance, receiver sensitivity, and EC. The Adaptive Data Rate (ADR) algorithm is a mechanism to configure the transmission parameters of EDs aiming to improve the PDR. Therefore, we introduce a new algorithm using the Multi-Armed Bandit (MAB) technique, to configure the EDs’ transmission parameters in a centralized manner on the Network Server (NS) side, while improving the EC, too. The performance of the proposed algorithm, the Low-Power Multi-Armed Bandit (LP-MAB), is evaluated through simulation results and is compared with other approaches in different scenarios. The simulation results indicate that the LP-MAB’s EC outperforms other algorithms while maintaining a relatively high PDR in various circumstances.

1. Introduction

The Internet of Things (IoT) refers to the network of physical objects—“things”—embedded with sensors and software that use the internet to transmit and receive data. With several billion connected IoT devices today, experts expect this number to grow to 22 billion by 2025 [1]. Various requirements are essential to IoT applications, including a long transmission range, low energy consumption, and a cost-effective design. Short-range communication technologies, such as Bluetooth, ZigBee, and Wi-Fi, are unsuitable for long-range communication. In contrast, cellular communication networks, such as 3G and 4G cellular networks, can provide a much wider transmission range at the expense of draining the ED’s battery quickly. Therefore, in order to meet the needs of IoT applications, long-range and low-power networks are required. Low-Power Wide-Area Networks (LPWANs) are wireless networks that enable long-range communications with low data rates and low energy consumption. An LPWAN can provide a transmission range of 40 and 10 kilometers (km) in sub-urban and urban environments, respectively [2], with a maximum battery life of ten years [3].
As shown in Figure 1, compared with other technologies, LPWAN was mainly designed for IoT applications that require the non-periodic transmission of short messages in extended radio coverage. LPWANs can operate on both licensed and unlicensed frequencies, such as LoRa [4], NarrowBand IoT (NB-IoT) [5,6], Sigfox [7], and LTE-M [8]. LoRa (short for Long Range) is a physical proprietary radio communication technique that utilizes spread-spectrum modulation derived from the Chirp Spread Spectrum (CSS) technology. LoRa enables long-range and low-power communication, where packets of up to 256 bytes can be sent with each message transfer at sub-GHz frequencies [9]. A number of network evaluation criteria can be used to assess the efficiency of LPWANs, including Packet Delivery Ratio (PDR) and Energy Consumption (EC).
In the LoRa network, several transmission parameters must be configured before a connection can be established. These transmission parameters are: Spreading Factor (SF), Transmission Power (TP), Carrier Frequency (CF), Coding Rate (CR), and BandWidth (BW). Various values can be assigned to each of these parameters. By configuring each transmission parameter to a different value, a relatively large state space of configurations, consisting of several hundreds of states, exists. The selection of each of these states can affect the network evaluation criteria, such as PDR and EC. As an example, if ED transmits with SF7 and TP2, i.e., using the minimum spreading factor and transmission power, respectively, the least noise sensitivity, the lowest transmission delay, and the least coverage area are achieved. In this regard, it is crucial to find the optimal configuration of the transmission parameters [11].
The Adaptive Data Rate (ADR) algorithm is a mechanism to adjust the transmission parameters of LoRa EDs with the objective of improving the values achieved for the network evaluation criteria. This mechanism was first proposed in the LoRaWAN specification v1.1 [4]. In this mechanism, the Network Server (NS) uses the highest Signal-to-Interference-plus-Noise Ratio (SINR) for link quality assessment after receiving the last 20 packets from each ED.
Machine Learning (ML) algorithms, which provide a self-learning process, are divided into supervised, unsupervised, and Reinforcement Learning (RL). There is no need for training data sets in RL-based methods, as learning happens through interaction with the environment. The RL agent can perceive and analyze its environment, take actions, and learn through trial and error [12]. Therefore, RL-based methods are the best choice for low-complexity network deployment. The literature review shows that RL techniques can improve resource allocation performance in LoRaWAN by allowing each ED to select the most appropriate configuration of transmission parameters.
In [13,14], an RL-based non-stationary resource allocation algorithm called LoRa-MAB is proposed, based on an adversarial environment suitable for LoRaWAN deployments. However, as a result of the long exploration process of the approach, LoRa-MAB experiences a high EC. Moreover, in the distributed solutions, the transmission parameters configuration is done on the EDs’ side, which are resource-constrained devices that are not designed to handle the computational overhead. Since many IoT applications employ battery-powered EDs that are used in large numbers for lengthy periods of time, EC plays an essential role in measuring the performance of such algorithms.
To meet the desired communication performance, it is challenging to determine the proper configuration of the transmission parameters so that EC is minimized and PDR is enhanced. In our previous works, [15,16], we have used centralized and distributed ADR approaches, respectively, to find a solution to this problem. In [15], a low-complexity ADR scheme was proposed, in which the NS attempts to obtain the optimal transmission parameters of the EDs, not by considering the history of the last 20 packets received, but by considering only the current environmental conditions of the communication, based on the transmission parameters of the last packet received. In [16], each ED individually tries to find the optimum transmission parameter configuration with the help of ML approaches.
This article focuses on improving the EC by combining non-stationary adversarial algorithms, suitable for the LoRa environment, with stochastic algorithms, which have the advantage of a short exploration time. Our work also reduces the overall computational overhead by migrating the implementation of the scheme to the NS without imposing any changes on the protocol design. Thereby, we improve upon our previous work that has been presented in [16], by reducing the overall EC and improving the PDR. The main contributions of our article can be summarised as follows:
  • In this paper, we propose a Low-Power Multi-Armed Bandit (henceforth, LP-MAB) ADR mechanism, a centralized adaptive configuration scheme in LoRaWAN. In particular, we employ EXPonential weights for EXPloration and EXPloitation (EXP3) along with the Successive Elimination (SE) technique. As a result, the proposed solution combines non-stationary adversarial and stochastic methods.
  • In order to assess the performance of LP-MAB, simulation results for LP-MAB and various other ADR schemes, namely, ADR-MAX [4], ADR-AVG [11], No-ADR, and ADR-Lite [15], have been compared. These results indicate that the LP-MAB’s EC outperforms other algorithms while maintaining a relatively high PDR in various circumstances, considering both stationary and mobile EDs. This is achieved by determining the effects of various parameters and conditions such as channel noise, simulation time, network size, and the number of daily sent packets by each ED. We also consider both an urban and a sub-urban environment for all the examined scenarios, while also studying the impact of network densification, i.e., the number of EDs in the simulation.
In general, this article follows the following structure: Background information and related works are presented in Section 2 and Section 3, respectively. Section 4 describes our LP-MAB algorithm. Then, the simulation setup and our results are presented in Section 5 and Section 6, respectively. Finally, Section 7 concludes this work.

2. Background

In this section, following a review of the LoRa and LoRaWAN protocol stack, an EC model will be discussed for LoRa EDs, since we are primarily concerned with optimizing the energy consumption of the LoRa network. This section will be concluded with a thorough description of the Adaptive Data Rate (ADR) mechanism.

2.1. LoRa Overview

The LoRa architecture is based on a star-of-stars topology, consisting of four components, i.e., EDs, GateWays (GWs), the NS, and an application server, as shown in Figure 2. Multiple GWs are located at different locations to receive the uplink data from EDs. As uplink messages are broadcast over the network, EDs are not assigned to a specific GW. The received LoRa packets by the GW, are then relayed to the NS over a backbone network, which can, for example, be implemented using IP over Ethernet, cellular, Wi-Fi, or 2.4-GHz radio communication. Packets are then routed to the relevant application by the NS, e.g., by using an Ethernet connection. Both uplink communication (ED to application) and downlink communication (application to ED) can be performed by the NS. The LoRa specification is documented in more detail in [4].
In LoRa, communication link quality is impacted by several transmission parameters [17], which are:
  • SF: SF can be described as the number of symbols that can appear in a single bit of transferred data, which can be set in the range of 7 to 12, depending on the environmental conditions between the ED and the GW.
  • TP: The TP of LoRa radio can be configured between 2 dBm and 14 dBm in steps of 3 dBm (the setup of each transmitter may vary). When the TP increases, the signal range increases while the battery lifetime of the EDs shortens, and vice versa.
  • CF: Lower frequency ranges result in decreased receiver antenna size, BW capacity, and latency, while increasing the coverage. A LoRa communication network can operate over radio frequency bands below 1 GHz, including frequencies of 433, 868, and 915 MHz, with different step sizes depending on the regulation rules.
  • CR: CR improves LoRa communication link robustness with Cyclic Redundancy Check (CRC). The values of CR may vary in the range of { 4 5 , 4 6 , 4 7 , 4 8 } .Adding such error correction coding will increase the transmission overhead, which can affect the performance.
Various parameters and conditions, such as BW, channel noise, simulation time, network size, packet length, ED speed, and the number of daily sent packets by each ED, can affect the network performance. The value of BW, in particular, may vary in the range of { 125 , 250 , 500 } kHz.

2.2. An EC Model for LoRa EDs

To have a realistic EC model for LoRa EDs, we assume the same sequence of working modes for the LoRa and LoRaWAN sensor nodes as the one presented in [18]. This sequence of working modes is illustrated in Figure 3. Thus, the total energy consumed by the EDs, E Total , is calculated as follows:
E Total = E Sleep + E Active ,
where E Sleep and E Active is the energy consumed by the EDs during the sleep and the active modes, respectively. The total EC of EDs in the active mode is calculated by the summation of the energy consumed during the relevant working modes of the EDs (from the ones shown in Figure 3). Thus, E Active is calculated as shown in the following equation [19]:
E Active = E W U + E m + E proc + E W U T + E T o A + E R .
Hence, in Equation (2), E W U , E m , E proc , E W U T , E T o A and E R , describe the EC of the wake-up of the device, the data measurement, the microcontroller processing, the LoRa transceiver’s wake-up, the transmission, and the reception mode, respectively, as shown in Figure 3. The consumed energy in the data transmission mode, E T o A , is expressed as follows [18]:
E T o A = P O N f M C U + P T o A × T T o A .
Here, P O N ( f M C U ) is the microcontroller’s EC depending on its processor frequency f M C U , while P T o A and T T o A are the consumed power in the transmission mode and its time duration, respectively [18]. The power utilization of LoRa sensors in the active mode depends on the Time-on-Air (ToA) duration. An ED requires time to transfer both the preamble and the payload message, i.e., T Preamble and T Payload , respectively [19], which leads to the following equation:
T ToA = T Preamble + T Payload .
T Preamble can be obtained as follows:
T Preamble = 4.25 + N P × T Symbol .
Let the number of preamble symbols be N P , and the symbol’s length be denoted by T Symbol , which is defined as the duration time for transmitting 2 S F chirps. Note that the B W is equal to the chirp rate. The symbol duration is calculated so that:
T Symbol = 2 S F B W .
Moreover, T Payload (in seconds) is calculated using this equation:
T Payload = T Symbol × N Payload .
N Payload is the number of symbols transmitted as message payload, except the preamble, specified as [19]:
N Payload = 8 + max Θ ( P L , S F ) Γ ( S F ) × 1 C R , 0 .
We use the following equation to calculate Θ ( P L , S F ) :
Θ ( P L , S F ) = 8 × P L 4 × S F + 16 + 28 20 × H .
In this equation, H is zero when the header is enabled, and H is equal to one when there is no header present. Γ ( S F ) can be calculated as S F 2 × D E , wherein D E is set as one when the low data rate optimization is enabled; otherwise, D E is set to zero.
As observed from Equations (1)–(9), higher SF values significantly increase the EC: Higher SF values exponentially increase T Symbol (Equation (6)), leading to long T Preamble (Equation (5)) and, thus, T T o A (Equation (4)). The increase in T T o A then leads to E T o A being higher (Equation (3)), making the E Active (Equation (2)) and E Total (Equation (1)) larger. Therefore, compared to lower SFs, transmitting the same amount of data with a higher SF requires a much higher T T o A and, thus, a much higher EC.

2.3. The ADR Mechanism

There are two methods to control the transmission parameters in LoRaWAN: distributed and centralized approaches [11]. In the distributed method, each ED tries to configure its own transmission parameters based on the NS’s ACKnowledgment (ACK) regarding the reception or non-reception of the uplink messages, e.g., in the ADR-AVG [11], and ADR-Lite [15] schemes. In the centralized method with a global knowledge of the network, the NS tries to configure the transmission parameters of each link according to the ACK messages individually, e.g., in the MIX-MAB [16], and LoRa-MAB [13] schemes. Our work significantly extends and revises the MIX-MAB work.
The Adaptive Data Rate (ADR) algorithm is a mechanism to configure the transmission parameters of EDs with the aim of improving PDR and EC as the two primary performance metrics. Through a centralized manner of configuring the EDs’ transmission parameters, ADR aims to optimize data rate and ED lifetime. ADR, for this purpose, evaluates the link-budget estimation between EDs and GWs in the uplink messages. In this regard, Media Access Control (MAC) commands will be used to control the data rate of the ED if the ADR bit is set. Each ED and the NS may set and unset the relevant ADR bit on demand. When the ADR bit is not set, the NS will not configure the ED’s transmission parameters, regardless of the signal quality received by the end device. However, the ADR scheme should be enabled whenever possible to maximize the network capacity and battery lifetime of EDs. More details about the ADR mechanism are provided in [20,21].

3. Related Works

There have been several studies to improve LoRaWAN performance, focusing on statistical and mathematical models [22], the effect of the number of GWs [23], optimization algorithms [24,25], and machine learning techniques [16]. Configuring the LoRaWAN transmission parameters to address scalability has been presented in [11]. In recent years, the ADR approach has been proposed in version 1.1 of the LoRaWAN Specification [4]. In [4], the maximum value of the latest twenty received packets’ Signal to Interference and Noise Ratio (SINR) is taken into account as an indicator to evaluate the link quality. However, in this optimistic approach, environmental changes cannot realistically be considered.
The proposed methods in [11] and [24] improved the original ADR mechanism by using the details, i.e., the SINR, of the last 20 received packets to adjust the transmission parameters. More specifically, by using the average SINR value of the last twenty packets in [11], in a method called ADR-AVG, instead of the maximum value of the SINR, resulted in better performance. In [24], the authors proposed a new ADR, called ADR-OWA, using the Ordered Weighted Averaging (OWA) function. However, when channel saturation in either an urban or a sub-urban environment is low, the EC of ADR-OWA will be higher than ADR-AVG.
In addition to the emerging ML techniques, the new IoT ED requirements for more reliability, as well as low latency demands, led to the development of more efficient optimization mechanisms [14]. Self-resource management is critical to improving the battery lifetime for LoRa EDs. Some works focused on using ML techniques, such as RL, to enable EDs to use innovative and inherently distributed techniques for the management of the transmission parameters [13,14]. MAB [13,14] and Q-learning [26] are two RL algorithms used in the literature to propose distributed radio resource allocation in LoRaWAN. In [26], the authors use RL by offering a Q-learning model combined with Carrier-Sense Multiple Access with Collision Avoidance (CSMA/CA), to decrease the collision rate and improve the PDR. However, in addition to the increased EC using the method in [26], Q-Learning requires a database to save its processing data, a requirement that is not compatible with resource-constrained IoT EDs.
The LoRa-MAB algorithm proposed in [13,14] is based on EXP3. As a non-stationary adversarial method, this approach suffers from a rather long exploration process resulting in high EC. In particular, in the aforementioned distributed approach, the EDs and the NS must frequently communicate, resulting in reducing the battery lifetime. SE is a non-stationary stochastic MAB-based algorithm presented in [27], for which, however, the adversarial environment of LoRa has not been taken into account.

4. Our Proposed LP-MAB Algorithm

This section proposes a centralized adaptive configuration algorithm in LoRaWAN. In our newly proposed RL-based adaptive configuration algorithm, the NS does not need to be provided with a predefined dataset, as it will learn by interacting with the EDs. In LP-MAB, the agent is the NS interacting with the environment, including EDs, to perform an action that can be defined as the determination of the set of transmission parameters to which an ED should be configured. The NS tries to achieve the optimum action, i.e., the optimal set of transmission parameters, for each ED by learning based on the relevant reward, which is based on the reception of the ACK messages. More specifically, a LoRa ED configures its transmission parameters based on the NS’s selected action. If the NS receives the packet, it sends back a confirmation ACK message to the ED, assigns a reward (which will be defined later) to the selected action, and uses it for the subsequent transmission parameters’ index. We model the adaptive configuration scheme utilised by the NS as a MAB problem, an RL-based technique, and formulate it using k multi-armed bandits, where k represents the total configuration’s state space. An agent selects from k different actions and, each time, receives a reward based on its chosen action.
Three general categories, of stochastic, adversarial, and switching bandit algorithms, can be used to address the MAB problems. EXP3 is a category of non-stationary adversarial MAB problems. LoRa can be placed in this category because the selection by two or more EDs of the same transmission parameter values, such as an equal SF, affects the transmission performance of all the relevant EDs. Stochastic MAB algorithms such as SE are unsuitable for LoRaWAN due to its adversarial nature. The long exploration process of EXP3 results in a high convergence time. On the other hand, the SE algorithm has the advantage of short-term exploration. So, inspired by the benefits of the EXP3 and SE algorithms used in [13,14] and [27], respectively, we combine these two approaches and propose a new algorithm called LP-MAB.
As can be seen in Algorithm 1, at the beginning, we assume there are | U | EDs in the simulation forming the set U = { E D 1 , E D 2 , , E D U } . NS aims to maximize the PDR of the network, while keeping the EC at the minimum possible value by learning to select the optimum transmission parameter set for u U . Assuming that each action is a vector of four transmission parameters, a k u = { S F k , T P k , C F k , C R k } denotes the kth action for the uth ED, in which S F k , T P k , C F k , and C R k are the values of SF, TP, CF, and CR in the kth action, respectively. We assume that there are | A | actions, whose set is denoted by A = { a 0 u , a 1 u , , a | A | 2 u , a | A | 1 u } . LP-MAB allows configuring the CF and CR in addition to the SF and TP, unlike [4,11,24], making the action set size | A | rather large. Actions are sorted in A in ascending order based on their EC according to Equation (8). Let N a k u indicate how many times the NS selects the kth action for the uth ED. W a k u ( t ) and P a k u ( t ) are the weight of the kth action and the probability of selecting the kth action for the uth ED at the simulation time of transmission period t, respectively. The transmission period t is initialised to zero. Thus, the visual representation of Algorithm 1 can be seen in Figure 4.
Algorithm 1: Initialization of LP-MAB.
Sensors 23 02363 i001
Our proposed LP-MAB algorithm is composed of two phases, exploration and exploitation, as described in Section 4.1 and Section 4.2, respectively.

4.1. Exploration Phase of the LP-MAB Algorithm

The goal of the first phase of our algorithm is to update the values of the W a k u ( t ) and P a k u ( t ) so that, in the exploitation phase, the NS can select the optimum configuration for the uth ED based on the information gathered from the environment. We assign probabilities to each action to obtain their weights (lines 10–12 of Algorithm 2), so that we can make a trade-off between exploration and exploitation. At the start of the simulation, the NS selects the first action for the uth ED, i.e., a k = 0 u , and then increase the value of N a k u by one, as shown in Figure 5  ( I ) . After the ACK reception or non-reception for the chosen action (demonstrated in Figure 5 by ✓ and ✗, respectively), the NS updates the reward value, R a k u ( t ) , based on line 16 of Algorithm 1. In this multi-reward strategy, the reception of the ACK for the action with the highest TP ( T P = 14 ), will be far less rewarded than the reception of the ACK for the action with the lowest TP ( T P = 2 ), aiming to minimize the EC as much as possible.
Algorithm 2: Exploration Phase of LP-MAB.
Sensors 23 02363 i002
After calculating the reward, the NS updates the weight and probability of the action according to the lines 10–12 of Algorithm 2, which have been extracted from the EXP3 algorithm. As a rule, the summation of all probabilities is equal to one. So the action’s probability should be normalized (line 11 of Algorithm 2). For the next packet of the uth ED, the NS selects the action a 1 u (Figure 5  ( I I ) ). This procedure for the uth ED continues until the NS has selected all the actions once (Figure 5  ( I V ) and Figure 6  ( I ) ).
The exploration phase of the LP-MAB scheme was adopted from the SE algorithm. As can be seen in line 3 of Algorithm 2, this phase is repeated for multiple rounds, up to the L EXP value, for each ED. A novel feature of our work in comparison to [16], is that we take into account, in the exploration phase, the potential occurrence of environmental changes, such as changes in the number of EDs and GWs, through the utilisation of L EXP . According to line 12 of Algorithm 1, L EXP can be calculated by dividing the total number of EDs by the product of the multiplication of the number of GWs by the constant value 100. The use of this constant value is intended to reduce the value of L EXP , which should not be too large, as it would significantly prolong the exploratory phase.
In general, as the number of EDs increases, the level of interference in a fixed network area grows higher, and thus it becomes necessary to extend the exploration phase, which is achieved through the use of L EXP , the value of which is directly proportional to the number of EDs. According to [28], the reception probability in LoRa networks is dependent on the number of GWs. Thus, in case of a high number of GWs, a shorter exploration phase is needed, which is achieved by L EXP being inversely proportional to the number of GWs. As seen in Figure 6  ( I I I ) , after L E X P rounds, at the end of the exploration phase, the value of N a k u ( a k u A ) is equal to L EXP . The weights and probabilities at the end of this phase for all the actions, will form the input of the exploitation phase.

4.2. Exploitation Phase of the LP-MAB Algorithm

In this phase, the actions are selected based on the relevant Probability Density Function (PDF), i.e., their probability, P a k u at the end of the exploration phase (line 2 of Algorithm 3). According to this step, which is derived from the EXP3 scheme, it is more likely that actions of high probability will be selected, i.e., more suitable actions have a higher chance to be selected. Same as in the exploration phase, also in this phase, both in the case of a successful ACK reception and in the case of a non-reception, in addition to the N a k u increment (line 3 of Algorithm 3), the NS will update the weight and probability of the relevant action (lines 6–8 of Algorithm 3).
Algorithm 3: Exploitation Phase of LP-MAB.
Sensors 23 02363 i003
Let us consider the case in which the probability of the kth action for the uth ED, i.e., of a k u , is smaller than the half of the maximum probability when all actions in A are considered. In that case, the NS will set the probability of action a k u to zero, so that action a k u will not be selected until the end of the exploitation phase (lines 9–11 of Algorithm 3). This removal process aims to eliminate actions with a low probability of leading to a successful transmission. It should be noted that the above threshold (line 9 of Algorithm 3) is derived heuristically from our simulation results, leading to the best performance.
The exploitation phase for the uth ED continues until the number of selections of at least one of the actions ( N a q u , for that action a q u ) reaches the value of L EE (line 1 of Algorithm 3). Thus, L EE should be considerably larger than L EXP for the first few transmission periods t considered, as a k u A , N a k u = L EXP at the beginning of the exploitation phase.
Our work differs from [16], in that we incorporate what we have learned from the environment during the exploration phase, through the use of the dynamic value of L EE . According to line 13 of Algorithm 1, L EE is calculated by multiplying the total number of actions by the quotient of the division of the exploration phase duration by the remaining time of the simulation. In general, the higher the number of actions is, the higher the number of successful transmissions, i.e., of potentially optimal configurations, will be, and, therefore, the NS will need to consider more actions to select the one that reduces interference the most; thus, a longer period of time should be spent on exploitation in this case. Moreover, due to the fact that our learning about the network increases as we get closer to the end of the simulation, to utilize the information obtained during the exploration phase more effectively, it is reasonable that the exploitation phase should take longer to finish.
As seen in Figure 7, for any arbitrary a q u , when N a q u reaches the value of L EE , the exploitation phase will be ended. At the end of this phase, the transmission period index t is incremented by one, so that the exploration phase can start again for a new transmission period. In this way, the actions that were removed from the previous execution will have a second chance. Note that all actions will have their N a k u set to zero, with no change in their weights or probabilities (line 14 of Algorithm 1). By not resetting the weight and probability values of the actions at the end of the exploitation phase, the previously gained knowledge is not eliminated by the proposed LP-MAB algorithm. Nevertheless, the weights and probabilities of all actions at the end of the new exploration phase, which correspond to the new transmission period, will be inputs for the exploitation phase corresponding to that transmission period.

5. Simulation Setup

We have used FLoRa [11] (a Framework for LoRa simulations) as a simulator tool. FLoRa, which is based on OMNeT++ [29], a discrete event network simulator, was proposed for the simulation of a LoRaWAN composed of EDs, GWs, and an NS according to the setup presented in [11]. More information regarding FLoRa is available at https://flora.aalto.fi/ (accessed on 11 January 2023). A LoRa link behavior model that considers the capture effect and inter-SF collisions in multiple network settings is presented in FLoRa. We customized FLoRa for simulating our adaptive configuration scheme based on artificial intelligence methods in the LoRa network. The LP-MAB framework is available at the following GitHub repository: https://github.com/reza-serati/LP-MAB (accessed on 11 January 2023).
We consider a LoRaWAN consisting of arbitrary numbers of GWs randomly placed in a square-shaped cell having different radius sizes based on urban and sub-urban environments, with up to 700 EDs uniformly distributed [11]. Simulations were conducted under the impact of the capture effect and inter-SF collisions to minimize the number of collisions [13]. The list of parameters that affect the performance of LoRaWAN are summarized in Table 1. It should be noted that, for the final results, a series of simulations was performed twenty times, and the resulting data were averaged.
To evaluate the proposed algorithm’s performance and compare it with other schemes, we use the following two metrics:
  • PDR (%): Defined as the total number of packets received by the NS divided by the total packets sent from all EDs during the simulation time.
  • EC (kJ): Defined as the total EC divided by PDR as discussed in [15].
We consider the following eight scenarios:
  • Scenario 1: The number of static EDs varies from 100 to 700, with a step size of 100.
  • Scenario 2: Based on the environmental conditions, for 100 static EDs, the channel saturation varies according to the values shown in Table 1.
  • Scenario 3: Considering that EDs are mobile, the number of nodes varies between 100 and 700.
  • Scenario 4: For 100 mobile EDs, the mobility speed can be changed using the values shown in Table 1.
  • Scenario 5: Comparatively to Scenarios 3 and 4, in which all nodes were mobile, two types of EDs were considered in the simulation environment: static and mobile, for varying network sizes.
  • Scenario 6: In contrast to the constant 12-day simulation time assumed in other Scenarios, in this Scenario, simulation days vary from 12 to 120, with a step size of 12 days, for 100 static EDs.
  • Scenario 7: Using the values shown in Table 1, the number of packets sent daily by each ED varies in a 365-day simulation time.
  • Scenario 8: Unlike traditional ADR approaches, in this Scenario, we have studied both the impact of CR and CF on LoRa network performance as well as the impact of SF and TP.
Additionally, all simulation scenarios are evaluated in urban and sub-urban environments, which operate differently in terms of path loss, channel saturation, and simulation radius parameters as demonstrated in Table 2. The parameter values selected for the two environments being simulated are such that make our work directly comparable to other works on the relevant scientific field, e.g., [11,15,24,30].

6. Simulation Results

Through simulations, we compare our proposed algorithm with the ADR-MAX [4], ADR-AVG [11], No-ADR (“No-ADR” indicates the absence of ADR; ADR is disabled and is not being used in this scheme.), and ADR-Lite [15] schemes in the eight aforementioned scenarios.

6.1. Scenario 1: Performance under a Varying Number of Static EDs

Figure 8a shows the PDR and EC in ADR-MAX, ADR-AVG, No-ADR, ADR-Lite, and LP-MAB in Scenario 1, for an urban environment with a radius of 480 m and σ equal to 3.56 . As observed, the PDR of our proposed solution is higher than others due to the use of a combination of a short-term initial exploration phase and a long exploitation phase, which follows the exploration phase. Additionally to that, in many RL techniques, the initial action probability is defined as a uniform distribution, i.e., P a k u ( t = 0 ) = 1 | A | , a k u A , u U [13]. Uniform probability initialization in such solutions as LoRa-MAB can take a long time to eliminate wrong choices from the actions, leading to increased convergence time. However, in LP-MAB, we do not assume an equal probability initialization for each action of an ED. Instead, we set the initial probabilities as an undefined number, i.e., n a n , which can be changed based on the ACK reception as demonstrated in the initial state shown in Figure 4.
Keeping a long system history from the start until the present enables the proposed method to maintain a more comprehensive understanding of the network and achieve a better performance than other ADR mechanisms, which have only a history of the last twenty packets received. We also see in Figure 8a that, in low network densification ( N < 300 ), the LP-MAB’s EC is lower than other approaches due to the fact that instead of making a decision based on only a portion of the previously received packets, we made decisions based on the entire history of received packets. In this way, from the first received packet to the last one, the NS tries to find the most optimal action to improve the network’s performance.
In Figure 8b, we illustrate the PDR and EC in ADR-MAX, ADR-AVG, No-ADR, ADR-Lite, and LP-MAB in Scenario 1, for a sub-urban environment with a radius of 9800 m. It can be seen that, due to the greedy manner of decision-making in ADR-Lite, it is possible to achieve a higher level of PDR, regardless of the network density, at the cost of a higher EC. Also, in the No-ADR scheme, because of its randomness and its lack of consideration for environmental changes, the result is entirely dependent on the initial transmission parameter values. Therefore, the No-ADR scheme achieves a better performance in the sub-urban environment compared to the urban environment. As a result of applying machine learning techniques such as RL, the NS can converge to the optimal state in terms of TP, resulting in the lowest EC of LP-MAB compared to other approaches. Another pertinent observation from Figure 8 is that the reduction in PDR and EC performance associated with increasing the number of the EDs is negligible, making the LP-MAB approach more scalable than others.

6.2. Scenario 2: Performance under Varying Values of Channel Saturation

In this Scenario, the channel noise, i.e., sigma ( σ ), takes the values of { 0 , 0.89 , 1.78 , 2.67 , 3.56 } and { 0 , 0.89 , 1.78 , 2.67 , 3.56 , 4.46 , 5.36 , 6.24 , 7.08 } for the urban and the sub-urban environment, respectively. In Figure 9, we illustrate the PDR and EC of different algorithms versus σ for the 100 static EDs used in Scenario 2. In the LP-MAB scheme, the reception of the NS’s ACKs by the EDs can directly influence the network’s performance. Thus, a higher rate of successful reception of the NS’s feedback by the EDs can contribute to a higher likelihood of determining the most optimal action.
Due to the lack of noise for σ = 0 , the EDs can receive most of the feedback, which can result in almost 100 % PDR and the lowest EC for LP-MAB compared to other schemes, in both urban and sub-urban environments. Increasing channel noise decreases the probability of successful feedback reception, so sub-optimal actions are selected for the EDs, resulting in reduced PDR and an increase in EC. In contrast to LP-MAB, when σ increases, the ADR-MAX’s EC also increases significantly, especially in noisy channels, because, as σ increases, the ADR-MAX’s PDR decreases, thus causing the EDs to choose less optimal actions, resulting in an increase in the EC.

6.3. Scenario 3: Performance under a Varying Number of Mobile EDs

A wide variety of applications require or apply mobility enabled by the IoT. Mobile applications are found in traffic monitoring, smart metering, and animal tracking [31,32]. Through this Scenario, we are investigating the effects of mobility on the performance of EDs, by comparing various ADR mechanisms. In this work, we assess the use of the Random Waypoint Mobility Model for simulating LoRaWAN [32]. This Scenario runs for σ = 7.08 and the number of EDs varies between 100 and 700 in both the urban and the sub-urban environment, while the EDs’ speed varies from zero to five meters per second and follows an exponential distribution.
Figure 10 shows the PDR and EC in ADR-MAX, ADR-AVG, No-ADR, ADR-Lite, and LP-MAB in Scenario 3. In LP-MAB, unlike ADR-Lite, the PDR, in both the urban and the sub-urban environments, does not degrade as the network densification increases because of the LP-MAB’s scalability feature. Thus, our proposed algorithm outperforms other ADR mechanisms in terms of PDR as the number of EDs increases. Compared to all other methods, our proposed algorithm achieved the lowest EC in the urban environment for low ED densification, and in the sub-urban environment for any number of EDs. We can attribute this to the multi-reward technique we have in place in our scheme, which means that the actions with the highest TP receive the lowest reward.

6.4. Scenario 4: Performance under Varying Values of Speed for Mobile EDs

For different IoT applications that require mobility, depending on the use case, the EDs may have varying speeds. For instance, in smart bicycles and animal monitoring applications, the EDs’ speed can be greater than 20 or lower than 5 km per hour, respectively. We examined the impact of different mobility speeds in a mobile Scenario, and the ways in which ADR mechanisms could be used to overcome the potential impact of mobile IoT devices. In this Scenario, EDs have a relatively low speed, between zero and twelve meters per second, with the speed varying in small steps of 2 m per second, i.e., of 7.2 km per hour, which is typical of IoT deployments in real-world environments. The varying speeds do not affect network performance in both the urban and the sub-urban environments, as can be seen in Figure 11.

6.5. Scenario 5: Performance under Varying Network Sizes and Different Mobility Speeds

Figure 12 shows the PDR and EC of the LP-MAB scheme for a variety of network sizes (small and large network areas) using 100 EDs, for different mobility speeds, based on the Random Waypoint Mobility Model [32]. As can be seen in Figure 12, by increasing the network size, the overall performance of the network will be degraded, regardless of its speed, the same as indicated in Scenario 4. It is important to note that the configured path loss model in our work is L o R a L o g N o r m a l S h a d o w i n , which is appropriate for small area networks as [11], unlike the L o R a P a t h L o s s O u l u path loss model used in [32], which is usually used in large area networks.

6.6. Scenario 6: Performance under a Varying Number of Simulation Days

In Figure 13, we illustrate the PDR and EC of different algorithms versus the number of simulation days for the 100 static EDs used in Scenario 6. According to the results, ADR-AVG performs better than other ADR mechanisms in terms of PDR in urban environments, as well as in terms of EC in both urban and sub-urban environments. It should be noted that these performance results were achieved in a low-density deployment of EDs. This may be incongruous with most IoT applications, requiring several hundred EDs, for which, as demonstrated in Scenarios 1 and 3, the performance of ADR-AVG may not be satisfactory.
Additionally, LP-MAB’s results are consistently second-best in this Scenario in terms of both PDR and EC, and may outperform other algorithms, if more EDs are included in this simulation scenario. Thus, we note that there seems to exist a trade-off between the ED densification and the number of days being simulated.

6.7. Scenario 7: Performance under a Varying Number of Packets Sent per Day

Figure 14 shows the PDR and EC in ADR-MAX, ADR-AVG, No-ADR, ADR-Lite, and LP-MAB in Scenario 7 for 100 static EDs in a 365-day simulation time. As discussed in Scenario 6, in an urban environment with low densification, ADR-AVG can outperform other algorithms. By lowering the average number of daily sent packets per ED, we can observe a throughput degradation of those algorithms whose performance directly depends on the reception of the feedback from the NS, i.e., ADR-AVG and LP-MAB, which are making decisions based on the last 20 received packets and the history of all the last received packets, respectively. Among the examined schemes, ADR-Lite as a low-complexity scheme that decides the following action based only on the last received packet, and No-ADR as a basic randomly deciding algorithm that does not apply any specific decision-making approach, exhibit a performance that remains unchanged by the reduction of the daily sent packets in the network.

6.8. Scenario 8: Performance under a Varying Number of Total Actions Available

Similar to the novel Scenario examined in our other work [15], here, we are also evaluating the impact of increasing the state space of the transmission parameters over a 120-day simulation time. In this way, we provide more freedom of choice in configuring the transmission parameters while applying no changes to the protocol design and adding no overhead to the LoRa packet’s header. Based on the parameter values shown in Table 1, the EDs can choose the SF, TP, CF, and CR using the following values: { 7 , 8 , 9 , 10 , 11 , 12 } , { 2 , 5 , 8 , 11 , 14 } , { 868.1 , 868.4 , 868.7 } , and { 4 5 , 4 6 , 4 7 , 4 8 } , respectively. In addition, it is important to note that as indicated in [33], BW cannot easily be altered due to the regularity limitations. Therefore, we have only considered the effects of SF, TP, CF, and CR with a cardinality of 6, 5, 3, and 4, respectively. For this Scenario, same as in [15], four different configurations were examined, namely Config-1, Config-2, Config-3, and Config-4, where the transmission parameters are: { S F + T P } , { S F + T P + C F } , { S F + T P + C R } , and { S F + T P + C F + C R } , respectively. Although, in real environments, the CF may not be adjustable for each ED, it can be modified during FLoRa simulations.
Figure 15 shows that, contrary to the initial assumptions about the higher degree of freedom in the choice of transmission parameters, the possibility of increasing PDR in both urban and sub-urban environments is rather limited. Config-1 uses the SF and TP as transmission parameters, which is the default configuration parameter set for ADR mechanisms, resulting in the same result as in Scenario 1. In general, increasing the number of CFs can result in less collision probability in each frequency, since inter-SF collisions, which are an important factor affecting the network’s performance, will be less likely. Therefore Config-2, which allows EDs to select different CFs for packet transmission, performs better than any other configuration in terms of both PDR and EC.
According to Equation (8), by increasing the CR, the physical message length will also increase, which will result in a longer T o A duration and, consequently, an increase in the chance of collision occurrence and a higher EC. Thus, in Config-3, in which the parameters’ selection state space has increased by allowing for the use of higher CR values for packet transmission, the overall network performance will be reduced. Albeit using multiple channels, i.e., a higher SF, can improve network performance even when a higher CR, i.e., a more effective error correction code, is selected. Nevertheless, also in this case, the total overhead of the network will also grow higher, resulting in unsatisfactory performance, as seen in Config-4’s results.
As illustrated in Figure 15 the PDR degradation in Config-1, Config-3, and Config-4, as well as the PDR growth in Config-2, in both the urban and the sub-urban environments, is more consistent in the LP-MAB approach in comparison to the ADR-Lite scheme due to LP-MAB’s scalability feature as discussed in Scenario 1. In general, however, our results validate the results presented in [15] regarding the ADR-Lite scheme’s performance in the four different configurations examined, as well as the general performance of ADR schemes in the context of these four configurations.

7. Conclusions and Future Works

This article introduces a centralized adaptive configuration algorithm to improve the PDR and EC in the context of LoRaWan, as these two metrics constitute the main performance metrics for LoRa networks. For this goal, we have presented an RL-based ADR algorithm that allows the NS to configure the EDs’ transmission parameters. This algorithm, which we have named LP-MAB, can achieve an efficient adaptive configuration using two MAB algorithms, SE and EXP3, after mapping the LoRa resource allocation problem to the MAB problem. By using SE and EXP3, the proposed solution can simultaneously benefit from the advantages of a short-term initial exploration phase and of a long exploitation phase, which follows the exploration phase.
Using several scenarios, we have evaluated the performance of the LP-MAB and compared it with other ADR mechanisms, namely the ADR-MAX [4], ADR-AVG [11], No-ADR, and ADR-Lite [15], in different circumstances. The simulation results indicate that the LP-MAB’s EC outperforms other algorithms while maintaining a relatively high PDR in various circumstances. LP-MAB is also more scalable than other approaches, since its PDR decreases relatively slowly as the number of the EDs increases. In IoT applications, where the battery lifetime is the most important factor, LP-MAB makes LoRa one of the best candidates for adoption as the main communication protocol among the vast number of EDs deployed.
As LoRa utilizes the ALOHA protocol as its Media Access Control (MAC) mechanism, the consequent dynamic value for the number of potential re-transmissions may lead to an increased PDR. Therefore, as a future work, we propose the use of an RL-based algorithm to specify the number of potential re-transmissions, which should be able to take into account the overall environmental conditions of the LoRa network, including the number of the EDs, the relevant noise, and the network size, on the one hand, as well as the trade-off between the number of potential packet re-transmissions and the EC, on the other hand.

Author Contributions

Conceptualization, B.T. and M.R.; Data curation, B.T. and R.S.; Funding acquisition, N.A.A.; Investigation, B.T. and R.S.; Methodology, B.T. and N.A.A.; Project administration, M.R.; Resources, N.A.A.; Software, B.T. and R.S.; Supervision, N.A.A. and M.R.; Validation, R.S. and N.A.A.; Writing—original draft, B.T., R.S., N.A.A. and M.R.; Writing—review & editing, N.A.A. and M.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the German Research Foundation–Deutsche Forschungsgemeinschaft (DFG), as part of Project 439892735 of the Priority Program 2253, and by the University of Oulu and the Academy of Finland Profi6 336449. The APC was funded by the Open Access Publication Fund of the University Library Passau.

Data Availability Statement

The FLoRa-based framework for simulating LP-MAB is available at https://github.com/reza-serati/LP-MAB (accessed on 11 January 2023).

Acknowledgments

We acknowledge support by the Open Access Publication Fund of the University Library of the University of Passau.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
3Gthird-generation cellular network
4Gfourth-generation cellular network
ACKACKnowledgment
ADRAdaptive Data Rate
BWBandWidth
CFCarrier Frequency
CRCoding Rate
CRCCyclic Redundancy Check
CSMA/CACarrier-Sense Multiple Access with Collision Avoidance
CSSChirp Spread Spectrum
ECEnergy Consumption
EDEnd Device
EXP3EXPonential weights for EXPloration and EXPloitation
FLoRaFramework for LoRa simulations
GWGateWay
IoTInternet of Things
kmkilometers
LoRaLong Range
LoRaWANLong-Range Wide-Area Network
LP-MABLow-Power Multi-Armed Bandit
LPWANLow-Power Wide-Area Network
mmeters
MABMulti-Armed Bandit
MACMedia Access Control
MLMachine Learning
NB-IoTNarrowBand IoT
NSNetwork Server
OWAOrdered Weighted Averaging
PDFProbability Density Function
PDRPacket Delivery Ratio
RLReinforcement Learning
SESuccessive Elimination
SINRSignal-to-Interference-plus-Noise Ratio
SFSpreading Factor
ToATime-on-Air
TPTransmission Power

References

  1. Nižetić, S.; Šolić, P.; López-de-Ipiña González-de Artaza, D.; Patrono, L. Internet of Things (IoT): Opportunities, issues and challenges towards a smart and sustainable future. J. Clean. Prod. 2020, 274, 122877. [Google Scholar] [CrossRef]
  2. Centenaro, M.; Vangelista, L.; Zanella, A.; Zorzi, M. Long-range communications in unlicensed bands: The rising stars in the IoT and smart city scenarios. IEEE Wirel. Commun. 2016, 23, 60–67. [Google Scholar] [CrossRef] [Green Version]
  3. Patel, D.; Won, M. Experimental Study on Low Power Wide Area Networks (LPWAN) for Mobile Internet of Things. In Proceedings of the 2017 IEEE 85th Vehicular Technology Conference (VTC Spring), Sydney, Australia, 4–7 June 2017; pp. 1–5. [Google Scholar]
  4. LoRa Alliance. LoRaWAN Specification (Version 1.1), 2017. Technical Specification. Available online: https://lora-alliance.org/resource_hub/lorawan-specification-v1-1/ (accessed on 11 January 2023).
  5. Hassan, M.B.; Ali, E.S.; Mokhtar, R.A.; Saeed, R.A.; Chaudhari, B.S. NB-IoT: Concepts, applications, and deployment challenges. In LPWAN Technologies for IoT and M2M Applications; Chaudhari, B.S., Zennaro, M., Eds.; Chapter 6; Academic Press: Cambridge, MA, USA, 2020; pp. 119–144. [Google Scholar]
  6. Ratasuk, R.; Vejlgaard, B.; Mangalvedhe, N.; Ghosh, A. NB-IoT system for M2M communication. In Proceedings of the 2016 IEEE Wireless Communications and Networking Conference, Doha, Qatar, 3–6 April 2016; pp. 1–5. [Google Scholar]
  7. UnaBiz SAS. Sigfox. Available online: https://www.sigfox.com/ (accessed on 11 January 2023).
  8. Lauridsen, M.; Kovács, I.Z.; Mogensen, P.; Sorensen, M.; Holst, S. Coverage and capacity analysis of LTE-M and NB-IoT in a rural area. In Proceedings of the 2016 IEEE 84th Vehicular Technology Conference (VTC-Fall), Montreal, QC, Canada, 18–21 September 2016; pp. 1–5. [Google Scholar]
  9. Semtech Corporation, Wireless Sensing and Timing Products Division. AN1200.22: LoRa Modulation Basics, May 2015. Rev. 2. Available online: https://www.frugalprototype.com/wp-content/uploads/2016/08/an1200.22.pdf (accessed on 11 January 2023).
  10. Haidine, A.; Hassani, S.E.; Aqqal, A.; Hannani, A.E. The Role of Communication Technologies in Building Future Smart Cities. In Smart Cities Technologies; Silva, I.N.D., Flauzino, R.A., Eds.; Chapter 4; IntechOpen: Rijeka, Croatia, 2016; pp. 55–76. [Google Scholar]
  11. Slabicki, M.; Premsankar, G.; Di Francesco, M. Adaptive configuration of LoRa networks for dense IoT deployments. In Proceedings of the NOMS 2018—2018 IEEE/IFIP Network Operations and Management Symposium, Taipei, Taiwan, 23–27 April 2018; pp. 1–9. [Google Scholar]
  12. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction, 2nd ed.; Adaptive Computation and Machine Learning Series; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  13. Ta, D.T.; Khawam, K.; Lahoud, S.; Adjih, C.; Martin, S. LoRa-MAB: A Flexible Simulator for Decentralized Learning Resource Allocation in IoT Networks. In Proceedings of the 2019 12th IFIP Wireless and Mobile Networking Conference (WMNC), Paris, France, 11–13 September 2019; pp. 55–62. [Google Scholar]
  14. Ta, D.T.; Khawam, K.; Lahoud, S.; Adjih, C.; Martin, S. LoRa-MAB: Toward an Intelligent Resource Allocation Approach for LoRaWAN. In Proceedings of the 2019 IEEE Global Communications Conference (GLOBECOM), Big Island, HI, USA, 9–13 December 2019; pp. 1–6. [Google Scholar]
  15. Serati, R.; Teymuri, B.; Anagnostopoulos, N.A.; Rasti, M. ADR-Lite: A Low-Complexity Adaptive Data Rate Scheme for the LoRa Network. In Proceedings of the 2022 18th International Conference on Wireless and Mobile Computing, Networking and Communications (WiMob 2022), Thessaloniki, Greece, 10–12 October 2022; pp. 296–301. [Google Scholar]
  16. Azizi, F.; Teymuri, B.; Aslani, R.; Rasti, M.; Tolvanen, J.; Nardelli, P.H.J. MIX-MAB: Reinforcement Learning-based Resource Allocation Algorithm for LoRaWAN. In Proceedings of the 2022 IEEE 95th Vehicular Technology Conference: (VTC2022-Spring), Helsinki, Finland, 19–22 June 2022; pp. 1–6. [Google Scholar]
  17. Augustin, A.; Yi, J.; Clausen, T.; Townsley, W.M. A Study of LoRa: Long Range & Low Power Networks for the Internet of Things. Sensors 2016, 16, 1466. [Google Scholar] [PubMed]
  18. Bouguera, T.; Diouris, J.F.; Chaillout, J.J.; Jaouadi, R.; Andrieux, G. Energy Consumption Model for Sensor Nodes Based on LoRa and LoRaWAN. Sensors 2018, 18, 2104. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Casals, L.; Mir, B.; Vidal, R.; Gomez, C. Modeling the Energy Performance of LoRaWAN. Sensors 2017, 17, 2364. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  20. LoRa Alliance. LoRaWAN Specification (Version 1.0.2), 2016. Technical Specification. Available online: https://lora-alliance.org/resource_hub/lorawan-specification-v1-0-2/ (accessed on 11 January 2023).
  21. Li, S.; Raza, U.; Khan, A. How agile is the adaptive data rate mechanism of LoRaWAN? In Proceedings of the 2018 IEEE Global Communications Conference (GLOBECOM), Abu Dhabi, United Arab Emirates, 9–13 December 2018; pp. 206–212. [Google Scholar]
  22. Sørensen, R.B.; Kim, D.M.; Nielsen, J.J.; Popovski, P. Analysis of latency and MAC-layer performance for class A LoRaWAN. IEEE Wirel. Commun. Lett. 2017, 6, 566–569. [Google Scholar] [CrossRef] [Green Version]
  23. Bor, M.C.; Roedig, U.; Voigt, T.; Alonso, J.M. Do LoRa Low-Power Wide-Area Networks Scale? In Proceedings of the 19th ACM International Conference on Modeling, Analysis and Simulation of Wireless and Mobile Systems (MSWiM 2016), Malta, 13–17 November 2016; pp. 59–67. [Google Scholar]
  24. Babaki, J.; Rasti, M.; Aslani, R. Dynamic Spreading Factor and Power Allocation of LoRa Networks for Dense IoT Deployments. In Proceedings of the 2020 IEEE 31st Annual International Symposium on Personal, Indoor and Mobile Radio Communications, London, UK, 31 August–3 September 2020; pp. 1–6. [Google Scholar]
  25. Ahmed, Z.E.; Saeed, R.A.; Mukherjee, A.; Ghorpade, S.N. Energy optimization in low-power wide area networks by using heuristic techniques. In LPWAN Technologies for IoT and M2M Applications; Chaudhari, B.S., Zennaro, M., Eds.; Academic Press: Cambridge, MA, USA, 2020; Chapter 10; pp. 199–223. [Google Scholar]
  26. Aihara, N.; Adachi, K.; Takyu, O.; Ohta, M.; Fujii, T. Q-Learning Aided Resource Allocation and Environment Recognition in LoRaWAN With CSMA/CA. IEEE Access 2019, 7, 152126–152137. [Google Scholar] [CrossRef]
  27. Allesiardo, R.; Féraud, R.; Maillard, O.A. The non-stationary stochastic multi-armed bandit problem. Int. J. Data Sci. Anal. 2017, 3, 267–283. [Google Scholar] [CrossRef] [Green Version]
  28. Hoeller, A.; Souza, R.D.; López, O.L.A.; Alves, H.; de Noronha Neto, M.; Brante, G. Analysis and performance optimization of LoRa networks with time and antenna diversity. IEEE Access 2018, 6, 32820–32829. [Google Scholar] [CrossRef]
  29. Varga, A. OMNeT++. In Modeling and Tools for Network Simulation; Wehrle, K., Günes, M., Gross, J., Eds.; Chapter 3; Springer Science & Business Media: Heidelberg, Germany, 2010; pp. 35–59. [Google Scholar]
  30. Al-Gumaei, Y.A.; Aslam, N.; Aljaidi, M.; Al-Saman, A.; Alsarhan, A.; Ashyap, A.Y. A Novel Approach to Improve the Adaptive-Data-Rate Scheme for IoT LoRaWAN. Electronics 2022, 11, 3521. [Google Scholar] [CrossRef]
  31. de Castro Tomé, M.; Nardelli, P.H.; Alves, H. Long-Range Low-Power Wireless Networks and Sampling Strategies in Electricity Metering. IEEE Trans. Ind. Electron. 2019, 66, 1629–1637. [Google Scholar] [CrossRef] [Green Version]
  32. Al Mojamed, M. On the Use of LoRaWAN for Mobile Internet of Things: The Impact of Mobility. Appl. Syst. Innov. 2022, 5, 5. [Google Scholar] [CrossRef]
  33. Talla, V.; Hessar, M.; Kellogg, B.; Najafi, A.; Smith, J.R.; Gollakota, S. LoRa Backscatter: Enabling The Vision of Ubiquitous Connectivity. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2017, 1, 1–24. [Google Scholar] [CrossRef]
Figure 1. Range of wireless protocols, according to [10] and our own knowledge and experience.
Figure 1. Range of wireless protocols, according to [10] and our own knowledge and experience.
Sensors 23 02363 g001
Figure 2. LoRaWAN network architecture.
Figure 2. LoRaWAN network architecture.
Sensors 23 02363 g002
Figure 3. The assumed working mode sequence for each ED, adopted from [18].
Figure 3. The assumed working mode sequence for each ED, adopted from [18].
Sensors 23 02363 g003
Figure 4. The initialization of LP-MAB for the uth ED.
Figure 4. The initialization of LP-MAB for the uth ED.
Sensors 23 02363 g004
Figure 5. Possible first round of the LP-MAB exploration phase for the uth ED.
Figure 5. Possible first round of the LP-MAB exploration phase for the uth ED.
Sensors 23 02363 g005
Figure 6. Possible next rounds of the LP-MAB exploration phase for the uth ED.
Figure 6. Possible next rounds of the LP-MAB exploration phase for the uth ED.
Sensors 23 02363 g006
Figure 7. Possible LP-MAB exploitation phase for the uth ED, with ( V ) representing the final round of the exploitation phase for this transmission period of the uth ED. In this extreme case used as an example, action a q u has been selected to be performed in all rounds.
Figure 7. Possible LP-MAB exploitation phase for the uth ED, with ( V ) representing the final round of the exploitation phase for this transmission period of the uth ED. In this extreme case used as an example, action a q u has been selected to be performed in all rounds.
Sensors 23 02363 g007
Figure 8. PDR & EC versus different numbers of static EDs in Scenario 1.
Figure 8. PDR & EC versus different numbers of static EDs in Scenario 1.
Sensors 23 02363 g008
Figure 9. PDR & EC versus different values of channel saturation in Scenario 2.
Figure 9. PDR & EC versus different values of channel saturation in Scenario 2.
Sensors 23 02363 g009
Figure 10. PDR & EC versus different numbers of mobile EDs in Scenario 3.
Figure 10. PDR & EC versus different numbers of mobile EDs in Scenario 3.
Sensors 23 02363 g010
Figure 11. PDR & EC versus different values for speed for mobile EDs in Scenario 4.
Figure 11. PDR & EC versus different values for speed for mobile EDs in Scenario 4.
Sensors 23 02363 g011
Figure 12. PDR & EC of the LP-MAB scheme versus varying network sizes and different mobility speeds in Scenario 5.
Figure 12. PDR & EC of the LP-MAB scheme versus varying network sizes and different mobility speeds in Scenario 5.
Sensors 23 02363 g012
Figure 13. PDR & EC versus different numbers of simulation days in Scenario 6.
Figure 13. PDR & EC versus different numbers of simulation days in Scenario 6.
Sensors 23 02363 g013
Figure 14. PDR & EC versus different values for number of sending message per day in Scenario 7.
Figure 14. PDR & EC versus different values for number of sending message per day in Scenario 7.
Sensors 23 02363 g014
Figure 15. PDR & EC versus different values for number of total actions in Scenario 8.
Figure 15. PDR & EC versus different values for number of total actions in Scenario 8.
Sensors 23 02363 g015
Table 1. Simulation Setup Parameters.
Table 1. Simulation Setup Parameters.
ParameterValue
Simulator PlatformOMNet++
Simulator ModelINET and FLORA
Repetitions20
Mobility ModelRandom Waypoint
ED Speed (v) { 0 , , 12 } m/s
Urban Cell Radius (r)480 m
Sub-urban Cell Radius (r)9800 m
Packet Length (L)20 bytes
BW ( B W )125 kHz
Simulations Time (T)12 days
Number of EDs (N) { 100 , , 700 }
Number of GWs ( # G W ) { 1 , , 10 }
Urban Environment’s Sigma ( σ ) { 0.0 , , 3.56 }
Sub-urban Environment’s Sigma ( σ ) { 0.0 , , 7.08 }
SFs ( SF ) { 7 , 8 , 9 , 10 , 11 , 12 }
TPs ( TP ) { 2 , 5 , 8 , 11 , 14 } dBm
CFs ( CF ) { 868.1 , 868.4 , 868.7 } MHz
CRs ( CR ) { 4 5 , 4 6 , 4 7 , 4 8 }
Number of Sent Packets per Day ( ϵ ) { 1 , , 192 } packets/day
Table 2. Standard deviation of the path loss ( σ ) in dB and other parameters for the different deployment scenarios. This table is partially adapted from [30] and based on the relevant values provided in [23,30].
Table 2. Standard deviation of the path loss ( σ ) in dB and other parameters for the different deployment scenarios. This table is partially adapted from [30] and based on the relevant values provided in [23,30].
Scenarios d 0 [m] P L ¯ ( d 0 ) [dB]n σ [dB]Cell Radius (r) [m]
Urban40 127.41 2.08 3.57 480
Sub-urban1000 128.95 2.32 7.08 9800
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Teymuri, B.; Serati, R.; Anagnostopoulos, N.A.; Rasti, M. LP-MAB: Improving the Energy Efficiency of LoRaWAN Using a Reinforcement-Learning-Based Adaptive Configuration Algorithm. Sensors 2023, 23, 2363. https://doi.org/10.3390/s23042363

AMA Style

Teymuri B, Serati R, Anagnostopoulos NA, Rasti M. LP-MAB: Improving the Energy Efficiency of LoRaWAN Using a Reinforcement-Learning-Based Adaptive Configuration Algorithm. Sensors. 2023; 23(4):2363. https://doi.org/10.3390/s23042363

Chicago/Turabian Style

Teymuri, Benyamin, Reza Serati, Nikolaos Athanasios Anagnostopoulos, and Mehdi Rasti. 2023. "LP-MAB: Improving the Energy Efficiency of LoRaWAN Using a Reinforcement-Learning-Based Adaptive Configuration Algorithm" Sensors 23, no. 4: 2363. https://doi.org/10.3390/s23042363

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop