The average number of free preambles NI is given by equation [2.3]:
From [2.1] and [2.2], we obtain the expected number of failed preambles NF:
[2.4]
The modeled system is an approximation of reality in many ways, especially where it concerns the fixed and limited number of access attempts. However, we preferred to simplify the model to make it more tractable (Figure 2.5).
Figure 2.5. System model. For a color version of this figure, see www.iste.co.uk/chalouf/intelligent.zip
The sub-system 1 represents the terminals
The proposed model is fluid: the quantities involved and the whole numbers are considered to be real (continuous) quantities. The parameters used are listed below:
– x1(t) is the number of waiting devices;
– x1,L(t) is the number of blocked devices, after a failure at the end of an access attempt (i.e. ACB);
– x2(t) is the total number of devices that pass the ACB control and wait to start the random access attempt (RA);
– x2,L(t) is the number of blocked devices, at instant t, after a failed RA attempt, awaiting a new attempt;
– λ is the arrival rate of IoT objects;
– μ is the rate of objects that can reattempt ACB after a failure;
– θ1 is the RA failure rate, which is equal to when θ is equal to 0 (see the point before last);
– θ2 is the rate of objects that can attempt access after a failure;
– θ is the rate at which the device abandons transmission after having reached the maximum number of RA attempts; in a correctly dimensioned system, we should have θ = 0;
– p is the ACB blocking factor.
We are now able to describe the evolution of state variables
[2.5]
In what follows, we suppose that θ = 0, to simplify the model. In fact, a system where the devices often reach the maximum number of attempts is an unstable system, which we naturally try to avoid.
The model described in [2.5] is nonlinear and not affine, in the control. It can easily be demonstrated that the model described is not observable, in view of its state [x1 x1,L x2 x2,L] which cannot be known precisely. It is also not controllable, as the blocking factor p can only act partially on the state. These properties render the synthesis of an optimal controller, guaranteeing the stability of the very complex system, described above, very complex.
Although the state is not observable, it is possible to produce an estimation of the average number of devices attempting access
2.5. Access controller for IoT terminals based on reinforcement learning
The difficulty of observing the system state, described in section 2.4, has led us to consider strategies making it possible to deduce the blocking factor even in the presence of very noisy measurements.
It is in this sense that we relied on deep learning techniques, which demonstrated great effectiveness in automatically extracting characteristics of system “features” in the presence of data tainted with noise or even of incomplete data (Rolnick et al. 2017).
Given the lack of data, we have considered the class of reinforcement learning techniques.
More particularly, we considered the “Twin Delayed Deep Deterministic policy gradient algorithm” (TD3) technique, which can tackle a continuous action space, and which has shown greater effectiveness in learning speed and in performance than existing approaches (Fujimoto et al. 2018).
We formulate, in what follows, the problem of access in the IoT as a reinforcement learning problem, in which an agent finds iteratively a sub-optimal blocking factor, making it possible to reduce the access conflict.
2.5.1. Formulating the problem
In reinforcement learning (Sutton et al. 2019), we have two main entities, an environment and an agent. The learning process happens through interaction between these entities so that the agent can optimize a total revenue. At each stage t, the agent obtains a representation of the state St of the environment and chooses an action at, based on this. Then the agent applies this action to the environment. Consequently, the environment passes to a new state St+1 and the agent receives a recompense rt. This interaction can be modeled as a Markovian decision process (MDP) M = (S, A, P, R), where S is the state space, A the action space, P the dynamic transition and R the revenue function. The behavior of the agent is defined by its policy π: S ⟶ A, which makes it possible to link a state to an action where a deterministic system is involved, or a distribution of action where it is probabilistic. The objective of such a system is to find the optimal policy π*, making it possible to maximize the accumulated revenue.
In the problem of controlling access to the IoT, we define a discrete MDP, where the state, the action and the revenue are defined as follows:
– The state: given that the number of terminals attempting access at a given instant k is unavailable, the state we are considering is based on measured estimates. Since a single measurement of this number is necessarily very noisy, we will consider a set of several measurements, which can better reveal the state present in the network. The state sk