Page 89 - ITU Journal Future and evolving technologies Volume 2 (2021), Issue 4 – AI and machine learning solutions in 5G and future networks
P. 89

ITU Journal on Future and Evolving Technologies, Volume 2 (2021), Issue 4




          From all the above-mentioned methods, XGBoost [32] was   features, each STA of every deployment is characterized by a
          selected  for  competing  in  the  challenge  because  it  of-  feature  vector  (   , …,   21+3     ),  with      denoting  the
                                                                                 
          fered  the  highest  performance  on  both  the  training  and   number  of  wireless  channels.  Note,  as  well,  that  the
          validation  stages  compared  with  the  other  models.  XG‑   entire data set is considered for training, thus combining
          boost is a gradient boosting framework available for mul‑   STAs  from  different  deployments.  The  rationale  is  that
          tiple platforms, thus providing high portability.    features such as the number of neighbors in the primary
                                                               channel,   the  SINR,  and  the  interference  should
          To train our model, we have analyzed the variance of the
                                                               differentiate STAs from different deployments.
          variables in the data set, and discarded the features with
          low or moderate variability.  Some of the considered fea‑   When it comes to the regression problem, Gossip uses a
          tures are the position of nodes, the primary channel, the   feed‑forward neural network with four layers.  The input
          distance  among  nodes,  or  the  power  received  by  neigh‑   layers pass the input features to two fully connected lay‑
          boring  devices.  To  preprocess  the  selected  features,  we   ers of neurons with a ReLU activation unit.  Finally, a sin‑
          have  applied  Yeo‑Johnson  transformation  [33]  and  nor‑   gle neuron receives the output of the hidden layers and
          malization.  These steps were done for predictors to im‑   generates the prediction of the throughput.  As a remark,
          prove their utility in the models.                   the last neuron has a linear activation.  It is important to
                                                               remark that the proposed neural network is mostly meant
          The  training  data  set  was  split  into  training  and  valida‑
                                                               to tackle linear regression problems. Nevertheless, even if
          tion, so that we could train the model on the whole train‑
                                                               the throughput prediction problem for WLANs is not lin‑
          ing  data  set  using  a  10‑fold  cross‑validation  procedure.
                                                               ear, we expect our model to properly identify local min‑
          Further, we used 100 and 300 combinations, respectively,
                                                               imum/maximum points that allow providing reasonable
          with 10‑fold cross‑validation.
                                                               prediction results.
          As  for  the  hyper‑parameter  setting  (e.g.,  max  depth  or
                                                               To  train  the  proposed  neural  network,  we  have  used
          minimum child weight), we tuned hyper‑parameters us‑
                                                               the RMSprop gradient descend method,  considering the
          ing a grid search.  More speci ically, the hyper‑parameter
                                                               Mean Squared Error (MSE) as a loss function.  Moreover,
          values  were  set  using  Latin  Hypercube  Sampling,  while
                                                               50 training episodes and a batch size of 50 STAs have been
          their  maximum  and  lower  range  value  for  each  hyper‑
                                                               considered. Thanks to Gossip design, the training data set
          parameter were mostly predetermined using default val‑
          ues from tidymodels [34].                            is populated with every STA of every deployment present
                                                               among all scenarios.
          Finally,  signi icant  efforts  have  been  put  to  deploy  our
          model.  In  particular,  we  have  used  docker  to  make  our
                                                               3.5  Net Intels
          model easy to (re)train and deploy.  All the code and doc‑
          umentation has been made publicly available [35].    To address the objective of predicting the throughput of
                                                               APs and STAs in typical dense environments, we explore
          3.4  UC3M NETCOM                                     a set of popular regression techniques.  With the help of
                                                               these techniques, we aim to build complex mathematical
          We formulate the throughput forecasting in WLANs as a   relationships among features and labels from the data set,
          linear  regression  problem,  which  assumes  a  linear  rela‑   so that performance of WLANs can be predicted at unseen
          tionship between the features and label(s) of a given data  deployments.  In particular, we propose using the follow‑
          set {   ,      ,      ,   ...,      ,    }, and a set of unknown parameters   ing techniques: 5
                   ,1
                
                        ,2
             to be learned (being    the bias).                 1. Arti icial Neural Network (ANN): The ANN method
                              0
          Our solution (named Gossip) is based on a linear regres‑   is  selected    ly  due  to  its  potential  and  versatil‑
          sion method, and it aims to predict the throughput of an   ity to model nonlinear and complex relationships in
          STA    in    a    given    WLAN    deployment    where  CB  is  ap‑   OBSS data elegantly. The proposed ANN is built using
          plied [36]. Based on STAs’ individual throughput, we de‑   Tensor low and Keras libraries in Python [38].  The
          rive  the  performance  of  each  AP  by  aggregating  the   NN model is designed with one input layer, 7 hidden
          values  of  their  associated  STAs.  In  particular,  Gossip   layers, and 1 output layer (see Fig. 5). The ReLU func‑
          derives the unknown bias and weight parameters    by (i)   tion is employed to activate hidden layers. In each of
          processing the WLAN data set, and (ii) applying a neural   the  irst six hidden layers, there are 1024 nodes. For
          network to perform regression.                           the seventh hidden layer, there are 512 nodes.  The
                                                                   model is trained using an Adam optimizer. The batch
          As  for  the  processing  part,  Gossip  takes  the  input  fea‑
                                                                   size and number of epochs for training, after multiple
          tures  generated  by  the  Komondor  simulator,  and  se‑
                                                                   trials, were set to 250 and 1000 respectively.
          lects/generates  the  most  relevant  ones:  the  position  of
          the STA, the AP to which the STA is associated,  the RSSI,   5 The code used to implement all the proposed methods by Net Intels is
          the SINR, the set of nodes using the same primary channel,   available in Github [37].
          and  the  set  of  allowed  channels.  After  processing  the






                                             © International Telecommunication Union, 2021                    73
   84   85   86   87   88   89   90   91   92   93   94