Telecommunications Science ›› 2022, Vol. 38 ›› Issue (6): 82-90.doi: 10.11959/j.issn.1000-0801.2022152

• Research and Development • Previous Articles     Next Articles

Value-difference learning based mMTC devices access algorithm in multi-cell network

Xin LI1, Jun SUN1,2   

  1. 1 College of Telecommunications &Information Engineering, Nanjing University of Posts and Telecommunications, Nanjing 210003, China
    2 Jiangsu Key Laboratory of Wireless Communications, Nanjing 210003, China
  • Revised:2022-04-06 Online:2022-06-20 Published:2022-06-01
  • Supported by:
    The National Natural Science Foundation of China(61771255);Provincial and Ministerial Key Laboratory Open Project(20190904)

Abstract:

In the massive machine type communication scenario of 5G, the access congestion problem of massive machine type communication devices (mMTCD) in multi-cell network is very important.A double deep Q network with value-difference based exploration (VDBE-DDQN) algorithm was proposed.The algorithm focused on the solution that could reduce the collision when a number of mMTCDs accessed to eNB in multi-cell network.The state transition process of the deep reinforcement learning algorithm was modeled as Markov decision process.Furthermore, the algorithm used a double deep Q network to fit the target state-action value function, and it employed an exploration strategy based on value-difference to adapt the change of the environment, which could take advantage of both current conditions and expected future needs.Moreover, each mMTCD updated the probability of exploration according to the difference between the current value function and the next value function estimated by the network, rather than using the same standard to select the best base eNB for the mMTCD.Simulation results show that the proposed algorithm can effectively improve the access success rate of the system.

Key words: mMTC, RA, reinforcement learning, eNB selection

CLC Number: 

No Suggested Reading articles found!