当前位置: X-MOL 学术Phys. Commun. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Joint resource allocation and power control for D2D communication with deep reinforcement learning in MCC
Physical Communication ( IF 2.0 ) Pub Date : 2020-12-26 , DOI: 10.1016/j.phycom.2020.101262
Dan Wang , Hao Qin , Bin Song , Ke Xu , Xiaojiang Du , Mohsen Guizani

Mission-critical communication (MCC) is one of the main goals in 5G, which can leverage multiple device-to-device (D2D) connections to enhance reliability for mission-critical communication. In MCC, D2D users can reuses the non-orthogonal wireless resources of cellular users without a base station (BS). Meanwhile, the D2D users will generate co-channel interference to cellular users and hence affect their quality-of-service (QoS). To comprehensively improve the user experience, we proposed a novel approach, which embraces resource allocation and power control along with Deep Reinforcement Learning (DRL). In this paper, multiple procedures are carefully designed to assist in developing our proposal. As a starter, a scenario with multiple D2D pairs and cellular users in a cell will be modeled; followed by the analysis of issues pertaining to resource allocation and power control as well as the formulation of our optimization goal; and finally, a DRL method based on spectrum allocation strategy will be created, which can ensure D2D users to obtain the sufficient resource for their QoS improvement. With the resource data provided, which D2D users capture by interacting with surroundings, the DRL method can help the D2D users autonomously selecting an available channel and power to maximize system capacity and spectrum efficiency while minimizing interference to cellular users. Experimental results show that our learning method performs well to improve resource allocation and power control significantly.



中文翻译:

MCC中具有深度强化学习的D2D通信联合资源分配和功率控制

关键任务通信(MCC)是5G的主要目标之一,它可以利用多个设备到设备(D2D)连接来增强关键任务通信的可靠性。在MCC中,D2D用户可以在没有基站(BS)的情况下重用蜂窝用户的非正交无线资源。同时,D2D用户将对蜂窝用户产生同信道干扰,从而影响其服务质量(QoS)。为了全面改善用户体验,我们提出了一种新颖的方法,该方法包括资源分配和功率控制以及深度强化学习(DRL)。在本文中,精心设计了多种程序来协助制定我们的提案。首先,将对一个单元中具有多个D2D对和蜂窝用户的场景进行建模。然后分析与资源分配和功率控制有关的问题,以及制定我们的优化目标;最后,提出了一种基于频谱分配策略的DRL方法,可以保证D2D用户获得足够的资源来改善他们的QoS。借助提供的D2D用户通过与周围环境交互捕获的资源数据,DRL方法可以帮助D2D用户自主选择可用的信道和功率,从而最大程度地提高系统容量和频谱效率,同时将对蜂窝用户的干扰降至最低。实验结果表明,我们的学习方法在改善资源分配和功率控制方面表现良好。将创建一种基于频谱分配策略的DRL方法,该方法可以确保D2D用户获得足够的资源来改善其QoS。借助提供的D2D用户通过与周围环境交互捕获的资源数据,DRL方法可以帮助D2D用户自主选择可用的信道和功率,从而最大程度地提高系统容量和频谱效率,同时将对蜂窝用户的干扰降至最低。实验结果表明,我们的学习方法在改善资源分配和功率控制方面表现良好。将创建一种基于频谱分配策略的DRL方法,该方法可以确保D2D用户获得足够的资源来改善其QoS。借助提供的D2D用户通过与周围环境交互捕获的资源数据,DRL方法可以帮助D2D用户自主选择可用的信道和功率,从而最大程度地提高系统容量和频谱效率,同时将对蜂窝用户的干扰降至最低。实验结果表明,我们的学习方法在改善资源分配和功率控制方面表现良好。DRL方法可以帮助D2D用户自主选择可用的信道和功率,以最大程度地提高系统容量和频谱效率,同时最大程度地减少对蜂窝用户的干扰。实验结果表明,我们的学习方法在改善资源分配和功率控制方面表现良好。DRL方法可以帮助D2D用户自主选择可用的信道和功率,以最大程度地提高系统容量和频谱效率,同时最大程度地减少对蜂窝用户的干扰。实验结果表明,我们的学习方法在改善资源分配和功率控制方面表现良好。

更新日期:2021-01-02
down
wechat
bug