通信学报 ›› 2022, Vol. 43 ›› Issue (6): 16-27.doi: 10.11959/j.issn.1000-436x.2022128

• 专题:面向6G的智能至简网络关键技术 • 上一篇    下一篇

6G联邦边缘学习新范式:基于任务导向的资源管理策略

王志勤1, 江甲沫1, 刘沛西2, 曹晓雯3, 李阳1, 韩凯峰1, 杜滢1, 朱光旭4   

  1. 1 中国信息通信研究院移动通信创新中心,北京 100191
    2 北京大学电子学院,北京 100871
    3 广东工业大学信息工程学院,广东 广州 510006
    4 深圳市大数据研究院,广东 深圳 518172
  • 修回日期:2022-04-27 出版日期:2022-06-01 发布日期:2022-06-01
  • 作者简介:王志勤(1970- ),女,北京人,博士,中国信息通信研究院教授级高级工程师,主要研究方向为无线移动通信技术和标准
    江甲沫(1985- ),男,吉林长春人,博士,中国信息通信研究院高级工程师,主要研究方向为面向6G的无线人工智能、通信感知一体化技术
    刘沛西(1991- ),男,湖北仙桃人,北京大学博士生,主要研究方向为面向6G的无线人工智能技术
    曹晓雯(1995- ),女,广东河源人,广东工业大学博士生,主要研究方向为面向6G的无线人工智能、无线空中计算技术
    李阳(1993- ),男,河南濮阳人,中国信息通信研究院助理工程师,主要研究方向为新一代移动通信与人工智能结合技术
    韩凯峰(1993- ),男,北京人,博士,中国信息通信研究院高级工程师,主要研究方向为面向6G的无线人工智能、通信感知一体化技术
    杜滢(1978- ),女,山东菏泽人,中国信息通信研究院教授级高级工程师,主要研究方向为无线移动通信技术和标准
    朱光旭(1989- ),男,广东广州人,博士,深圳市大数据研究院研究员,主要研究方向为面向6G的无线人工智能技术
  • 基金资助:
    国家重点研发计划基金资助项目(2020YFB1807100);国家自然科学基金资助项目(62001310);广东省基础与应用基础研究基金资助项目(2022A1515010109)

New design paradigm for federated edge learning towards 6G:task-oriented resource management strategies

Zhiqin WANG1, Jiamo JIANG1, Peixi LIU2, Xiaowen CAO3, Yang LI1, Kaifeng HAN1, Ying DU1, Guangxu ZHU4   

  1. 1 Mobile Communication Innovation Center, China Academy of Information and Communications Technology, Beijing 100191, China
    2 School of Electronics, Peking University, Beijing 100871, China
    3 School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China
    4 Shenzhen Research Institute of Big Data, Shenzhen 518172, China
  • Revised:2022-04-27 Online:2022-06-01 Published:2022-06-01
  • Supported by:
    The National Key Research and Development Program of China(2020YFB1807100);The National Natural Science Foundation of China(62001310);Guangdong Basic and Applied Basic Research Foundation(2022A1515010109)

摘要:

目的:为充分利用分布在网络边缘的丰富数据使之服务于人工智能模型训练,以联邦边缘学习为代表的边缘智能技术应运而生,本文综述了面向 6G 的联邦边缘学习技术,能够充分利用分布在网络边缘的丰富数据使之服务于人工智能模型训练,并以最优化学习性能(如优化模型训练时间、学习收敛性等)为目标设计无线资源管理策略。

方法:利用联邦学习网络架构,本文对资源分配和用户调度方案进行分析:1.对于资源分配问题分析在以最小化总训练时间为目标时的通信轮数和每轮时延之间的折中关系,为了满足每轮训练的时间约束,应该给计算能力低的设备分配更多的频率带宽,通过减小通信时间补偿计算时间,反之亦然。因此,设备间的带宽分配应该同时考虑信道条件和计算资源,这与传统只考虑信道条件的带宽分配工作截然不同。为此,需要建模总训练时间最小化问题,优化量化级数和带宽分配,设计交替优化算法解决该问题。2.对于用户调度问题通过将数据重要性和仍需通信轮次相联系、信道质量和单轮通信延时相联系,利用理论模型将两者统一,建模通信时间最小化优化问题。通过求解该优化问题发现,最优的调度策略在前期会更多地注重数据重要性,而在后期更注重信道质量。此外,还将所提的单设备调度算法也扩展至多设备调度场景当中。

结果:1.对于资源分配问题,当带宽分配最优时,通过仿真得到的总训练时间与量化级数的关系,在每个量化级数上运行相同的训练过程至少5次,且存在使总训练时间最小化的最优量化级数。总训练时间 T= N ϵ T d ,且 N ϵ 是量化水平 q 的递减函数,而Td是 q 的递增函数。此外,通过理论优化中得到的最优量化级数与仿真结果一致,同时验证了所提算法的有效性。根据损失函数最优值间隔与训练时间的关系所示,通过求解训练时间最小化问题得到了最优量化级数和最优带宽分配策略。2.对于用户调度问题,仿真中将所提用户调度(TLM)方案与其他3个常见的调度方案比较,并分别显示了在通信时间为6000s和14000s时的平均精度,其中平均精度通过测量预测值与真实值的联合交叉(IoU, intersection of union)得到。其中CA方案在信道衰减最大的1号车产生了最差的精度,而 IA 方案在数据不太重要的4号车展现了最低的精度。ICA方案的目的是在CA和IA之间找到一个平衡点,但由于其启发式的性质,性能低于TLM方案。

结论:1.最优量化级数和最优带宽分配下的训练损失在更短的时间内达到了预定的阈值,并且能取得最高的测试准确率。其次,非最优量化级数和最优带宽分配下的训练性能会优于最优量化级数和平均带宽分配的性能,这同时验证了资源分配的必要性。2.TLM方案在训练早期取得了稍好的性能,充分训练后明显优于所有其他方案。这是由于所提TLM方案中固有的前瞻性质比现有的CA、IA和ICA方案中的近视性质更有优势。

关键词: 6G智能网络, 任务导向的联邦边缘学习, 资源管理策略, 用户调度策略

Abstract:

Objectives: To make full use of the abundant data distributed at the edge of the network to serve the training of artificial intelligence models, edge intelligence technology represented by federated edge learning emerges as the times require. The rich data at the edge enables it to serve artificial intelligence model training, and design wireless resource management strategies with the goal of optimizing learning performance (such as optimizing model training time, learning convergence, etc.).

Methods: Using the federated learning network architecture, this paper analyzes the resource allocation and user scheduling schemes: 1. For the resource allocation problem, the trade-off relationship between the number of communication rounds and the delay per round when the goal of minimizing the total training time is analyzed, To meet the time constraints of each round of training, more frequency bandwidth should be allocated to devices with low computational power, compensating for a computational time by reducing communication time, and vice versa. Therefore, bandwidth allocation between devices should consider both channel conditions and computing resources, which is completely different from the traditional bandwidth allocation work that only considers channel conditions. To this end, it is necessary to model the total training time minimization problem, optimize the quantization series and bandwidth allocation, and design an alternate optimization algorithm to solve this problem. 2. For the user scheduling problem, the communication time minimization optimization problem is modeled by linking the importance of data with the number of communication rounds, channel quality and single-round communication delay, and using a theoretical model to unify the two. By solving the optimization problem, it is found that the optimal scheduling strategy will pay more attention to the importance of data in the early stage, and pay more attention to the channel quality in the later stage. In addition, the proposed single-device scheduling algorithm is also extended to multi-device scheduling scenarios.

Results: 1. For the resource allocation problem, when the bandwidth allocation is optimal, the relationship between the total training time and the quantization level obtained by simulation, run the same training process at least 5 times on each quantization level, and there is a total training time. The total training time is T= N ϵ T d , and N ϵ is a decreasing function of quantization level q,and Td is an increasing function of q . In addition, the optimal quantization series obtained through theoretical optimization is consistent with the simulation results, and the effectiveness of the proposed algorithm is verified. According to the relationship between the optimal value interval of the loss function and the training time, the optimal quantization series and the optimal bandwidth allocation strategy are obtained by solving the training time minimization problem. 2. For the user scheduling problem, the proposed user scheduling (TLM) scheme is compared with three other common scheduling schemes in the simulation, and the average precision is shown when the communication time is 6 000 s and 14 000 s, where the average Accuracy is obtained by measuring the IoU, the intersection of union, of the predicted value and the true value. The CA scheme yields the worst accuracy on car 1 with the largest channel attenuation, while the IA scheme exhibits the lowest accuracy on car 4 where the data is less important. The ICA scheme aims to find a balance between CA and IA, but due to its heuristic nature, the performance is lower than that of the TLM scheme.

Conclusions: 1. The training loss under the optimal quantization level and optimal bandwidth allocation reaches the predetermined threshold in a shorter time and can achieve the highest test accuracy. Secondly, the training performance under the non-optimal quantization level and optimal bandwidth allocation will be better than the performance of the optimal quantization leveland average bandwidth allocation, which also verifies the necessity of resource allocation. 2. The TLM scheme achieves slightly better performance early in training and significantly outperforms all other schemes after full training. This is due to the inherent prospective nature in the proposed TLM protocol which is advantageous over the myopic nature in the existing CA, IA and ICA protocols.

Key words: intelligent network towards 6G, task-oriented federated edge learning, resource management strategy, user scheduling strategy

中图分类号: 

No Suggested Reading articles found!