论文标题
自适应子载波,参数和电力分配,用于宽带渠道的分区边缘学习
Adaptive Subcarrier, Parameter, and Power Allocation for Partitioned Edge Learning Over Broadband Channels
论文作者
论文摘要
在本文中,我们考虑了分区的边缘学习(Partel),该学习在无线网络中实现了一种众所周知的分布式学习方法。因此,Partel利用边缘设备分布的计算资源来训练大型人工智能(AI)模型,通过将模型动态分配到参数块中以进行分离的设备更新。针对宽带通道,我们考虑参数分配,子通道分配和传输功率的关节控制以提高Partel的性能。具体而言,在最小学习潜伏期的标准下,优化了联合子载波,参数和功率分配(支持)的策略。考虑了两种情况。首先,对于可分解模型(例如逻辑回归)的情况,延迟最小化问题是一个混合构成程序和非convex。由于其难以理解,我们通过整数松弛并将其转换为等效凸大问题的延迟约束下的等效凸问题来开发实用的解决方案。因此,低复杂算法旨在计算支持策略。其次,考虑可以通过引入一些辅助变量来使用Partel训练的深神经网络(DNN)模型的情况。但是,这引入了对模型分区的限制,从而降低了参数分配的粒度。前面的策略扩展到了DNN模型,通过将所提出的负载舍入和比例调整的技术应用于载荷粒度约束引起的潜伏期扩张。
In this paper, we consider partitioned edge learning (PARTEL), which implements parameter-server training, a well known distributed learning method, in a wireless network. Thereby, PARTEL leverages distributed computation resources at edge devices to train a large-scale artificial intelligence (AI) model by dynamically partitioning the model into parametric blocks for separated updating at devices. Targeting broadband channels, we consider the joint control of parameter allocation, sub-channel allocation, and transmission power to improve the performance of PARTEL. Specifically, the policies for joint SUbcarrier, Parameter, and POweR allocaTion (SUPPORT) are optimized under the criterion of minimum learning latency. Two cases are considered. First, for the case of decomposable models (e.g., logistic regression), the latency-minimization problem is a mixed-integer program and non-convex. Due to its intractability, we develop a practical solution by integer relaxation and transforming it into an equivalent convex problem of model size maximization under a latency constraint. Thereby, a low-complexity algorithm is designed to compute the SUPPORT policy. Second, consider the case of deep neural network (DNN) models which can be trained using PARTEL by introducing some auxiliary variables. This, however, introduces constraints on model partitioning reducing the granularity of parameter allocation. The preceding policy is extended to DNN models by applying the proposed techniques of load rounding and proportional adjustment to rein in latency expansion caused by the load granularity constraints.