论文标题

plad:学习使用伪标签和近似分布来推断形状程序

PLAD: Learning to Infer Shape Programs with Pseudo-Labels and Approximate Distributions

论文作者

Jones, R. Kenny, Walke, Homer, Ritchie, Daniel

论文摘要

推断产生2D和3D形状的程序对于反向工程,编辑等非常重要。执行此任务的培训模型很复杂,因为配对(形状,程序)数据不容易用于许多域,因此确切的监督学习不可行。但是,可以通过损害分配的程序标签或形状分布的准确性来获取配对数据。尾流方法使用来自形状程序的生成模型的样品来近似真实形状的分布。在自我训练中,形状通过识别模型,该模型预测了这些形状被视为“伪标签”的程序。与这些方法相关,我们引入了一种针对程序推理独特的新型自我训练变体,其中程序伪标记与其执行的输出形状配对,以近似形状分布为代价避免了标签不匹配。我们建议在单个概念框架下对这些制度进行分组,在该概念框架下进行训练,并以伪标签或近似分布(PLAD)得出的最大似然更新(PLAD)进行训练。我们在多个2D和3D形状程序推理域上评估了这些技术。与政策梯度增强学习相比,我们表明PLAD技术可以推断出更准确的形状程序,并更快地收敛。最后,我们建议将来自不同PLAD方法的更新结合在单个模型的培训中,并发现此方法的表现优于任何个人技术。

Inferring programs which generate 2D and 3D shapes is important for reverse engineering, editing, and more. Training models to perform this task is complicated because paired (shape, program) data is not readily available for many domains, making exact supervised learning infeasible. However, it is possible to get paired data by compromising the accuracy of either the assigned program labels or the shape distribution. Wake-sleep methods use samples from a generative model of shape programs to approximate the distribution of real shapes. In self-training, shapes are passed through a recognition model, which predicts programs that are treated as "pseudo-labels" for those shapes. Related to these approaches, we introduce a novel self-training variant unique to program inference, where program pseudo-labels are paired with their executed output shapes, avoiding label mismatch at the cost of an approximate shape distribution. We propose to group these regimes under a single conceptual framework, where training is performed with maximum likelihood updates sourced from either Pseudo-Labels or an Approximate Distribution (PLAD). We evaluate these techniques on multiple 2D and 3D shape program inference domains. Compared with policy gradient reinforcement learning, we show that PLAD techniques infer more accurate shape programs and converge significantly faster. Finally, we propose to combine updates from different PLAD methods within the training of a single model, and find that this approach outperforms any individual technique.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源