论文标题

大型模型的演变

Evolution through Large Models

论文作者

Lehman, Joel, Gordon, Jonathan, Jain, Shawn, Ndousse, Kamal, Yeh, Cathy, Stanley, Kenneth O.

论文摘要

本文追求了这样的见解:大型语言模型(LLMS)训练以生成代码可以极大地提高突变操作员在基因编程(GP)中应用程序的有效性。因为这样的LLM受益于包括顺序更改和修改的培训数据,因此它们可以近似人类会做出的变化。为了强调通过大型模型(ELM)在主要实验ELM中与MAP-ELITE相结合的含义的广度,生成了数十万个Python程序的功能示例,这些程序在Sodarace域中在Sodarace域中输出卧床机器人,而原始LLM在预训练中从未见过。这些示例然后有助于引导培训一种新的条件语言模型,该模型可以为特定地形输出合适的步行者。引导新模型可以在以前可用的零培训数据中为给定上下文中输出适当的工件的新模型具有对开放性,深度学习和增强学习的影响。在这里深入探讨了这些含义,以期激发榆树现在打开的新研究方向。

This paper pursues the insight that large language models (LLMs) trained to generate code can vastly improve the effectiveness of mutation operators applied to programs in genetic programming (GP). Because such LLMs benefit from training data that includes sequential changes and modifications, they can approximate likely changes that humans would make. To highlight the breadth of implications of such evolution through large models (ELM), in the main experiment ELM combined with MAP-Elites generates hundreds of thousands of functional examples of Python programs that output working ambulating robots in the Sodarace domain, which the original LLM had never seen in pre-training. These examples then help to bootstrap training a new conditional language model that can output the right walker for a particular terrain. The ability to bootstrap new models that can output appropriate artifacts for a given context in a domain where zero training data was previously available carries implications for open-endedness, deep learning, and reinforcement learning. These implications are explored here in depth in the hope of inspiring new directions of research now opened up by ELM.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源