Language Model Crossover: Variation through Few-Shot Prompting
Document Type
Article
Publication Date
11-28-2024
Abstract
This article pursues the insight that language models naturally enable an intelligent variation operator similar in spirit to evolutionary crossover. In particular, language models of sufficient scale demonstrate in-context learning, i.e., they can learn from associations between a small number of input patterns to generate outputs incorporating such associations (also called few-shot prompting). This ability can be leveraged to form a simple but powerful variation operator, i.e., to prompt a language model with a few text-based genotypes (such as code, plain-text sentences, or equations), and to parse its corresponding output as those genotypes' offspring. The promise of such language model crossover (which is simple to implement and can leverage many different open source language models) is that it enables a simple mechanism to evolve semantically rich text representations (with few domain-specific tweaks), and naturally benefits from current progress in language models. Experiments in this article highlight the versatility of language-model crossover, through evolving binary bit-strings, sentences, equations, text-to-image prompts, and Python code. The conclusion is that language model crossover is a flexible and effective method for evolving genomes representable as text.
Identifier
85210697343 (Scopus)
Publication Title
ACM Transactions on Evolutionary Learning and Optimization
External Full Text Location
https://doi.org/10.1145/3694791
e-ISSN
26883007
ISSN
2688299X
Issue
4
Volume
4
Grant
1948017
Fund Ref
National Science Foundation
Recommended Citation
Meyerson, Elliot; Nelson, Mark J.; Bradley, Herbie; Gaier, Adam; Moradi, Arash; Hoover, Amy K.; and Lehman, Joel, "Language Model Crossover: Variation through Few-Shot Prompting" (2024). Faculty Publications. 80.
https://digitalcommons.njit.edu/fac_pubs/80