AI: post transformers

CODEGEN: Open Language Model for Code Synthesis


Listen Later

This source introduces CODEGEN, a family of large language models developed by Salesforce Research, designed for program synthesis. The models, varying in size up to 16.1B parameters, are trained on extensive natural language and programming language datasets, and the training library, JAXFORMER, is open-sourced to promote accessibility. A key contribution is the exploration of multi-turn program synthesis, where complex problems are broken into smaller, interactable steps, enhancing user intent understanding and synthesis accuracy. To evaluate this, the authors created Multi-Turn Programming Benchmark (MTPB), demonstrating that multi-turn prompting significantly improves performance over single-turn specifications, particularly for more challenging problems. The research highlights the scalability of program synthesis capacity with increasing model and data size, making powerful code generation more attainable for wider research and application.


Source: 2023 - https://arxiv.org/pdf/2203.13474

...more
View all episodesView all episodes
Download on the App Store

AI: post transformersBy mcgrof