Menu
Sign In Search Podcasts Charts People & Topics Add Podcast API Pricing
Podcast Image

AI: post transformers

CODEGEN: Open Language Model for Code Synthesis

08 Aug 2025

Description

This source introduces CODEGEN, a family of large language models developed by Salesforce Research, designed for program synthesis. The models, varying in size up to 16.1B parameters, are trained on extensive natural language and programming language datasets, and the training library, JAXFORMER, is open-sourced to promote accessibility. A key contribution is the exploration of multi-turn program synthesis, where complex problems are broken into smaller, interactable steps, enhancing user intent understanding and synthesis accuracy. To evaluate this, the authors created Multi-Turn Programming Benchmark (MTPB), demonstrating that multi-turn prompting significantly improves performance over single-turn specifications, particularly for more challenging problems. The research highlights the scalability of program synthesis capacity with increasing model and data size, making powerful code generation more attainable for wider research and application.Source: 2023 - https://arxiv.org/pdf/2203.13474

Audio
Featured in this Episode

No persons identified in this episode.

Transcription

This episode hasn't been transcribed yet

Help us prioritize this episode for transcription by upvoting it.

0 upvotes
🗳️ Sign in to Upvote

Popular episodes get transcribed faster

Comments

There are no comments yet.

Please log in to write the first comment.