🔇Paper Reading: Code Pretraining in LLMs
Jun 10, 2024
| Oct 10, 2024
Words 145Read Time 1 min
type
status
date
Jun 10, 2024 06:38 AM
slug
summary
tags
category
icon
password
Paper Information If LLM Is the Wizard, Then Code Is the Wand: A Survey on How Code Empowers Large Language Models to Serve as Intelligent Agents https://arxiv.org/abs/2401.00812
This surveys discusses multiple benefits of including code in LLM pretraining from three aspects:
  • improving programming and reasoning skills
  • using external tools with executable steps
  • performing self-improvement from feedback

Functionalities of Code Pretraining

  1. Performance gain: programming, reasoning, structure generation
  1. Tool Use: generate executable steps during decision-making (Planning, CoT)
  1. Interaction (Agent): self-improvement
notion image
Furthermore, there are several benefits in building intelligent agents:
  1. enhance perception and planning skills
  1. direct action primitive grounding and modular memory organization
  1. providing self-correction and self-improvement

Code improves performance

notion image
Advantages of learning from code:
  1. Directly improve programming skills
  1. Empower complex reasoning (CoT → PoT)
  1. Capture structured knowledge
    1. structural reasoning
    2. markup language understanding

Code improves tool using

notion image
Type of tools

Code improves self-correction

notion image
How to learn from feedback?
  1. Selection based method
    1. majority voting
    2. re-ranking
  1. Prompted-based method
  1. Fine-tuning

Challenges

  1. Causality between code pretraining and LLM’s performance improvement
  1. Enhancing reasoning beyond code
  1. Improving multi-turn interactions
 
  • LLM
  • Paper
  • Survey
  • Alibaba Qwen: Generalizing an LLM from 8k to 1M Context using Qwen-AgentLong Context Challenge in LLMs
    Loading...