LLM Pretraining with Continuous Concepts

Meta Info

URL: https://arxiv.org/abs/2502.08524

First author: Jihoon Tack

Affiliation: FAIR at Meta, KAIST, UC San Diego

Code

https://github.com/facebookresearch/RAM/tree/main/projects/cocomix

Overview Framwork

Understand the paper

Background

  • natural language tokens are often superficial (e.g., function words like “the” or “a”), and need a more efficient way for high-level reasoning and conceptual understanding

  • LLMs inherently encode high-level concepts and reasoning processes in their latent representations

  • Sparse Autoencoders (SAEs) can effectively isolate meaningful latent features in LLMs by capturing the high-level semantic concepts

Key Question

Can we augment the next token prediction objective to explicitly model concepts in a latent representation space, thereby bridging semantic abstraction and fine-grained token-level guidance?

Objection

  • Enrich or supplement natural language tokens so that LLMs can learn more abstract reasoning abilities and long-range dependencies.

Method

  • We extract semantic concepts using a pretrained SAE and select the most influential ones based on attribution scores, which quantify each concept’s influence on the model’s output.

  • The model is then trained to predict these selected concepts from its hidden state using a cross-entropy loss.

  • Once multiple concepts are predicted, we compress them into a single continuous concept and mix (or insert) them into the hidden states by interleaving with token embeddings, thereby directly contributing to the next token prediction

Comment

  • While Coconut (Training Large Language Models to Reason in a Continuous Latent Space) directly feeds the hidden layers as ”reasoning thought”, CocoMix (this work) chooses the most significant ”concepts” from the hidden layers.

  • Still, CocoMix does not emphasize the reasoning chain, but on the extra condition for generating the next token.

  • Next step: refine the reasoning chain?

Last updated