Dataset Viewer
Auto-converted to Parquet Duplicate
Search is not available for this dataset
text
int64
0
14.9M
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
11,726,849
End of preview. Expand in Data Studio

Llama1B 1M Training Activations

This repository contains activation data accompanying the paper Learning a Generative Meta-Model of LLM Activations.

Quick Start

With this data, you can train a GLP on Llama-3.2-1B activations from Layer 07. The activations are derived from FineWeb. GLPs are activation diffusion models useful for applications like on-manifold steering and sparse probing.

# download data
huggingface-cli download generative-latent-prior/llama1b-layer07-fineweb-1M \
    --repo-type dataset  \
    --local-dir data/llama1b-layer07-fineweb-1M \
    --local-dir-use-symlinks False
# launch training
conda activate glp
python3 glp_train.py config=configs/train_llama1b_static.yaml

Citation

@article{luo2026glp,
  title={Learning a Generative Meta-Model of LLM Activations},
  author={Grace Luo and Jiahai Feng and Trevor Darrell and Alec Radford and Jacob Steinhardt},
  journal={arXiv preprint arXiv:2602.06964},
  year={2026}
}
Downloads last month
21

Paper for generative-latent-prior/llama1b-layer07-fineweb-1M