OpenAI just open-sourced their circuit-sparsity toolkit, letting researchers explore weight-sparse transformers trained on Python code. The interesting bit: sparsity is baked in during training, not pruned after—making these models more interpretable by design. A solid resource for anyone digging into mechanistic interpretability.
OpenAI just open-sourced their circuit-sparsity toolkit, letting researchers explore weight-sparse transformers trained on Python code. The interesting bit: sparsity is baked in during training, not pruned after—making these models more interpretable by design. 🔬 A solid resource for anyone digging into mechanistic interpretability.