# LLCs of Compiled Neural Networks

Studying the Local Learning Coefficient in neural networks compiled from known programs.

**Type**: Applied

**Difficulty**: Medium

**Status**: Unstarted

This project aims to study the Local Learning Coefficient (LLC) in neural networks that are compiled from known programs. By using tools like Tracr or other techniques for compiling programs into neural networks, we can create a controlled environment to study the relationship between algorithmic structure and the LLC.

Key research questions:

- How does the LLC of a compiled neural network correlate with known measures of program complexity?
- How does the LLC vary with model size for compiled networks implementing the same algorithm?
- Can we use LLC analysis to detect the presence of specific algorithmic structures in neural networks?
- How does the complexity vary between compiled networks and the same networks trained on the relevant task?

Methodology:

- Implement a system for compiling a variety of algorithms into neural networks of different sizes.
- Estimate LLCs for these compiled networks.
- Compare LLC estimates to established measures of program complexity (e.g., Kolmogorov complexity & time/space complexity when available).
- Investigate how LLCs change as we vary the size of the compiled network while keeping the underlying algorithm constant.
- Attempt to use LLC analysis to detect the presence of specific algorithmic structures in more general neural networks.

Expected outcomes:

- Empirical data on the relationship between LLCs and algorithmic complexity.
- Insights into how LLCs reflect the underlying structure of compiled neural networks.
- Potential development of LLC-based techniques for detecting algorithmic structure in neural networks.

This research could provide a bridge between the abstract notion of the LLC and more concrete measures of algorithmic complexity, potentially enhancing our ability to interpret the internal structure of neural networks.

## Where to begin:

- Quantifying degeneracy (Lau et al. 2023) ,
- Tracr: Compiled Transformers as a Laboratory for Interpretability

If you have decided to start working on this, please let us know in the Discord. We'll update this listing so that other people who are interested in this project can find you.