Network Controll Through Symbiosis

February 23, 2025 • Gleb Razgar and Hiroki Kozuki

1. Background

Brain emulations promise to make way for an interesting future. Unfortunately current methods render emulations un-interpretable. Say you wanted to find which circuits of the fly’s brain are responsible for pleasure, and then modify them to see the overall effect on the simulation. With todays interpretability methods this would be hardly possible.

Learning to read and control networks - human or otherwise is a logical goal we should strive to accomplish. But how do we?

Hitherto, interpreting and steering AIs has mainly been done through two key lenses: Mechanistic and Representational. Both deal with different units of analysis. Representation engineering (RepE) is a top-down study of the network effects exhibited by the populations of neurons, whilst Mechanistic Interpretability (MI) is a bottom-up study attempting to reverse engineer those effects from individual neurons and circuits. MI is precise but impractical, RepE is practical but imprecise. If we were to make it practical and precise to control AI models, both of the techniques are insufficient. To solve this, we propose network Symbiosis as a possible solution.

Alt text

Figure 1: SymbioteNet (left) steering the primary neural network (right) by interacting with it's weights and observing it's outputs


2. Network Symbiosis

2.1 Intuition

To understand network Symbiosis, consider brain-computer interfaces. Putting neuroplasticity aside, BCIs learn your brain representations that map to a desired output. The idea behind Symbiotic networks is similar. There are always at least two networks involved (Figure 1):

  1. Primary feed-forward network - acting like the brain
  2. Symbiote feed-forward network - acting like the BCI

The Primary network is trained to map input to output as per usual, whilst the SymbioteNet uses the Primary model’s activations or weights as its input and tries to manipulate them to get the Primary network to behave in the desired fashion.

2.2 Example

A simple example would illustrate the concept. Like in Figure 1, say we’re training a classic MNIST CNN to classify digits 1-10. Once we’ve trained the Primary CNN, we would then train a SymbioteNet by feeding in a digit we would like the Primary network to misclassify and backpropagate the error depending on how well it managed to inhibit the digit (by inhibiting all the CNN’s neurons apart from the ones in the first and last layers). An important detail is that the cross-entropy loss should punish the model for misclassifying ALL digits and instead be rewarded if it classifies them all correctly except for the undesired digit.


3. Interpreting Networks

SymbioteNet’s role is to elucidate the low-level representations of the Primary network. This implies being able to capture mechanistically universal concepts among networks first highlighted by Chris Olah and his team:

Claim 1: Features Features are the fundamental unit of information of a network. Can SymbioteNet reveal the features of the studied model?

Claim 2: Circuits Features are connected by weights, forming circuits. Can SymbioteNet reveal circuits?

Claim 3: Universality Analogous features and circuits form across models and tasks. Can SymbioteNet reveal those circuits recursively?

These claims have been shown to be true across the networks, even though by causation and with low degrees of success. SymbioteNet tries solving exactly that.


4. Experiment

4.1 Basics

Before scaling to larger networks like transformers, we conducted our experiment on the setup described in Section 2. Having trained and frozen the weights of the Primary MNIST CNN, we taught the SymbioteNet to inhibit each digit iteratively without misclassifying the unstated digit until its turn.

Alt text

Figure 2: Training and validation loss of SymbioteNet. As can be deduced from the validation loss, it takes the model significantly longer to learn the weight representations than for the CNN to learn classification.

4.2 Polysemanticity

Smaller networks have the propensity to learn concepts easily, so, motivated by the BIMT paper, we induce polysemanticity by progressively pruning the weaker weights away during training of the Primary network (Figure 3). This forces the network to become more parameter efficient and forces concepts to superimpose in a mixed representation.

Alt text

Figure 3: Enforcement of polysemanticity through BIMT-like weight pruning.

4.3 Results

After training both networks and freezing their weights, we test them on the unseen dataset by checking how many times SymbioteNet manages to successfully steer the Primary network away from correctly classifying the digit 3.

Importantly, SymbioteNet is made two times smaller than the Primary network, such that it is forced to learn the compressed representation of the subsidiary model’s weights rather than simply memorizing the digit activations. That’s where the advantage of training SymbioteNet on the CNN’s weights rather than activations really shines.

Alt text

Figure 4: Confusion matrix showing SymbioteNet's success in learning the circuits of the Primary network and inhibiting them to correctly misclassify the digit 3 on command.

As seen in the confusion matrix, after training, SymbioteNet managed to prevent the Primary model from classifying 100% of the digits 3 from an equal data distribution. Combined with the loss observations, and equally spreading the model’s capacity to inhibit the weights across layers with the exclusion of the first and last, the only reasonable deduction is that SymbioteNet has learned the intrinsic representations of both features and circuits.


5. Future Work

This brief study has just scratched the surface of symbiotic networks. The next steps in the inquiry are rooted in scaling this approach to much larger models, such as transformers, as though our mimicry of polysemanticity is representative of larger networks, transformers encode information in a different dimension of complexity. If successfully scaled to transformers, it’s evident that network symbiosis would make for a new way to interpret and steer models.