| transformer_lens/ActivationCache.py |
ActivationCache.__init__ |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.remove_batch_dim |
9 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.__getitem__ |
8 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.__len__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.to |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/ActivationCache.py |
ActivationCache.toggle_autodiff |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.keys |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.values |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.items |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.__iter__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.apply_slice_to_batch_dim |
7 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.accumulated_resid |
25 |
0 |
0 |
14 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.logit_attrs |
22 |
0 |
0 |
16 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.decompose_resid |
34 |
0 |
0 |
22 |
2 |
96% |
| transformer_lens/ActivationCache.py |
ActivationCache.compute_head_results |
14 |
4 |
0 |
10 |
1 |
62% |
| transformer_lens/ActivationCache.py |
ActivationCache.stack_head_results |
27 |
0 |
0 |
16 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.stack_activation |
6 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/ActivationCache.py |
ActivationCache.get_neuron_results |
12 |
0 |
0 |
8 |
2 |
90% |
| transformer_lens/ActivationCache.py |
ActivationCache.stack_neuron_results |
30 |
1 |
0 |
20 |
1 |
96% |
| transformer_lens/ActivationCache.py |
ActivationCache.apply_ln_to_stack |
33 |
1 |
0 |
24 |
5 |
89% |
| transformer_lens/ActivationCache.py |
ActivationCache.get_full_resid_decomposition |
32 |
0 |
0 |
16 |
2 |
96% |
| transformer_lens/ActivationCache.py |
(no function) |
34 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.__call__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.to_tokens |
11 |
0 |
0 |
4 |
1 |
93% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.forward |
18 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.run_with_cache |
10 |
0 |
0 |
2 |
1 |
92% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.run_with_cache.ForwardWrapper.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.run_with_cache.ForwardWrapper.__enter__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.run_with_cache.ForwardWrapper.__enter__.wrapped_forward |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/BertNextSentencePrediction.py |
BertNextSentencePrediction.run_with_cache.ForwardWrapper.__exit__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/BertNextSentencePrediction.py |
(no function) |
11 |
0 |
34 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
TensorLike.ndim |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
TensorLike.shape |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
TensorLike.size |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
TensorLike.unsqueeze |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
TensorLike.squeeze |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
TensorLike.broadcast_to |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
TensorLike.__matmul__ |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__init__ |
16 |
1 |
0 |
0 |
0 |
94% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__matmul__ |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__matmul__ |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__matmul__ |
8 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__rmatmul__ |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__rmatmul__ |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__rmatmul__ |
8 |
1 |
0 |
6 |
1 |
86% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__mul__ |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__rmul__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.AB |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.BA |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.T |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.svd |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.S |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.Vh |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.eigenvalues |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix._convert_to_slice |
6 |
0 |
0 |
4 |
1 |
90% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__getitem__ |
13 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.norm |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.__repr__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.make_even |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.get_corner |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.ndim |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.collapse_l |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.collapse_r |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.unsqueeze |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
FactoredMatrix.pair |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/FactoredMatrix.py |
(no function) |
43 |
0 |
47 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.__init__ |
13 |
5 |
0 |
8 |
3 |
52% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder._ensure_numpy |
12 |
4 |
0 |
10 |
2 |
64% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.to_frames |
40 |
15 |
0 |
18 |
6 |
57% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.encoder_output |
13 |
3 |
0 |
6 |
1 |
68% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.forward |
17 |
5 |
0 |
10 |
3 |
63% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.run_with_cache |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.to |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.cuda |
5 |
4 |
0 |
4 |
0 |
11% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.cpu |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.mps |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.from_pretrained |
23 |
3 |
0 |
8 |
4 |
77% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.W_K |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.W_Q |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.W_V |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.W_O |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.W_in |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.W_out |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.b_K |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.b_Q |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.b_V |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.b_O |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.b_in |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.b_out |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.QK |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.OV |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
HookedAudioEncoder.all_head_labels |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedAudioEncoder.py |
(no function) |
62 |
0 |
10 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder._get_blocks |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.__init__ |
30 |
3 |
0 |
16 |
5 |
83% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.to_tokens |
10 |
9 |
0 |
2 |
0 |
8% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.encoder_output |
11 |
10 |
0 |
6 |
0 |
6% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.forward |
20 |
19 |
0 |
8 |
0 |
4% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.run_with_cache |
5 |
4 |
0 |
2 |
0 |
14% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.to |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.cuda |
5 |
4 |
0 |
4 |
0 |
11% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.cpu |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.mps |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.from_pretrained |
14 |
13 |
0 |
6 |
0 |
5% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.b_U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_E |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_pos |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_E_pos |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_in |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.W_out |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.b_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.b_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.b_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.b_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.b_in |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.b_out |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.QK |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.OV |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
HookedEncoder.all_head_labels |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoder.py |
(no function) |
72 |
0 |
36 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.__init__ |
30 |
29 |
0 |
18 |
0 |
2% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.to_tokens |
8 |
7 |
0 |
2 |
0 |
10% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.forward |
38 |
37 |
0 |
24 |
0 |
2% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.generate |
51 |
50 |
0 |
26 |
0 |
1% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.run_with_cache |
5 |
4 |
0 |
2 |
0 |
14% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.to |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.cuda |
5 |
4 |
0 |
4 |
0 |
11% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.cpu |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.mps |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.from_pretrained |
16 |
15 |
0 |
8 |
0 |
4% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.b_U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_E |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_pos |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_in |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.W_out |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.b_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.b_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.b_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.b_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.b_in |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.b_out |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.QK |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.OV |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
HookedEncoderDecoder.all_head_labels |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedEncoderDecoder.py |
(no function) |
73 |
0 |
28 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.__init__ |
49 |
11 |
0 |
34 |
9 |
73% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.check_hooks_to_add |
8 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.get_pos_offset |
10 |
1 |
0 |
4 |
1 |
86% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.get_residual |
22 |
5 |
0 |
14 |
2 |
75% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.input_to_embed |
27 |
3 |
0 |
18 |
3 |
87% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.forward |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.forward |
32 |
4 |
0 |
22 |
3 |
83% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.loss_fn |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.run_with_cache |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.run_with_cache |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.set_tokenizer |
22 |
3 |
0 |
16 |
4 |
82% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.to_tokens |
11 |
1 |
0 |
6 |
1 |
88% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.to_string |
8 |
0 |
0 |
6 |
1 |
93% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.to_str_tokens |
26 |
2 |
0 |
16 |
2 |
90% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.to_single_token |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.to_single_str_token |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.get_token_position |
17 |
1 |
0 |
12 |
2 |
90% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.tokens_to_residual_directions |
13 |
1 |
0 |
8 |
1 |
90% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.to |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.cuda |
5 |
4 |
0 |
4 |
0 |
11% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.cpu |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.mps |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.move_model_modules_to_device |
10 |
0 |
0 |
6 |
1 |
94% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.from_pretrained |
53 |
32 |
0 |
32 |
9 |
38% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.from_pretrained_no_processing |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.init_weights |
14 |
11 |
0 |
14 |
2 |
18% |
| transformer_lens/HookedTransformer.py |
HookedTransformer._init_weights_gpt2 |
3 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer._init_weights_xavier |
7 |
7 |
0 |
8 |
0 |
0% |
| transformer_lens/HookedTransformer.py |
HookedTransformer._init_weights_kaiming |
7 |
7 |
0 |
8 |
0 |
0% |
| transformer_lens/HookedTransformer.py |
HookedTransformer._init_weights_muP |
13 |
13 |
0 |
12 |
0 |
0% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.load_and_process_state_dict |
41 |
23 |
0 |
30 |
8 |
45% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.fill_missing_keys |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.fold_layer_norm |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.center_writing_weights |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.center_unembed |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.fold_value_biases |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.refactor_factored_attn_matrices |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.set_use_attn_result |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.set_use_split_qkv_input |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.set_use_hook_mlp_in |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.set_use_attn_in |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.set_ungroup_grouped_query_attention |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.process_weights_ |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.generate |
121 |
9 |
0 |
64 |
11 |
89% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.generate._logits_to_tuple |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.generate_stream |
66 |
65 |
0 |
36 |
0 |
1% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.b_U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_E |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_pos |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_E_pos |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer._get_blocks |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_in |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_gate |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.W_out |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.b_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.b_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.b_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.b_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.b_in |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.b_out |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.QK |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.OV |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.accumulated_bias |
11 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.all_composition_scores |
12 |
11 |
0 |
6 |
0 |
6% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.all_head_labels |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.load_sample_training_dataset |
5 |
5 |
0 |
2 |
0 |
0% |
| transformer_lens/HookedTransformer.py |
HookedTransformer.sample_datapoint |
8 |
7 |
0 |
4 |
0 |
8% |
| transformer_lens/HookedTransformer.py |
(no function) |
136 |
0 |
74 |
0 |
0 |
100% |
| transformer_lens/SVDInterpreter.py |
SVDInterpreter.__init__ |
5 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/SVDInterpreter.py |
SVDInterpreter.get_singular_vectors |
14 |
1 |
0 |
8 |
1 |
91% |
| transformer_lens/SVDInterpreter.py |
SVDInterpreter._get_singular_vectors_from_matrix |
7 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/SVDInterpreter.py |
SVDInterpreter._get_OV_matrix |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/SVDInterpreter.py |
SVDInterpreter._get_w_in_matrix |
6 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/SVDInterpreter.py |
SVDInterpreter._get_w_out_matrix |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/SVDInterpreter.py |
(no function) |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/__init__.py |
(no function) |
19 |
2 |
0 |
0 |
0 |
89% |
| transformer_lens/benchmarks/__init__.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/activation_cache.py |
benchmark_run_with_cache |
33 |
9 |
0 |
24 |
10 |
67% |
| transformer_lens/benchmarks/activation_cache.py |
benchmark_activation_cache |
28 |
6 |
0 |
12 |
4 |
75% |
| transformer_lens/benchmarks/activation_cache.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/audio.py |
benchmark_audio_forward |
27 |
27 |
0 |
14 |
0 |
0% |
| transformer_lens/benchmarks/audio.py |
benchmark_audio_cache |
24 |
24 |
0 |
10 |
0 |
0% |
| transformer_lens/benchmarks/audio.py |
benchmark_audio_representation_stability |
21 |
21 |
0 |
4 |
0 |
0% |
| transformer_lens/benchmarks/audio.py |
benchmark_audio_representation_stability._extract_states |
7 |
7 |
0 |
6 |
0 |
0% |
| transformer_lens/benchmarks/audio.py |
benchmark_audio_feature_extractor |
24 |
24 |
0 |
12 |
0 |
0% |
| transformer_lens/benchmarks/audio.py |
benchmark_audio_ctc_decode |
26 |
26 |
0 |
8 |
0 |
0% |
| transformer_lens/benchmarks/audio.py |
run_audio_benchmarks |
21 |
21 |
0 |
12 |
0 |
0% |
| transformer_lens/benchmarks/audio.py |
(no function) |
10 |
10 |
0 |
0 |
0 |
0% |
| transformer_lens/benchmarks/backward_gradients.py |
benchmark_backward_hooks |
85 |
18 |
0 |
48 |
13 |
72% |
| transformer_lens/benchmarks/backward_gradients.py |
benchmark_critical_backward_hooks |
71 |
13 |
0 |
44 |
13 |
76% |
| transformer_lens/benchmarks/backward_gradients.py |
benchmark_gradient_computation |
35 |
10 |
0 |
18 |
7 |
60% |
| transformer_lens/benchmarks/backward_gradients.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/component_benchmark.py |
benchmark_all_components |
30 |
20 |
0 |
18 |
3 |
27% |
| transformer_lens/benchmarks/component_benchmark.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentTestResult.get_failure_severity |
11 |
10 |
0 |
10 |
0 |
5% |
| transformer_lens/benchmarks/component_outputs.py |
BenchmarkReport.pass_rate |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/benchmarks/component_outputs.py |
BenchmarkReport.print_summary |
19 |
18 |
0 |
10 |
0 |
3% |
| transformer_lens/benchmarks/component_outputs.py |
BenchmarkReport._print_component_result |
16 |
15 |
0 |
8 |
0 |
4% |
| transformer_lens/benchmarks/component_outputs.py |
BenchmarkReport.get_component_type_summary |
10 |
1 |
0 |
6 |
1 |
88% |
| transformer_lens/benchmarks/component_outputs.py |
BenchmarkReport.get_failure_by_severity |
7 |
6 |
0 |
6 |
0 |
8% |
| transformer_lens/benchmarks/component_outputs.py |
BenchmarkReport.print_detailed_analysis |
22 |
21 |
0 |
12 |
0 |
3% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentBenchmarker.__init__ |
32 |
15 |
0 |
10 |
3 |
48% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentBenchmarker.benchmark_all_components |
37 |
3 |
0 |
28 |
7 |
85% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentBenchmarker._test_component_recursive |
67 |
32 |
0 |
42 |
12 |
54% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentBenchmarker._test_component |
46 |
11 |
0 |
20 |
4 |
74% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentBenchmarker._run_component |
49 |
27 |
0 |
24 |
6 |
49% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentBenchmarker._get_test_input_for_component |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentBenchmarker._generate_test_inputs |
9 |
2 |
0 |
0 |
0 |
78% |
| transformer_lens/benchmarks/component_outputs.py |
ComponentBenchmarker._compare_outputs |
11 |
1 |
0 |
2 |
1 |
85% |
| transformer_lens/benchmarks/component_outputs.py |
benchmark_model |
17 |
16 |
0 |
2 |
0 |
5% |
| transformer_lens/benchmarks/component_outputs.py |
(no function) |
43 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/forward_pass.py |
_is_encoder_decoder |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/benchmarks/forward_pass.py |
_get_decoder_input_ids |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/benchmarks/forward_pass.py |
benchmark_forward_pass |
52 |
39 |
0 |
26 |
4 |
22% |
| transformer_lens/benchmarks/forward_pass.py |
benchmark_loss_equivalence |
18 |
9 |
0 |
10 |
3 |
43% |
| transformer_lens/benchmarks/forward_pass.py |
benchmark_logits_equivalence |
16 |
8 |
0 |
10 |
3 |
42% |
| transformer_lens/benchmarks/forward_pass.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/generation.py |
benchmark_generation |
15 |
4 |
0 |
6 |
3 |
67% |
| transformer_lens/benchmarks/generation.py |
benchmark_generation_with_kv_cache |
9 |
3 |
0 |
4 |
2 |
62% |
| transformer_lens/benchmarks/generation.py |
benchmark_multiple_generation_calls |
12 |
3 |
0 |
6 |
2 |
72% |
| transformer_lens/benchmarks/generation.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/granular_weight_processing.py |
WeightProcessingConfig.__str__ |
12 |
12 |
0 |
10 |
0 |
0% |
| transformer_lens/benchmarks/granular_weight_processing.py |
run_granular_weight_processing_benchmarks |
132 |
132 |
0 |
74 |
0 |
0% |
| transformer_lens/benchmarks/granular_weight_processing.py |
(no function) |
17 |
17 |
0 |
0 |
0 |
0% |
| transformer_lens/benchmarks/hook_registration.py |
benchmark_hook_registry |
20 |
9 |
0 |
10 |
3 |
47% |
| transformer_lens/benchmarks/hook_registration.py |
benchmark_forward_hooks |
59 |
15 |
0 |
38 |
14 |
68% |
| transformer_lens/benchmarks/hook_registration.py |
benchmark_gated_hooks_fire |
56 |
10 |
0 |
28 |
5 |
82% |
| transformer_lens/benchmarks/hook_registration.py |
benchmark_critical_forward_hooks |
60 |
18 |
0 |
38 |
12 |
67% |
| transformer_lens/benchmarks/hook_registration.py |
benchmark_hook_functionality |
18 |
6 |
0 |
4 |
1 |
59% |
| transformer_lens/benchmarks/hook_registration.py |
benchmark_hook_functionality.ablation_hook |
14 |
8 |
0 |
4 |
1 |
39% |
| transformer_lens/benchmarks/hook_registration.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/hook_structure.py |
benchmark_forward_hooks_structure |
56 |
55 |
0 |
36 |
0 |
1% |
| transformer_lens/benchmarks/hook_structure.py |
benchmark_backward_hooks_structure |
59 |
58 |
0 |
36 |
0 |
1% |
| transformer_lens/benchmarks/hook_structure.py |
benchmark_activation_cache_structure |
29 |
28 |
0 |
16 |
0 |
2% |
| transformer_lens/benchmarks/hook_structure.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/main_benchmark.py |
should_skip_ht_comparison |
6 |
1 |
0 |
0 |
0 |
83% |
| transformer_lens/benchmarks/main_benchmark.py |
get_auto_model_class |
7 |
1 |
0 |
0 |
0 |
86% |
| transformer_lens/benchmarks/main_benchmark.py |
_fixup_custom_model |
25 |
23 |
0 |
18 |
1 |
7% |
| transformer_lens/benchmarks/main_benchmark.py |
run_comparison_benchmarks |
180 |
110 |
0 |
96 |
21 |
34% |
| transformer_lens/benchmarks/main_benchmark.py |
run_comparison_benchmarks.add_result |
3 |
0 |
0 |
2 |
1 |
80% |
| transformer_lens/benchmarks/main_benchmark.py |
run_benchmark_suite |
566 |
359 |
0 |
326 |
86 |
33% |
| transformer_lens/benchmarks/main_benchmark.py |
run_benchmark_suite.get_memory_mb |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/benchmarks/main_benchmark.py |
run_benchmark_suite.should_run_phase |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/main_benchmark.py |
run_benchmark_suite.add_result |
5 |
0 |
0 |
4 |
1 |
89% |
| transformer_lens/benchmarks/main_benchmark.py |
run_benchmark_suite.cleanup_tensors |
17 |
16 |
0 |
16 |
0 |
3% |
| transformer_lens/benchmarks/main_benchmark.py |
run_benchmark_suite.cleanup_model |
92 |
34 |
0 |
74 |
29 |
56% |
| transformer_lens/benchmarks/main_benchmark.py |
run_benchmark_suite._cleanup_bridge_unprocessed |
3 |
3 |
0 |
2 |
0 |
0% |
| transformer_lens/benchmarks/main_benchmark.py |
update_model_registry |
23 |
22 |
0 |
10 |
0 |
3% |
| transformer_lens/benchmarks/main_benchmark.py |
main |
14 |
14 |
0 |
2 |
0 |
0% |
| transformer_lens/benchmarks/main_benchmark.py |
(no function) |
28 |
1 |
0 |
2 |
1 |
93% |
| transformer_lens/benchmarks/multimodal.py |
_create_test_image |
5 |
5 |
0 |
0 |
0 |
0% |
| transformer_lens/benchmarks/multimodal.py |
_prepare_test_inputs |
20 |
20 |
0 |
10 |
0 |
0% |
| transformer_lens/benchmarks/multimodal.py |
benchmark_multimodal_forward |
20 |
20 |
0 |
10 |
0 |
0% |
| transformer_lens/benchmarks/multimodal.py |
benchmark_multimodal_generation |
19 |
19 |
0 |
10 |
0 |
0% |
| transformer_lens/benchmarks/multimodal.py |
benchmark_multimodal_cache |
17 |
17 |
0 |
8 |
0 |
0% |
| transformer_lens/benchmarks/multimodal.py |
(no function) |
8 |
8 |
0 |
0 |
0 |
0% |
| transformer_lens/benchmarks/text_quality.py |
_load_scoring_model |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/text_quality.py |
_compute_continuation_perplexity |
17 |
2 |
0 |
2 |
1 |
84% |
| transformer_lens/benchmarks/text_quality.py |
_compute_repetition_penalty |
13 |
3 |
0 |
8 |
3 |
71% |
| transformer_lens/benchmarks/text_quality.py |
_perplexity_to_score |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/benchmarks/text_quality.py |
benchmark_text_quality |
59 |
11 |
0 |
26 |
9 |
74% |
| transformer_lens/benchmarks/text_quality.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/utils.py |
is_tiny_test_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/utils.py |
filter_expected_missing_hooks |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/utils.py |
safe_allclose |
7 |
4 |
0 |
4 |
2 |
45% |
| transformer_lens/benchmarks/utils.py |
BenchmarkResult.__str__ |
12 |
11 |
0 |
6 |
0 |
6% |
| transformer_lens/benchmarks/utils.py |
BenchmarkResult.print_immediate |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/utils.py |
make_capture_hook |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/utils.py |
make_capture_hook.hook_fn |
6 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/benchmarks/utils.py |
make_grad_capture_hook |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/utils.py |
make_grad_capture_hook.hook_fn |
6 |
0 |
0 |
6 |
2 |
83% |
| transformer_lens/benchmarks/utils.py |
_squeeze_batch_dim |
7 |
1 |
0 |
6 |
1 |
85% |
| transformer_lens/benchmarks/utils.py |
compare_activation_dicts |
15 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/benchmarks/utils.py |
compare_tensors |
16 |
10 |
0 |
8 |
4 |
42% |
| transformer_lens/benchmarks/utils.py |
compare_scalars |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/benchmarks/utils.py |
format_results |
38 |
37 |
0 |
14 |
0 |
2% |
| transformer_lens/benchmarks/utils.py |
(no function) |
37 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_weight_processing |
28 |
21 |
0 |
16 |
2 |
20% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_weight_sharing |
47 |
21 |
0 |
16 |
5 |
49% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_weight_modification |
49 |
32 |
0 |
16 |
3 |
31% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_layer_norm_folding |
33 |
11 |
0 |
14 |
5 |
62% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_attention_output_centering |
23 |
5 |
0 |
10 |
4 |
73% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_mlp_output_centering |
32 |
12 |
0 |
22 |
8 |
52% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_unembed_centering |
15 |
5 |
0 |
6 |
2 |
57% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_value_bias_folding |
20 |
7 |
0 |
12 |
5 |
56% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_no_nan_inf |
19 |
9 |
0 |
12 |
3 |
48% |
| transformer_lens/benchmarks/weight_processing.py |
benchmark_weight_magnitudes |
41 |
13 |
0 |
28 |
7 |
65% |
| transformer_lens/benchmarks/weight_processing.py |
(no function) |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/cache/__init__.py |
(no function) |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/cache/key_value_cache.py |
TransformerLensKeyValueCache.init_cache |
7 |
4 |
0 |
4 |
1 |
36% |
| transformer_lens/cache/key_value_cache.py |
TransformerLensKeyValueCache.freeze |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/cache/key_value_cache.py |
TransformerLensKeyValueCache.unfreeze |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/cache/key_value_cache.py |
TransformerLensKeyValueCache.append_attention_mask |
5 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/cache/key_value_cache.py |
TransformerLensKeyValueCache.__getitem__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/cache/key_value_cache.py |
(no function) |
18 |
0 |
2 |
0 |
0 |
100% |
| transformer_lens/cache/key_value_cache_entry.py |
TransformerLensKeyValueCacheEntry.init_cache_entry |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/cache/key_value_cache_entry.py |
TransformerLensKeyValueCacheEntry.append |
6 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/cache/key_value_cache_entry.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/__init__.py |
(no function) |
22 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.__init__ |
67 |
18 |
0 |
32 |
10 |
68% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.OV |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.QK |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.forward |
70 |
28 |
0 |
44 |
12 |
56% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention._apply_qk_norm |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.calculate_qkv_matrices |
21 |
12 |
0 |
12 |
4 |
39% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.calculate_attention_scores |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.calculate_z_scores |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.apply_causal_mask |
13 |
2 |
0 |
6 |
2 |
79% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.calculate_sin_cos_rotary |
49 |
35 |
0 |
10 |
5 |
32% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.rotate_every_two |
8 |
2 |
0 |
2 |
1 |
70% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.apply_rotary |
19 |
2 |
0 |
6 |
2 |
84% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention._extend_rotary_embeddings |
5 |
4 |
0 |
0 |
0 |
20% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention._extend_mask |
8 |
7 |
0 |
6 |
0 |
7% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.create_alibi_slope |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.create_alibi_multipliers |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/abstract_attention.py |
AbstractAttention.create_alibi_bias |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/abstract_attention.py |
(no function) |
51 |
2 |
0 |
2 |
1 |
94% |
| transformer_lens/components/attention.py |
Attention.__init__ |
10 |
3 |
0 |
2 |
1 |
67% |
| transformer_lens/components/attention.py |
(no function) |
10 |
1 |
0 |
2 |
1 |
83% |
| transformer_lens/components/bert_block.py |
BertBlock.__init__ |
16 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_block.py |
BertBlock.forward |
17 |
4 |
0 |
2 |
1 |
74% |
| transformer_lens/components/bert_block.py |
(no function) |
12 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_embed.py |
BertEmbed.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_embed.py |
BertEmbed.forward |
10 |
9 |
0 |
2 |
0 |
8% |
| transformer_lens/components/bert_embed.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_mlm_head.py |
BertMLMHead.__init__ |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_mlm_head.py |
BertMLMHead.forward |
4 |
3 |
0 |
0 |
0 |
25% |
| transformer_lens/components/bert_mlm_head.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_nsp_head.py |
BertNSPHead.__init__ |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_nsp_head.py |
BertNSPHead.forward |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/components/bert_nsp_head.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_pooler.py |
BertPooler.__init__ |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/bert_pooler.py |
BertPooler.forward |
4 |
3 |
0 |
0 |
0 |
25% |
| transformer_lens/components/bert_pooler.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/embed.py |
Embed.__init__ |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/components/embed.py |
Embed.forward |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/components/embed.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.__init__ |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.W_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.W_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.W_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.W_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.b_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.b_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.b_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.b_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.calculate_qkv_matrices |
10 |
4 |
0 |
2 |
1 |
58% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.calculate_attention_scores |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention.calculate_z_scores |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/components/grouped_query_attention.py |
GroupedQueryAttention._apply_qk_norm |
4 |
3 |
0 |
0 |
0 |
25% |
| transformer_lens/components/grouped_query_attention.py |
(no function) |
30 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/layer_norm.py |
LayerNorm.__init__ |
10 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/components/layer_norm.py |
LayerNorm.forward |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/components/layer_norm.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/layer_norm_pre.py |
LayerNormPre.__init__ |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/layer_norm_pre.py |
LayerNormPre.forward |
5 |
1 |
0 |
2 |
1 |
71% |
| transformer_lens/components/layer_norm_pre.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/can_be_used_as_mlp.py |
CanBeUsedAsMLP.__init__ |
5 |
1 |
0 |
2 |
1 |
71% |
| transformer_lens/components/mlps/can_be_used_as_mlp.py |
CanBeUsedAsMLP.forward |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/can_be_used_as_mlp.py |
CanBeUsedAsMLP.select_activation_function |
6 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/components/mlps/can_be_used_as_mlp.py |
(no function) |
19 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/gated_mlp.py |
GatedMLP.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/gated_mlp.py |
GatedMLP.forward |
9 |
3 |
0 |
4 |
2 |
62% |
| transformer_lens/components/mlps/gated_mlp.py |
(no function) |
14 |
1 |
0 |
2 |
1 |
88% |
| transformer_lens/components/mlps/gated_mlp_4bit.py |
GatedMLP4Bit.__init__ |
11 |
10 |
0 |
0 |
0 |
9% |
| transformer_lens/components/mlps/gated_mlp_4bit.py |
GatedMLP4Bit.forward |
7 |
6 |
0 |
2 |
0 |
11% |
| transformer_lens/components/mlps/gated_mlp_4bit.py |
(no function) |
14 |
2 |
0 |
2 |
1 |
81% |
| transformer_lens/components/mlps/gpt_oss_moe.py |
GptOssExpert.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/gpt_oss_moe.py |
GptOssExpert.forward |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/gpt_oss_moe.py |
GptOssMoE.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/gpt_oss_moe.py |
GptOssMoE.forward |
21 |
1 |
0 |
4 |
1 |
92% |
| transformer_lens/components/mlps/gpt_oss_moe.py |
(no function) |
16 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/mlp.py |
MLP.__init__ |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/mlp.py |
MLP.forward |
6 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/components/mlps/mlp.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/moe.py |
MoEGatedMLP.__init__ |
12 |
1 |
0 |
2 |
1 |
86% |
| transformer_lens/components/mlps/moe.py |
MoEGatedMLP.forward |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/moe.py |
MoE.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/mlps/moe.py |
MoE.forward |
19 |
1 |
0 |
4 |
1 |
91% |
| transformer_lens/components/mlps/moe.py |
(no function) |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/pos_embed.py |
PosEmbed.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/pos_embed.py |
PosEmbed.forward |
10 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/components/pos_embed.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/rms_norm.py |
RMSNorm.__init__ |
9 |
8 |
0 |
2 |
0 |
9% |
| transformer_lens/components/rms_norm.py |
RMSNorm.forward |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/components/rms_norm.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/rms_norm_pre.py |
RMSNormPre.__init__ |
5 |
4 |
0 |
0 |
0 |
20% |
| transformer_lens/components/rms_norm_pre.py |
RMSNormPre.forward |
4 |
3 |
0 |
2 |
0 |
17% |
| transformer_lens/components/rms_norm_pre.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/t5_attention.py |
T5Attention.__init__ |
16 |
15 |
0 |
6 |
0 |
5% |
| transformer_lens/components/t5_attention.py |
T5Attention._relative_position_bucket |
12 |
11 |
0 |
2 |
0 |
7% |
| transformer_lens/components/t5_attention.py |
T5Attention.compute_relative_attention_bias |
9 |
8 |
0 |
2 |
0 |
9% |
| transformer_lens/components/t5_attention.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/t5_block.py |
T5Block.__init__ |
24 |
23 |
0 |
4 |
0 |
4% |
| transformer_lens/components/t5_block.py |
T5Block.forward |
27 |
26 |
0 |
8 |
0 |
3% |
| transformer_lens/components/t5_block.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/token_typed_embed.py |
TokenTypeEmbed.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/token_typed_embed.py |
TokenTypeEmbed.forward |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/token_typed_embed.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/transformer_block.py |
TransformerBlock.__init__ |
48 |
16 |
0 |
32 |
5 |
59% |
| transformer_lens/components/transformer_block.py |
TransformerBlock.forward |
40 |
7 |
0 |
22 |
6 |
79% |
| transformer_lens/components/transformer_block.py |
TransformerBlock.apply_mlp |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/components/transformer_block.py |
(no function) |
18 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/unembed.py |
Unembed.__init__ |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/unembed.py |
Unembed.forward |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/components/unembed.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/HookedTransformerConfig.py |
HookedTransformerConfig.__post_init__ |
37 |
4 |
0 |
28 |
4 |
88% |
| transformer_lens/config/HookedTransformerConfig.py |
HookedTransformerConfig.unwrap |
5 |
0 |
0 |
4 |
1 |
89% |
| transformer_lens/config/HookedTransformerConfig.py |
HookedTransformerConfig.from_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/HookedTransformerConfig.py |
HookedTransformerConfig.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/HookedTransformerConfig.py |
HookedTransformerConfig.__repr__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/config/HookedTransformerConfig.py |
HookedTransformerConfig.set_seed_everywhere |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/config/HookedTransformerConfig.py |
HookedTransformerConfig.is_layer_norm_activation |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/HookedTransformerConfig.py |
(no function) |
85 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/TransformerBridgeConfig.py |
TransformerBridgeConfig.__init__ |
71 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/TransformerBridgeConfig.py |
TransformerBridgeConfig.__post_init__ |
4 |
1 |
0 |
4 |
2 |
62% |
| transformer_lens/config/TransformerBridgeConfig.py |
TransformerBridgeConfig.head_dim |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/TransformerBridgeConfig.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/TransformerLensConfig.py |
TransformerLensConfig.__post_init__ |
8 |
1 |
0 |
8 |
1 |
88% |
| transformer_lens/config/TransformerLensConfig.py |
TransformerLensConfig.unwrap |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/TransformerLensConfig.py |
TransformerLensConfig.from_dict |
12 |
2 |
0 |
4 |
1 |
81% |
| transformer_lens/config/TransformerLensConfig.py |
TransformerLensConfig.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/TransformerLensConfig.py |
TransformerLensConfig.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/TransformerLensConfig.py |
(no function) |
39 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/config/__init__.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/__init__.py |
(no function) |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/__init__.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/arithmetic_tensor_conversion.py |
ArithmeticTensorConversion.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/arithmetic_tensor_conversion.py |
ArithmeticTensorConversion.handle_conversion |
9 |
0 |
0 |
8 |
1 |
94% |
| transformer_lens/conversion_utils/conversion_steps/arithmetic_tensor_conversion.py |
ArithmeticTensorConversion.revert |
11 |
11 |
0 |
8 |
0 |
0% |
| transformer_lens/conversion_utils/conversion_steps/arithmetic_tensor_conversion.py |
ArithmeticTensorConversion.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/arithmetic_tensor_conversion.py |
(no function) |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/attention_auto_conversion.py |
AttentionAutoConversion.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/attention_auto_conversion.py |
AttentionAutoConversion.handle_conversion |
19 |
11 |
0 |
10 |
2 |
34% |
| transformer_lens/conversion_utils/conversion_steps/attention_auto_conversion.py |
AttentionAutoConversion.revert_conversion |
13 |
12 |
0 |
10 |
0 |
4% |
| transformer_lens/conversion_utils/conversion_steps/attention_auto_conversion.py |
AttentionAutoConversion.clear_state |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/conversion_utils/conversion_steps/attention_auto_conversion.py |
AttentionAutoConversion.get_conversion_info |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/attention_auto_conversion.py |
AttentionAutoConversion.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/attention_auto_conversion.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/base_tensor_conversion.py |
BaseTensorConversion.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/base_tensor_conversion.py |
BaseTensorConversion.convert |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/base_tensor_conversion.py |
BaseTensorConversion.handle_conversion |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/base_tensor_conversion.py |
BaseTensorConversion.revert |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/base_tensor_conversion.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/callable_tensor_conversion.py |
CallableTensorConversion.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/callable_tensor_conversion.py |
CallableTensorConversion.handle_conversion |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/callable_tensor_conversion.py |
CallableTensorConversion.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/callable_tensor_conversion.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/chain_tensor_conversion.py |
ChainTensorConversion.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/chain_tensor_conversion.py |
ChainTensorConversion.handle_conversion |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/conversion_utils/conversion_steps/chain_tensor_conversion.py |
ChainTensorConversion.revert |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/conversion_utils/conversion_steps/chain_tensor_conversion.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/rearrange_tensor_conversion.py |
RearrangeTensorConversion.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/rearrange_tensor_conversion.py |
RearrangeTensorConversion.handle_conversion |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/rearrange_tensor_conversion.py |
RearrangeTensorConversion.revert |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/rearrange_tensor_conversion.py |
RearrangeTensorConversion.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/rearrange_tensor_conversion.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/repeat_tensor_conversion.py |
RepeatTensorConversion.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/repeat_tensor_conversion.py |
RepeatTensorConversion.handle_conversion |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/repeat_tensor_conversion.py |
RepeatTensorConversion.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/repeat_tensor_conversion.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/split_tensor_conversion.py |
SplitTensorConversion.__init__ |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/split_tensor_conversion.py |
SplitTensorConversion.handle_conversion |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/conversion_utils/conversion_steps/split_tensor_conversion.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/tensor_conversion_set.py |
TensorConversionSet.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/tensor_conversion_set.py |
TensorConversionSet.get_component |
11 |
10 |
0 |
6 |
0 |
6% |
| transformer_lens/conversion_utils/conversion_steps/tensor_conversion_set.py |
TensorConversionSet.handle_conversion |
5 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/tensor_conversion_set.py |
TensorConversionSet.process_conversion_action |
6 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/tensor_conversion_set.py |
TensorConversionSet.process_conversion |
7 |
4 |
0 |
4 |
1 |
36% |
| transformer_lens/conversion_utils/conversion_steps/tensor_conversion_set.py |
TensorConversionSet.get_conversion_action |
4 |
3 |
0 |
2 |
0 |
17% |
| transformer_lens/conversion_utils/conversion_steps/tensor_conversion_set.py |
TensorConversionSet.__repr__ |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/tensor_conversion_set.py |
(no function) |
14 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/ternary_tensor_conversion.py |
TernaryTensorConversion.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/ternary_tensor_conversion.py |
TernaryTensorConversion.handle_conversion |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/ternary_tensor_conversion.py |
TernaryTensorConversion.handle_primary_conversion |
5 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/ternary_tensor_conversion.py |
TernaryTensorConversion.handle_fallback_conversion |
7 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/ternary_tensor_conversion.py |
TernaryTensorConversion.find_context_field |
5 |
0 |
0 |
4 |
1 |
89% |
| transformer_lens/conversion_utils/conversion_steps/ternary_tensor_conversion.py |
TernaryTensorConversion.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/ternary_tensor_conversion.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/transpose_tensor_conversion.py |
TransposeTensorConversion.handle_conversion |
5 |
2 |
0 |
4 |
2 |
56% |
| transformer_lens/conversion_utils/conversion_steps/transpose_tensor_conversion.py |
TransposeTensorConversion.revert |
5 |
2 |
0 |
4 |
2 |
56% |
| transformer_lens/conversion_utils/conversion_steps/transpose_tensor_conversion.py |
TransposeTensorConversion.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/transpose_tensor_conversion.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/zeros_like_conversion.py |
ZerosLikeConversion.handle_conversion |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/zeros_like_conversion.py |
ZerosLikeConversion.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/conversion_steps/zeros_like_conversion.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/helpers/__init__.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/helpers/find_property.py |
find_property |
6 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/conversion_utils/helpers/find_property.py |
(no function) |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/helpers/merge_quantiziation_fields.py |
merge_quantization_fields |
15 |
1 |
0 |
10 |
1 |
92% |
| transformer_lens/conversion_utils/helpers/merge_quantiziation_fields.py |
(no function) |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/hook_conversion_utils.py |
get_weight_conversion_field_set |
14 |
3 |
0 |
10 |
1 |
75% |
| transformer_lens/conversion_utils/hook_conversion_utils.py |
model_info_cfg |
2 |
2 |
0 |
0 |
0 |
0% |
| transformer_lens/conversion_utils/hook_conversion_utils.py |
model_info |
2 |
2 |
0 |
0 |
0 |
0% |
| transformer_lens/conversion_utils/hook_conversion_utils.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/param_processing_conversion.py |
ParamProcessingConversion.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/param_processing_conversion.py |
ParamProcessingConversion._resolve_key |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/conversion_utils/param_processing_conversion.py |
ParamProcessingConversion.convert |
5 |
1 |
0 |
2 |
1 |
71% |
| transformer_lens/conversion_utils/param_processing_conversion.py |
ParamProcessingConversion.revert |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/conversion_utils/param_processing_conversion.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/evals.py |
sanity_check |
2 |
2 |
0 |
0 |
0 |
0% |
| transformer_lens/evals.py |
make_wiki_data_loader |
5 |
5 |
0 |
0 |
0 |
0% |
| transformer_lens/evals.py |
make_owt_data_loader |
5 |
5 |
0 |
0 |
0 |
0% |
| transformer_lens/evals.py |
make_pile_data_loader |
5 |
5 |
0 |
0 |
0 |
0% |
| transformer_lens/evals.py |
make_code_data_loader |
5 |
5 |
0 |
0 |
0 |
0% |
| transformer_lens/evals.py |
make_mmlu_data_loader |
21 |
4 |
0 |
10 |
1 |
84% |
| transformer_lens/evals.py |
evaluate_on_dataset |
10 |
10 |
0 |
4 |
0 |
0% |
| transformer_lens/evals.py |
induction_loss |
11 |
11 |
0 |
4 |
0 |
0% |
| transformer_lens/evals.py |
evaluate |
9 |
9 |
0 |
4 |
0 |
0% |
| transformer_lens/evals.py |
IOIDataset.__init__ |
8 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/evals.py |
IOIDataset.__len__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/evals.py |
IOIDataset.__getitem__ |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/evals.py |
IOIDataset.get_sample |
14 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/evals.py |
IOIDataset.get_default_names |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/evals.py |
IOIDataset.get_default_templates |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/evals.py |
IOIDataset.get_default_nouns |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/evals.py |
ioi_eval |
29 |
0 |
0 |
8 |
1 |
97% |
| transformer_lens/evals.py |
ioi_eval.collate |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/evals.py |
mmlu_eval |
46 |
3 |
0 |
16 |
3 |
90% |
| transformer_lens/evals.py |
(no function) |
40 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/factories/__init__.py |
(no function) |
0 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/factories/activation_function_factory.py |
ActivationFunctionFactory.pick_activation_function |
9 |
2 |
0 |
6 |
2 |
73% |
| transformer_lens/factories/activation_function_factory.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/factories/architecture_adapter_factory.py |
ArchitectureAdapterFactory.select_architecture_adapter |
5 |
0 |
0 |
4 |
1 |
89% |
| transformer_lens/factories/architecture_adapter_factory.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/factories/mlp_factory.py |
MLPFactory.create_mlp |
7 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/factories/mlp_factory.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/head_detector.py |
detect_head |
31 |
2 |
0 |
18 |
2 |
92% |
| transformer_lens/head_detector.py |
get_previous_token_head_detection_pattern |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/head_detector.py |
get_duplicate_token_head_detection_pattern |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/head_detector.py |
get_induction_head_detection_pattern |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/head_detector.py |
get_supported_heads |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/head_detector.py |
compute_head_attention_similarity_score |
16 |
0 |
0 |
10 |
0 |
100% |
| transformer_lens/head_detector.py |
(no function) |
23 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
_ScaledGradientTensor.__init__ |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/hook_points.py |
_ScaledGradientTensor.sum |
4 |
4 |
0 |
2 |
0 |
0% |
| transformer_lens/hook_points.py |
_ScaledGradientTensor.__getattr__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/hook_points.py |
_ScaledGradientTensor.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
_HookFunctionProtocol.__call__ |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
_AliasedHookPoint.__init__ |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/hook_points.py |
_AliasedHookPoint.name |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
_AliasedHookPoint.ctx |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
_AliasedHookPoint.hook_conversion |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/hook_points.py |
_AliasedHookPoint.layer |
4 |
3 |
0 |
2 |
0 |
17% |
| transformer_lens/hook_points.py |
HookPoint.__init__ |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookPoint.add_perma_hook |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookPoint.add_hook |
25 |
2 |
0 |
14 |
3 |
87% |
| transformer_lens/hook_points.py |
HookPoint.add_hook.full_hook |
19 |
7 |
0 |
16 |
2 |
63% |
| transformer_lens/hook_points.py |
HookPoint.has_hooks |
8 |
0 |
0 |
6 |
1 |
93% |
| transformer_lens/hook_points.py |
HookPoint.has_hooks._has_hooks_in_direction |
7 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/hook_points.py |
HookPoint.remove_hooks |
7 |
0 |
0 |
6 |
2 |
85% |
| transformer_lens/hook_points.py |
HookPoint.remove_hooks._remove_hooks |
8 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/hook_points.py |
HookPoint.clear_context |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookPoint.enable_reshape |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookPoint.forward |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookPoint.layer |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/hook_points.py |
HookedRootModule.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.setup |
9 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.hook_points |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.remove_all_hook_fns |
2 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.clear_contexts |
2 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.reset_hooks |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.check_and_add_hook |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.check_hooks_to_add |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.add_hook |
7 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.add_perma_hook |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule._enable_hook_with_name |
7 |
2 |
0 |
4 |
2 |
64% |
| transformer_lens/hook_points.py |
HookedRootModule._enable_hooks_for_points |
3 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule._enable_hook |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/hook_points.py |
HookedRootModule.hooks |
10 |
0 |
0 |
6 |
1 |
94% |
| transformer_lens/hook_points.py |
HookedRootModule.run_with_hooks |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/hook_points.py |
HookedRootModule.add_caching_hooks |
21 |
20 |
0 |
16 |
0 |
3% |
| transformer_lens/hook_points.py |
HookedRootModule.add_caching_hooks.save_hook |
7 |
7 |
0 |
4 |
0 |
0% |
| transformer_lens/hook_points.py |
HookedRootModule.run_with_cache |
7 |
1 |
0 |
2 |
1 |
78% |
| transformer_lens/hook_points.py |
HookedRootModule.get_caching_hooks |
27 |
5 |
0 |
18 |
5 |
78% |
| transformer_lens/hook_points.py |
HookedRootModule.get_caching_hooks.save_hook |
14 |
2 |
0 |
10 |
3 |
79% |
| transformer_lens/hook_points.py |
HookedRootModule.cache_all |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/hook_points.py |
HookedRootModule.cache_some |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/hook_points.py |
(no function) |
77 |
0 |
4 |
0 |
0 |
100% |
| transformer_lens/lit/__init__.py |
serve |
26 |
25 |
0 |
16 |
0 |
2% |
| transformer_lens/lit/__init__.py |
LITWidget.__init__ |
10 |
9 |
0 |
6 |
0 |
6% |
| transformer_lens/lit/__init__.py |
LITWidget.url |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/lit/__init__.py |
LITWidget.render |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/__init__.py |
LITWidget.stop |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/__init__.py |
(no function) |
21 |
1 |
0 |
2 |
1 |
91% |
| transformer_lens/lit/constants.py |
(no function) |
80 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
_ensure_lit_available |
2 |
2 |
0 |
2 |
0 |
0% |
| transformer_lens/lit/dataset.py |
SimpleTextDataset.__init__ |
6 |
5 |
0 |
4 |
0 |
10% |
| transformer_lens/lit/dataset.py |
SimpleTextDataset.examples |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
SimpleTextDataset.__len__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
SimpleTextDataset.__iter__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
SimpleTextDataset.description |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
SimpleTextDataset.spec |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
SimpleTextDataset.from_strings |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/lit/dataset.py |
SimpleTextDataset.from_file |
9 |
8 |
0 |
4 |
0 |
8% |
| transformer_lens/lit/dataset.py |
PromptCompletionDataset.__init__ |
6 |
5 |
0 |
4 |
0 |
10% |
| transformer_lens/lit/dataset.py |
PromptCompletionDataset._add_example |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/lit/dataset.py |
PromptCompletionDataset.examples |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
PromptCompletionDataset.__len__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
PromptCompletionDataset.__iter__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
PromptCompletionDataset.description |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
PromptCompletionDataset.spec |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
PromptCompletionDataset.from_pairs |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/lit/dataset.py |
IOIDataset.__init__ |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/lit/dataset.py |
IOIDataset.examples |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
IOIDataset.__len__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
IOIDataset.__iter__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
IOIDataset.description |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
IOIDataset.spec |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
IOIDataset.add_example |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/lit/dataset.py |
IOIDataset.generate |
9 |
8 |
0 |
2 |
0 |
9% |
| transformer_lens/lit/dataset.py |
InductionDataset.__init__ |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/lit/dataset.py |
InductionDataset.examples |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
InductionDataset.__len__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
InductionDataset.__iter__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
InductionDataset.description |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
InductionDataset.spec |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/dataset.py |
InductionDataset.add_example |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/lit/dataset.py |
InductionDataset.generate_simple |
11 |
10 |
0 |
2 |
0 |
8% |
| transformer_lens/lit/dataset.py |
LITDatasetWrapper.__init__ |
4 |
4 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
LITDatasetWrapper.init_spec |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
LITDatasetWrapper.spec |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
LITDatasetWrapper.description |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
LITDatasetWrapper.examples |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
LITDatasetWrapper.__len__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
LITDatasetWrapper.__iter__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
wrap_for_lit |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/dataset.py |
(no function) |
92 |
15 |
3 |
4 |
2 |
82% |
| transformer_lens/lit/model.py |
_ensure_lit_available |
2 |
2 |
0 |
2 |
0 |
0% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.__init__ |
13 |
12 |
0 |
6 |
0 |
5% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.supports_concurrent_predictions |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.description |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.init_spec |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.input_spec |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.output_spec |
20 |
19 |
0 |
10 |
0 |
3% |
| transformer_lens/lit/model.py |
HookedTransformerLIT._get_embedding_layers |
8 |
7 |
0 |
6 |
0 |
7% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.predict |
2 |
1 |
0 |
2 |
0 |
25% |
| transformer_lens/lit/model.py |
HookedTransformerLIT._predict_single |
20 |
19 |
0 |
8 |
0 |
4% |
| transformer_lens/lit/model.py |
HookedTransformerLIT._get_top_k_per_position |
14 |
13 |
0 |
8 |
0 |
5% |
| transformer_lens/lit/model.py |
HookedTransformerLIT._extract_embeddings |
12 |
11 |
0 |
2 |
0 |
7% |
| transformer_lens/lit/model.py |
HookedTransformerLIT._extract_attention |
6 |
5 |
0 |
2 |
0 |
12% |
| transformer_lens/lit/model.py |
HookedTransformerLIT._compute_gradients |
23 |
22 |
0 |
6 |
0 |
3% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.max_minibatch_size |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.get_embedding_table |
14 |
13 |
0 |
4 |
0 |
6% |
| transformer_lens/lit/model.py |
HookedTransformerLIT.from_pretrained |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/lit/model.py |
HookedTransformerLITBatched.__init__ |
3 |
3 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/model.py |
HookedTransformerLITBatched.description |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/model.py |
HookedTransformerLITBatched.init_spec |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/model.py |
HookedTransformerLITBatched.input_spec |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/model.py |
HookedTransformerLITBatched.output_spec |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/model.py |
HookedTransformerLITBatched.max_minibatch_size |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/model.py |
HookedTransformerLITBatched.predict_minibatch |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/model.py |
HookedTransformerLITBatched.from_pretrained |
3 |
3 |
0 |
0 |
0 |
0% |
| transformer_lens/lit/model.py |
(no function) |
65 |
16 |
3 |
6 |
3 |
73% |
| transformer_lens/lit/utils.py |
check_lit_installed |
5 |
1 |
0 |
0 |
0 |
80% |
| transformer_lens/lit/utils.py |
tensor_to_numpy |
7 |
0 |
0 |
6 |
1 |
92% |
| transformer_lens/lit/utils.py |
numpy_to_tensor |
10 |
3 |
0 |
8 |
3 |
67% |
| transformer_lens/lit/utils.py |
get_tokens_from_model |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/lit/utils.py |
clean_token_string |
7 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/lit/utils.py |
clean_token_strings |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/utils.py |
extract_attention_from_cache |
6 |
5 |
0 |
4 |
0 |
10% |
| transformer_lens/lit/utils.py |
extract_embeddings_from_cache |
11 |
10 |
0 |
8 |
0 |
5% |
| transformer_lens/lit/utils.py |
compute_token_gradients |
13 |
12 |
0 |
2 |
0 |
7% |
| transformer_lens/lit/utils.py |
get_top_k_predictions |
8 |
7 |
0 |
2 |
0 |
10% |
| transformer_lens/lit/utils.py |
validate_input_example |
5 |
4 |
0 |
4 |
0 |
11% |
| transformer_lens/lit/utils.py |
batch_examples |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/utils.py |
unbatch_outputs |
15 |
2 |
0 |
10 |
2 |
84% |
| transformer_lens/lit/utils.py |
get_hook_name_for_layer |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/utils.py |
filter_cache_by_pattern |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/utils.py |
get_model_info |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/lit/utils.py |
(no function) |
22 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/loading_from_pretrained.py |
_get_rope_theta |
7 |
1 |
0 |
4 |
2 |
73% |
| transformer_lens/loading_from_pretrained.py |
make_model_alias_map |
7 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/loading_from_pretrained.py |
get_official_model_name |
5 |
1 |
0 |
2 |
1 |
71% |
| transformer_lens/loading_from_pretrained.py |
convert_hf_model_config |
173 |
80 |
0 |
140 |
39 |
52% |
| transformer_lens/loading_from_pretrained.py |
convert_neel_model_config |
11 |
2 |
0 |
4 |
2 |
73% |
| transformer_lens/loading_from_pretrained.py |
get_pretrained_model_config |
60 |
22 |
0 |
38 |
7 |
64% |
| transformer_lens/loading_from_pretrained.py |
get_num_params_of_pretrained |
4 |
3 |
0 |
2 |
0 |
17% |
| transformer_lens/loading_from_pretrained.py |
get_checkpoint_labels |
21 |
20 |
0 |
14 |
0 |
3% |
| transformer_lens/loading_from_pretrained.py |
get_pretrained_state_dict |
117 |
75 |
0 |
100 |
17 |
34% |
| transformer_lens/loading_from_pretrained.py |
fill_missing_keys |
9 |
1 |
0 |
6 |
1 |
87% |
| transformer_lens/loading_from_pretrained.py |
get_basic_config |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/loading_from_pretrained.py |
(no function) |
44 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/__init__.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.__init__ |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter._merge_default_config |
3 |
0 |
0 |
4 |
1 |
86% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter._qkvo_weight_conversions |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.preprocess_weights |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.get_component_mapping |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.get_remote_component |
20 |
6 |
0 |
14 |
2 |
65% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.get_component_from_list_module |
32 |
6 |
0 |
22 |
7 |
76% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.get_generalized_component |
31 |
10 |
0 |
24 |
3 |
65% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.get_component |
23 |
9 |
0 |
18 |
5 |
56% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.translate_transformer_lens_path |
76 |
18 |
0 |
56 |
10 |
73% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter._preprocess_parameter_path |
91 |
25 |
0 |
46 |
12 |
72% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.convert_hf_key_to_tl_key |
36 |
6 |
0 |
30 |
10 |
76% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.prepare_loading |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.prepare_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.create_stateful_cache |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter.setup_component_testing |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter._enable_ht_attention |
20 |
20 |
0 |
8 |
0 |
0% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter._extract_qkv_gpt2_style |
11 |
11 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter._extract_output_proj |
5 |
5 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/architecture_adapter.py |
ArchitectureAdapter._disable_hook_conversions |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/architecture_adapter.py |
(no function) |
31 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
_resolve_attr_path |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
build_alias_to_canonical_map |
9 |
1 |
0 |
8 |
2 |
82% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.__init__ |
36 |
8 |
0 |
12 |
4 |
71% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.boot_transformers |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.original_model |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.original_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._register_aliases |
20 |
0 |
0 |
12 |
1 |
97% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._set_processed_weight_attributes |
33 |
32 |
0 |
14 |
0 |
2% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._register_all_aliases_recursive |
5 |
0 |
0 |
6 |
1 |
91% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.__setattr__ |
10 |
2 |
0 |
6 |
1 |
81% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._initialize_hook_registry |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._collect_component_aliases |
15 |
0 |
0 |
12 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._compute_hook_aliases_cached |
19 |
3 |
0 |
12 |
2 |
84% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._collect_hook_aliases_from_registry |
7 |
1 |
0 |
2 |
1 |
78% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._add_aliases_to_hooks |
20 |
1 |
0 |
12 |
3 |
88% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._scan_existing_hooks |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._scan_existing_hooks.scan_module |
38 |
0 |
0 |
26 |
1 |
98% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.hook_dict |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.clear_hook_registry |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._initialize_hooks_to_cache |
49 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.__getattr__ |
16 |
5 |
0 |
10 |
2 |
65% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.__str__ |
4 |
3 |
0 |
0 |
0 |
25% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.enable_compatibility_mode |
11 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.enable_compatibility_mode.set_compatibility_mode |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._setup_hook_compatibility |
19 |
2 |
0 |
20 |
2 |
85% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.process_weights |
42 |
11 |
0 |
28 |
11 |
69% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._calculate_loss |
9 |
9 |
0 |
2 |
0 |
0% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._extract_hf_weights |
10 |
10 |
0 |
8 |
0 |
0% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.to_tokens |
18 |
0 |
0 |
16 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.to_string |
7 |
1 |
0 |
6 |
2 |
77% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.to_str_tokens |
19 |
13 |
0 |
12 |
2 |
26% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.to_single_token |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.get_token_position |
17 |
16 |
0 |
12 |
0 |
3% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.to_single_str_token |
5 |
4 |
0 |
2 |
0 |
14% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.blocks_with |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.stack_params_for |
12 |
1 |
0 |
6 |
1 |
89% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._stack_block_params |
17 |
0 |
0 |
10 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._reshape_qkv |
4 |
2 |
0 |
2 |
1 |
50% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._reshape_o |
4 |
2 |
0 |
2 |
1 |
50% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_in |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_gate |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_out |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.b_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.b_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.b_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.b_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.b_in |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.b_out |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.b_U |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.W_E |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.QK |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.OV |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.QK_for_attn_layers |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.OV_for_attn_layers |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.tokens_to_residual_directions |
13 |
3 |
0 |
8 |
1 |
71% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._get_block_variant_bias |
14 |
2 |
0 |
10 |
2 |
83% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.accumulated_bias |
17 |
0 |
0 |
12 |
1 |
97% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.all_composition_scores |
27 |
1 |
0 |
8 |
1 |
94% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.all_composition_scores._stack |
10 |
2 |
0 |
6 |
2 |
75% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.composition_layer_indices |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.block_hooks |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.block_submodules |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.layer_types |
7 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.all_head_labels |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.attn_head_labels |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.parameters |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.named_parameters |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.tl_parameters |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.tl_named_parameters |
1 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.forward |
114 |
40 |
0 |
80 |
14 |
63% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.get_hook_point |
12 |
11 |
0 |
6 |
0 |
6% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.loss_fn |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_cache |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_cache |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_cache |
110 |
28 |
0 |
84 |
13 |
71% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_cache.create_names_filter_fn |
18 |
6 |
0 |
14 |
3 |
66% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_cache.make_cache_hook |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_cache.make_cache_hook.cache_hook |
14 |
10 |
0 |
10 |
2 |
25% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_cache.stop_hook |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_hooks |
29 |
4 |
0 |
12 |
4 |
80% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_hooks.add_hook_to_point |
15 |
7 |
0 |
8 |
1 |
57% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_hooks.stop_hook |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_hooks.apply_hooks |
24 |
4 |
0 |
16 |
3 |
82% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.run_with_hooks.apply_hooks.wrapped_hook_fn |
7 |
7 |
0 |
4 |
0 |
0% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._generate_tokens |
68 |
8 |
0 |
36 |
4 |
88% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.generate |
104 |
22 |
0 |
60 |
12 |
76% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.generate._logits_to_tuple |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.generate_stream |
61 |
18 |
0 |
36 |
9 |
64% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.generate_stream._maybe_decode |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.hf_generate |
62 |
21 |
0 |
44 |
12 |
59% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.prepare_multimodal_inputs |
6 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.to |
28 |
4 |
0 |
22 |
5 |
78% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.cuda |
5 |
4 |
0 |
4 |
0 |
11% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.cpu |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.mps |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.add_hook |
24 |
12 |
0 |
16 |
4 |
45% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.reset_hooks |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.reset_hooks.remove_hooks_recursive |
4 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.hooks |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.hooks._hooks_context |
10 |
0 |
0 |
4 |
1 |
93% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.hooks._hooks_context.add_hook_to_point |
8 |
5 |
0 |
4 |
1 |
33% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.hooks._hooks_context.apply_hooks |
20 |
10 |
0 |
14 |
3 |
44% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.set_use_attn_result |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.set_use_split_qkv_input |
6 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.set_use_attn_in |
6 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._propagate_attention_flag |
12 |
5 |
0 |
8 |
3 |
60% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._validate_attention_fork_supported |
23 |
4 |
0 |
14 |
4 |
78% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._is_valid_bridge_path |
32 |
5 |
0 |
22 |
6 |
76% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge._normalize_bridge_key_to_hf |
21 |
1 |
0 |
20 |
3 |
90% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.state_dict |
15 |
2 |
0 |
10 |
2 |
84% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.load_state_dict |
18 |
18 |
0 |
10 |
0 |
0% |
| transformer_lens/model_bridge/bridge.py |
TransformerBridge.get_params |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/bridge.py |
(no function) |
159 |
0 |
2 |
0 |
0 |
100% |
| transformer_lens/model_bridge/compat.py |
patch_dynamic_cache_v5 |
14 |
2 |
0 |
6 |
0 |
90% |
| transformer_lens/model_bridge/compat.py |
patch_dynamic_cache_v5._from_legacy_cache |
5 |
5 |
0 |
4 |
0 |
0% |
| transformer_lens/model_bridge/compat.py |
patch_dynamic_cache_v5._get_usable_length |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/compat.py |
patch_dynamic_cache_v5._to_legacy_cache |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/compat.py |
(no function) |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/component_setup.py |
replace_remote_component |
10 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/component_setup.py |
set_original_components |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/component_setup.py |
setup_submodules |
45 |
10 |
0 |
28 |
6 |
75% |
| transformer_lens/model_bridge/component_setup.py |
setup_components |
15 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/model_bridge/component_setup.py |
setup_blocks_bridge |
13 |
1 |
0 |
4 |
1 |
88% |
| transformer_lens/model_bridge/component_setup.py |
(no function) |
16 |
0 |
2 |
0 |
0 |
100% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__torch_function__ |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.shape |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.device |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.dtype |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__getitem__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__getattr__ |
5 |
2 |
0 |
0 |
0 |
60% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__gt__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__lt__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__ge__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__le__ |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__eq__ |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__ne__ |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/model_bridge/composition_scores.py |
CompositionScores.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/composition_scores.py |
(no function) |
21 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/exceptions.py |
StopAtLayerException.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/exceptions.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/__init__.py |
(no function) |
38 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/alibi_joint_qkv_attention.py |
ALiBiJointQKVAttentionBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/alibi_joint_qkv_attention.py |
ALiBiJointQKVAttentionBridge.forward |
14 |
4 |
0 |
6 |
2 |
60% |
| transformer_lens/model_bridge/generalized_components/alibi_joint_qkv_attention.py |
ALiBiJointQKVAttentionBridge._reconstruct_attention |
31 |
2 |
0 |
10 |
3 |
88% |
| transformer_lens/model_bridge/generalized_components/alibi_joint_qkv_attention.py |
ALiBiJointQKVAttentionBridge.get_random_inputs |
12 |
0 |
0 |
4 |
2 |
88% |
| transformer_lens/model_bridge/generalized_components/alibi_joint_qkv_attention.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/alibi_utils.py |
build_alibi_slopes |
10 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/alibi_utils.py |
build_alibi_tensor |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/alibi_utils.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.__init__ |
25 |
0 |
0 |
6 |
1 |
97% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.set_original_component |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.setup_hook_compatibility |
5 |
0 |
0 |
4 |
1 |
89% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.get_random_inputs |
26 |
18 |
0 |
18 |
5 |
30% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._setup_qkv_hook_reshaping |
40 |
6 |
0 |
26 |
5 |
80% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._setup_qkv_hook_reshaping.ReshapeForAttentionHeads.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._setup_qkv_hook_reshaping.ReshapeForAttentionHeads.handle_conversion |
5 |
0 |
0 |
4 |
1 |
89% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._setup_qkv_hook_reshaping.ReshapeForAttentionHeads.revert |
5 |
0 |
0 |
4 |
1 |
89% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._setup_qkv_hook_reshaping.TransposeRotaryHeads.handle_conversion |
3 |
1 |
0 |
2 |
1 |
60% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._setup_qkv_hook_reshaping.TransposeRotaryHeads.revert |
3 |
0 |
0 |
2 |
1 |
80% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._update_kv_cache |
9 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._reshape_qkv_to_heads |
16 |
1 |
0 |
6 |
1 |
91% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._apply_attn_dropout |
7 |
0 |
0 |
6 |
1 |
92% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._apply_output_projection |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._softmax_dropout_pattern |
10 |
0 |
0 |
6 |
1 |
94% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._reshape_attn_output |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._apply_reconstruct_attention_mask |
18 |
2 |
0 |
12 |
2 |
87% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._get_n_heads |
7 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._reshape_weight_to_3d |
7 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._project_per_head_qkv |
14 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._compute_per_head_result |
13 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.forward |
35 |
8 |
0 |
24 |
8 |
66% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.W_Q |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.W_K |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.W_V |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.W_O |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge._reshape_bias |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.b_Q |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.b_K |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.b_V |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
AttentionBridge.b_O |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/attention.py |
(no function) |
47 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/audio_feature_extractor.py |
AudioFeatureExtractorBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/audio_feature_extractor.py |
AudioFeatureExtractorBridge.forward |
8 |
7 |
0 |
4 |
0 |
8% |
| transformer_lens/model_bridge/generalized_components/audio_feature_extractor.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.__init__ |
18 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent._register_hook |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent._register_aliases |
35 |
10 |
0 |
20 |
1 |
73% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.get_hooks |
8 |
0 |
0 |
8 |
2 |
88% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent._is_getattr_called_internally |
4 |
3 |
0 |
4 |
0 |
12% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.set_original_component |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.original_component |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.add_hook |
5 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.remove_hooks |
8 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.set_processed_weights |
25 |
10 |
0 |
24 |
6 |
59% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.forward |
31 |
10 |
0 |
18 |
7 |
61% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.__getattr__ |
22 |
6 |
0 |
14 |
3 |
69% |
| transformer_lens/model_bridge/generalized_components/base.py |
GeneralizedComponent.__setattr__ |
21 |
2 |
0 |
12 |
0 |
94% |
| transformer_lens/model_bridge/generalized_components/base.py |
(no function) |
29 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/block.py |
BlockBridge.__init__ |
14 |
1 |
0 |
8 |
2 |
86% |
| transformer_lens/model_bridge/generalized_components/block.py |
BlockBridge.forward |
7 |
1 |
0 |
2 |
1 |
78% |
| transformer_lens/model_bridge/generalized_components/block.py |
BlockBridge._apply_output_hook |
11 |
0 |
0 |
8 |
2 |
89% |
| transformer_lens/model_bridge/generalized_components/block.py |
BlockBridge._check_stop_at_layer |
15 |
4 |
0 |
12 |
3 |
67% |
| transformer_lens/model_bridge/generalized_components/block.py |
BlockBridge._hook_input_hidden_states |
6 |
2 |
0 |
4 |
1 |
50% |
| transformer_lens/model_bridge/generalized_components/block.py |
BlockBridge._filter_kwargs_for_forward |
14 |
2 |
0 |
4 |
1 |
83% |
| transformer_lens/model_bridge/generalized_components/block.py |
MLABlockBridge.__init__ |
5 |
0 |
0 |
4 |
1 |
89% |
| transformer_lens/model_bridge/generalized_components/block.py |
ParallelBlockBridge.__init__ |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/model_bridge/generalized_components/block.py |
(no function) |
24 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/bloom_attention.py |
BloomAttentionBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/bloom_attention.py |
BloomAttentionBridge.forward |
22 |
21 |
0 |
10 |
0 |
3% |
| transformer_lens/model_bridge/generalized_components/bloom_attention.py |
BloomAttentionBridge._reconstruct_attention |
37 |
3 |
0 |
12 |
4 |
86% |
| transformer_lens/model_bridge/generalized_components/bloom_attention.py |
BloomAttentionBridge.set_processed_weights |
30 |
29 |
0 |
4 |
0 |
3% |
| transformer_lens/model_bridge/generalized_components/bloom_attention.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/bloom_block.py |
BloomBlockBridge.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/bloom_block.py |
BloomBlockBridge.build_alibi_tensor |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/bloom_block.py |
BloomBlockBridge.forward |
29 |
28 |
0 |
16 |
0 |
2% |
| transformer_lens/model_bridge/generalized_components/bloom_block.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/bloom_mlp.py |
BloomMLPBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/bloom_mlp.py |
BloomMLPBridge.forward |
10 |
9 |
0 |
6 |
0 |
6% |
| transformer_lens/model_bridge/generalized_components/bloom_mlp.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/clip_vision_encoder.py |
CLIPVisionEncoderLayerBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/clip_vision_encoder.py |
CLIPVisionEncoderLayerBridge.forward |
8 |
7 |
0 |
4 |
0 |
8% |
| transformer_lens/model_bridge/generalized_components/clip_vision_encoder.py |
CLIPVisionEncoderBridge.__init__ |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/model_bridge/generalized_components/clip_vision_encoder.py |
CLIPVisionEncoderBridge.forward |
10 |
9 |
0 |
6 |
0 |
6% |
| transformer_lens/model_bridge/generalized_components/clip_vision_encoder.py |
(no function) |
14 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/codegen_attention.py |
_rotate_every_two |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/codegen_attention.py |
_apply_rotary_pos_emb |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/codegen_attention.py |
CodeGenAttentionBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/codegen_attention.py |
CodeGenAttentionBridge.get_random_inputs |
11 |
10 |
0 |
4 |
0 |
7% |
| transformer_lens/model_bridge/generalized_components/codegen_attention.py |
CodeGenAttentionBridge.set_original_component |
3 |
0 |
0 |
2 |
1 |
80% |
| transformer_lens/model_bridge/generalized_components/codegen_attention.py |
CodeGenAttentionBridge.forward |
15 |
4 |
0 |
6 |
2 |
62% |
| transformer_lens/model_bridge/generalized_components/codegen_attention.py |
CodeGenAttentionBridge._reconstruct_attention |
35 |
3 |
0 |
6 |
3 |
85% |
| transformer_lens/model_bridge/generalized_components/codegen_attention.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/conv1d.py |
Conv1DBridge.forward |
6 |
5 |
0 |
2 |
0 |
12% |
| transformer_lens/model_bridge/generalized_components/conv1d.py |
Conv1DBridge.__repr__ |
8 |
7 |
0 |
2 |
0 |
10% |
| transformer_lens/model_bridge/generalized_components/conv1d.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/conv_pos_embed.py |
ConvPosEmbedBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/conv_pos_embed.py |
ConvPosEmbedBridge.forward |
8 |
7 |
0 |
4 |
0 |
8% |
| transformer_lens/model_bridge/generalized_components/conv_pos_embed.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/depthwise_conv1d.py |
DepthwiseConv1DBridge.forward |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/model_bridge/generalized_components/depthwise_conv1d.py |
DepthwiseConv1DBridge.__repr__ |
10 |
9 |
0 |
2 |
0 |
8% |
| transformer_lens/model_bridge/generalized_components/depthwise_conv1d.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/embedding.py |
EmbeddingBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/embedding.py |
EmbeddingBridge.W_E |
13 |
6 |
0 |
8 |
3 |
48% |
| transformer_lens/model_bridge/generalized_components/embedding.py |
EmbeddingBridge.forward |
20 |
6 |
0 |
8 |
4 |
64% |
| transformer_lens/model_bridge/generalized_components/embedding.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/gated_delta_net.py |
GatedDeltaNetBridge.__init__ |
12 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/gated_delta_net.py |
GatedDeltaNetBridge.forward |
5 |
1 |
0 |
4 |
2 |
67% |
| transformer_lens/model_bridge/generalized_components/gated_delta_net.py |
GatedDeltaNetBridge._native_forward |
14 |
6 |
0 |
10 |
3 |
46% |
| transformer_lens/model_bridge/generalized_components/gated_delta_net.py |
GatedDeltaNetBridge._hooked_forward |
55 |
54 |
0 |
12 |
0 |
1% |
| transformer_lens/model_bridge/generalized_components/gated_delta_net.py |
GatedDeltaNetBridge.compute_effective_attention |
27 |
26 |
0 |
6 |
0 |
3% |
| transformer_lens/model_bridge/generalized_components/gated_delta_net.py |
(no function) |
13 |
0 |
2 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/gated_mlp.py |
resolve_activation_fn |
16 |
0 |
0 |
14 |
1 |
97% |
| transformer_lens/model_bridge/generalized_components/gated_mlp.py |
resolve_activation_fn.gelu_tanh |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/gated_mlp.py |
GatedMLPBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/gated_mlp.py |
GatedMLPBridge.forward |
29 |
21 |
0 |
10 |
2 |
26% |
| transformer_lens/model_bridge/generalized_components/gated_mlp.py |
GatedMLPBridge.set_processed_weights |
45 |
13 |
0 |
24 |
9 |
59% |
| transformer_lens/model_bridge/generalized_components/gated_mlp.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/gated_rms_norm.py |
GatedRMSNormBridge.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/gated_rms_norm.py |
GatedRMSNormBridge.forward |
7 |
1 |
0 |
4 |
2 |
73% |
| transformer_lens/model_bridge/generalized_components/gated_rms_norm.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_gate_up_mlp.py |
JointGateUpMLPBridge.__init__ |
13 |
1 |
0 |
2 |
1 |
87% |
| transformer_lens/model_bridge/generalized_components/joint_gate_up_mlp.py |
JointGateUpMLPBridge._filter_gate_up_state_dict |
4 |
3 |
0 |
2 |
0 |
17% |
| transformer_lens/model_bridge/generalized_components/joint_gate_up_mlp.py |
JointGateUpMLPBridge._default_split_gate_up |
18 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_gate_up_mlp.py |
JointGateUpMLPBridge.set_original_component |
8 |
7 |
0 |
4 |
0 |
8% |
| transformer_lens/model_bridge/generalized_components/joint_gate_up_mlp.py |
JointGateUpMLPBridge._resolve_activation_fn |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/model_bridge/generalized_components/joint_gate_up_mlp.py |
JointGateUpMLPBridge.forward |
11 |
10 |
0 |
4 |
0 |
7% |
| transformer_lens/model_bridge/generalized_components/joint_gate_up_mlp.py |
(no function) |
16 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge.__init__ |
24 |
0 |
0 |
8 |
1 |
97% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge.__deepcopy__ |
16 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._filter_qkv_state_dict |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._create_qkv_conversion_rule |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._create_qkv_conversion_rule.ConditionalRearrangeConversion.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._create_qkv_conversion_rule.ConditionalRearrangeConversion.handle_conversion |
5 |
1 |
0 |
4 |
2 |
67% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._create_qkv_conversion_rule.ConditionalRearrangeConversion.revert |
5 |
1 |
0 |
4 |
2 |
67% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._default_split_qkv_matrix |
34 |
4 |
0 |
14 |
7 |
77% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge.set_original_component |
8 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge.forward |
9 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._is_split_qkv_fork_active |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._split_forward_qkv |
24 |
2 |
0 |
4 |
1 |
89% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._process_output |
19 |
2 |
0 |
14 |
7 |
73% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._apply_attention_input_hook |
9 |
2 |
0 |
6 |
2 |
73% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
JointQKVAttentionBridge._reconstruct_attention |
33 |
3 |
0 |
8 |
2 |
88% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_attention.py |
(no function) |
22 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_position_embeddings_attention.py |
JointQKVPositionEmbeddingsAttentionBridge.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_position_embeddings_attention.py |
JointQKVPositionEmbeddingsAttentionBridge.get_random_inputs |
21 |
20 |
0 |
6 |
0 |
4% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_position_embeddings_attention.py |
JointQKVPositionEmbeddingsAttentionBridge._apply_rotary_pos_emb |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_position_embeddings_attention.py |
JointQKVPositionEmbeddingsAttentionBridge._apply_rotary_pos_emb.rotate_half |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_position_embeddings_attention.py |
JointQKVPositionEmbeddingsAttentionBridge._reconstruct_attention |
28 |
3 |
0 |
6 |
2 |
85% |
| transformer_lens/model_bridge/generalized_components/joint_qkv_position_embeddings_attention.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/linear.py |
LinearBridge.forward |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/model_bridge/generalized_components/linear.py |
LinearBridge.__repr__ |
9 |
0 |
0 |
2 |
1 |
91% |
| transformer_lens/model_bridge/generalized_components/linear.py |
LinearBridge.set_processed_weights |
24 |
12 |
0 |
18 |
6 |
48% |
| transformer_lens/model_bridge/generalized_components/linear.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/mla_attention.py |
_rotate_half |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/mla_attention.py |
_apply_rotary_pos_emb |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/mla_attention.py |
MLAAttentionBridge.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/mla_attention.py |
MLAAttentionBridge.forward |
74 |
7 |
0 |
18 |
4 |
86% |
| transformer_lens/model_bridge/generalized_components/mla_attention.py |
MLAAttentionBridge.get_random_inputs |
30 |
29 |
0 |
14 |
0 |
2% |
| transformer_lens/model_bridge/generalized_components/mla_attention.py |
MLAAttentionBridge.__getattr__ |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/mla_attention.py |
(no function) |
16 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/mlp.py |
MLPBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/mlp.py |
MLPBridge.forward |
14 |
1 |
0 |
6 |
3 |
80% |
| transformer_lens/model_bridge/generalized_components/mlp.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/moe.py |
MoEBridge.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/moe.py |
MoEBridge.get_random_inputs |
6 |
5 |
0 |
4 |
0 |
10% |
| transformer_lens/model_bridge/generalized_components/moe.py |
MoEBridge.forward |
27 |
8 |
0 |
14 |
4 |
61% |
| transformer_lens/model_bridge/generalized_components/moe.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/mpt_alibi_attention.py |
_build_mpt_alibi_tensor |
10 |
1 |
0 |
2 |
1 |
83% |
| transformer_lens/model_bridge/generalized_components/mpt_alibi_attention.py |
MPTALiBiAttentionBridge.forward |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/model_bridge/generalized_components/mpt_alibi_attention.py |
MPTALiBiAttentionBridge.set_original_component |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/model_bridge/generalized_components/mpt_alibi_attention.py |
MPTALiBiAttentionBridge._reconstruct_attention |
22 |
3 |
0 |
6 |
3 |
79% |
| transformer_lens/model_bridge/generalized_components/mpt_alibi_attention.py |
MPTALiBiAttentionBridge.get_random_inputs |
10 |
9 |
0 |
4 |
0 |
7% |
| transformer_lens/model_bridge/generalized_components/mpt_alibi_attention.py |
(no function) |
18 |
2 |
0 |
0 |
0 |
89% |
| transformer_lens/model_bridge/generalized_components/normalization.py |
NormalizationBridge.__init__ |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/normalization.py |
NormalizationBridge.forward |
25 |
4 |
0 |
14 |
5 |
77% |
| transformer_lens/model_bridge/generalized_components/normalization.py |
NormalizationBridge._hf_autograd_forward_with_hooks |
27 |
5 |
0 |
12 |
4 |
77% |
| transformer_lens/model_bridge/generalized_components/normalization.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/pos_embed.py |
PosEmbedBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/pos_embed.py |
PosEmbedBridge.W_pos |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/model_bridge/generalized_components/pos_embed.py |
PosEmbedBridge.forward |
13 |
1 |
0 |
8 |
2 |
86% |
| transformer_lens/model_bridge/generalized_components/pos_embed.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/position_embedding_hooks_mixin.py |
PositionEmbeddingHooksMixin._init_position_embedding_hooks |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/position_embedding_hooks_mixin.py |
PositionEmbeddingHooksMixin.set_rotary_emb |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/position_embedding_hooks_mixin.py |
PositionEmbeddingHooksMixin._apply_position_embedding_hooks |
6 |
0 |
0 |
2 |
1 |
88% |
| transformer_lens/model_bridge/generalized_components/position_embedding_hooks_mixin.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
_setup_eager_attention_hook_wrapper |
11 |
2 |
0 |
2 |
0 |
85% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
_setup_eager_attention_hook_wrapper.hooked_eager_attention_forward |
8 |
7 |
0 |
6 |
0 |
7% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
PositionEmbeddingsAttentionBridge.__init__ |
9 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
PositionEmbeddingsAttentionBridge.set_original_component |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
PositionEmbeddingsAttentionBridge._validate_submodule_declarations |
6 |
1 |
0 |
6 |
2 |
75% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
PositionEmbeddingsAttentionBridge._decide_qk_norm_phase |
18 |
6 |
0 |
12 |
3 |
57% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
PositionEmbeddingsAttentionBridge._apply_pre_reshape_qk_norm |
6 |
5 |
0 |
2 |
0 |
12% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
PositionEmbeddingsAttentionBridge.forward |
135 |
16 |
0 |
50 |
10 |
84% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
PositionEmbeddingsAttentionBridge.get_random_inputs |
23 |
22 |
0 |
6 |
0 |
3% |
| transformer_lens/model_bridge/generalized_components/position_embeddings_attention.py |
(no function) |
22 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/rms_normalization.py |
RMSNormalizationBridge.__init__ |
3 |
0 |
0 |
2 |
1 |
80% |
| transformer_lens/model_bridge/generalized_components/rms_normalization.py |
(no function) |
6 |
0 |
2 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/rotary_embedding.py |
RotaryEmbeddingBridge.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/rotary_embedding.py |
RotaryEmbeddingBridge.get_random_inputs |
17 |
16 |
0 |
10 |
0 |
4% |
| transformer_lens/model_bridge/generalized_components/rotary_embedding.py |
RotaryEmbeddingBridge.forward |
16 |
4 |
0 |
10 |
4 |
62% |
| transformer_lens/model_bridge/generalized_components/rotary_embedding.py |
RotaryEmbeddingBridge.get_dummy_inputs |
5 |
4 |
0 |
2 |
0 |
14% |
| transformer_lens/model_bridge/generalized_components/rotary_embedding.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/siglip_vision_encoder.py |
SiglipVisionEncoderLayerBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/siglip_vision_encoder.py |
SiglipVisionEncoderLayerBridge.forward |
8 |
7 |
0 |
4 |
0 |
8% |
| transformer_lens/model_bridge/generalized_components/siglip_vision_encoder.py |
SiglipVisionEncoderBridge.__init__ |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/model_bridge/generalized_components/siglip_vision_encoder.py |
SiglipVisionEncoderBridge.forward |
10 |
9 |
0 |
6 |
0 |
6% |
| transformer_lens/model_bridge/generalized_components/siglip_vision_encoder.py |
(no function) |
14 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/ssm2_mixer.py |
SSM2MixerBridge.forward |
16 |
7 |
0 |
12 |
4 |
46% |
| transformer_lens/model_bridge/generalized_components/ssm2_mixer.py |
SSM2MixerBridge.compute_effective_attention |
46 |
1 |
0 |
6 |
1 |
96% |
| transformer_lens/model_bridge/generalized_components/ssm2_mixer.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/ssm_block.py |
SSMBlockBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/ssm_block.py |
SSMBlockBridge.forward |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/model_bridge/generalized_components/ssm_block.py |
SSMBlockBridge._apply_output_hook |
9 |
5 |
0 |
6 |
2 |
40% |
| transformer_lens/model_bridge/generalized_components/ssm_block.py |
SSMBlockBridge._hook_input_hidden_states |
6 |
2 |
0 |
4 |
1 |
50% |
| transformer_lens/model_bridge/generalized_components/ssm_block.py |
SSMBlockBridge._check_stop_at_layer |
17 |
5 |
0 |
12 |
3 |
66% |
| transformer_lens/model_bridge/generalized_components/ssm_block.py |
(no function) |
14 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/ssm_mixer.py |
SSMMixerBridge.forward |
16 |
7 |
0 |
12 |
4 |
46% |
| transformer_lens/model_bridge/generalized_components/ssm_mixer.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/symbolic.py |
SymbolicBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/symbolic.py |
SymbolicBridge.forward |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/symbolic.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/t5_block.py |
T5BlockBridge.__init__ |
8 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/t5_block.py |
T5BlockBridge.set_original_component |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/t5_block.py |
T5BlockBridge._patch_t5_block_forward |
5 |
1 |
0 |
2 |
1 |
71% |
| transformer_lens/model_bridge/generalized_components/t5_block.py |
T5BlockBridge._patch_t5_block_forward.patched_forward |
50 |
13 |
0 |
26 |
10 |
64% |
| transformer_lens/model_bridge/generalized_components/t5_block.py |
T5BlockBridge.forward |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/model_bridge/generalized_components/t5_block.py |
T5BlockBridge.get_expected_parameter_names |
5 |
4 |
0 |
2 |
0 |
14% |
| transformer_lens/model_bridge/generalized_components/t5_block.py |
T5BlockBridge.get_list_size |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/model_bridge/generalized_components/t5_block.py |
(no function) |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/unembedding.py |
UnembeddingBridge.__init__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/unembedding.py |
UnembeddingBridge.set_original_component |
6 |
0 |
0 |
2 |
1 |
88% |
| transformer_lens/model_bridge/generalized_components/unembedding.py |
UnembeddingBridge.W_U |
10 |
4 |
0 |
6 |
2 |
50% |
| transformer_lens/model_bridge/generalized_components/unembedding.py |
UnembeddingBridge.forward |
13 |
3 |
0 |
4 |
2 |
71% |
| transformer_lens/model_bridge/generalized_components/unembedding.py |
UnembeddingBridge.b_U |
17 |
10 |
0 |
8 |
3 |
40% |
| transformer_lens/model_bridge/generalized_components/unembedding.py |
(no function) |
12 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/vision_projection.py |
VisionProjectionBridge.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/generalized_components/vision_projection.py |
VisionProjectionBridge.forward |
8 |
7 |
0 |
4 |
0 |
8% |
| transformer_lens/model_bridge/generalized_components/vision_projection.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/get_params_util.py |
_get_n_kv_heads |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/get_params_util.py |
_reshape_kv_weight |
8 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/get_params_util.py |
_get_or_create_bias |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/get_params_util.py |
get_bridge_params |
87 |
1 |
0 |
20 |
3 |
96% |
| transformer_lens/model_bridge/get_params_util.py |
get_bridge_params._get_device_dtype |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/get_params_util.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/sources/__init__.py |
(no function) |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/sources/transformers.py |
map_default_transformer_lens_config |
111 |
25 |
0 |
92 |
20 |
74% |
| transformer_lens/model_bridge/sources/transformers.py |
determine_architecture_from_hf_config |
14 |
2 |
0 |
12 |
4 |
77% |
| transformer_lens/model_bridge/sources/transformers.py |
get_hf_model_class_for_architecture |
15 |
8 |
0 |
10 |
3 |
48% |
| transformer_lens/model_bridge/sources/transformers.py |
boot |
184 |
66 |
0 |
100 |
29 |
64% |
| transformer_lens/model_bridge/sources/transformers.py |
setup_tokenizer |
22 |
6 |
0 |
16 |
6 |
68% |
| transformer_lens/model_bridge/sources/transformers.py |
list_supported_models |
8 |
7 |
0 |
0 |
0 |
12% |
| transformer_lens/model_bridge/sources/transformers.py |
check_model_support |
12 |
11 |
0 |
2 |
0 |
7% |
| transformer_lens/model_bridge/sources/transformers.py |
(no function) |
25 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/__init__.py |
(no function) |
55 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/apertus.py |
ApertusArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/apertus.py |
ApertusArchitectureAdapter.prepare_loading |
16 |
15 |
0 |
4 |
0 |
5% |
| transformer_lens/model_bridge/supported_architectures/apertus.py |
ApertusArchitectureAdapter.prepare_loading._patched_init |
19 |
19 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/apertus.py |
ApertusArchitectureAdapter.prepare_loading._patched_forward |
4 |
4 |
0 |
2 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/apertus.py |
ApertusArchitectureAdapter._xielu_needs_patch |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/model_bridge/supported_architectures/apertus.py |
ApertusArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/apertus.py |
(no function) |
14 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
_BaichuanAttentionBridge._reconstruct_attention |
46 |
6 |
0 |
14 |
4 |
83% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
_patch_init_weights_for_baichuan |
14 |
10 |
0 |
10 |
1 |
29% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
_patch_init_weights_for_baichuan.safe_init_weights |
4 |
4 |
0 |
2 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
BaichuanArchitectureAdapter.__init__ |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
BaichuanArchitectureAdapter._split_baichuan_w_pack |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
BaichuanArchitectureAdapter._split_baichuan_w_pack._make_linear |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
BaichuanArchitectureAdapter.setup_component_testing |
10 |
9 |
0 |
6 |
0 |
6% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
BaichuanArchitectureAdapter.prepare_loading |
20 |
4 |
0 |
6 |
1 |
81% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
BaichuanArchitectureAdapter.prepare_model |
27 |
5 |
0 |
14 |
2 |
73% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
BaichuanArchitectureAdapter.preprocess_weights |
52 |
3 |
0 |
22 |
6 |
85% |
| transformer_lens/model_bridge/supported_architectures/baichuan.py |
(no function) |
20 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/bert.py |
BertArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/bert.py |
BertArchitectureAdapter.prepare_model |
4 |
3 |
0 |
2 |
0 |
17% |
| transformer_lens/model_bridge/supported_architectures/bert.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/bloom.py |
BloomArchitectureAdapter.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/bloom.py |
BloomArchitectureAdapter.split_qkv_matrix |
21 |
20 |
0 |
0 |
0 |
5% |
| transformer_lens/model_bridge/supported_architectures/bloom.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/codegen.py |
CodeGenArchitectureAdapter.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/codegen.py |
CodeGenArchitectureAdapter.split_qkv_matrix |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/codegen.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/cohere.py |
CohereArchitectureAdapter.__init__ |
22 |
1 |
0 |
4 |
2 |
88% |
| transformer_lens/model_bridge/supported_architectures/cohere.py |
CohereArchitectureAdapter.preprocess_weights |
7 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/cohere.py |
CohereArchitectureAdapter.setup_component_testing |
7 |
6 |
0 |
6 |
0 |
8% |
| transformer_lens/model_bridge/supported_architectures/cohere.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/deepseek_v3.py |
DeepSeekV3ArchitectureAdapter.__init__ |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/deepseek_v3.py |
DeepSeekV3ArchitectureAdapter.setup_component_testing |
7 |
6 |
0 |
6 |
0 |
8% |
| transformer_lens/model_bridge/supported_architectures/deepseek_v3.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
_patch_decoder_inplace_add |
7 |
1 |
0 |
2 |
1 |
78% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
_patch_decoder_inplace_add.patched_forward |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
FalconArchitectureAdapter.__init__ |
26 |
3 |
0 |
10 |
4 |
81% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
FalconArchitectureAdapter.prepare_model |
7 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
FalconArchitectureAdapter.prepare_model._make_patched_linear |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
FalconArchitectureAdapter.prepare_model._make_patched_linear.patched_forward |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
FalconArchitectureAdapter._split_falcon_qkv |
32 |
19 |
0 |
10 |
3 |
38% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
FalconArchitectureAdapter._split_falcon_qkv.build_linear |
5 |
1 |
0 |
2 |
1 |
71% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
FalconArchitectureAdapter.setup_component_testing |
9 |
8 |
0 |
8 |
0 |
6% |
| transformer_lens/model_bridge/supported_architectures/falcon.py |
(no function) |
12 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gemma1.py |
Gemma1ArchitectureAdapter.__init__ |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gemma1.py |
Gemma1ArchitectureAdapter.setup_hook_compatibility |
8 |
7 |
0 |
2 |
0 |
10% |
| transformer_lens/model_bridge/supported_architectures/gemma1.py |
Gemma1ArchitectureAdapter.setup_hook_compatibility.EmbeddingScaleConversion.__init__ |
2 |
2 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/gemma1.py |
Gemma1ArchitectureAdapter.setup_hook_compatibility.EmbeddingScaleConversion.handle_conversion |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/gemma1.py |
Gemma1ArchitectureAdapter.setup_hook_compatibility.EmbeddingScaleConversion.revert |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/gemma1.py |
Gemma1ArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/gemma1.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gemma2.py |
Gemma2ArchitectureAdapter.__init__ |
14 |
2 |
0 |
4 |
2 |
78% |
| transformer_lens/model_bridge/supported_architectures/gemma2.py |
Gemma2ArchitectureAdapter.setup_hook_compatibility |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/model_bridge/supported_architectures/gemma2.py |
Gemma2ArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/gemma2.py |
(no function) |
10 |
0 |
2 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gemma3.py |
Gemma3ArchitectureAdapter.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gemma3.py |
Gemma3ArchitectureAdapter.setup_hook_compatibility |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gemma3.py |
Gemma3ArchitectureAdapter.setup_component_testing |
19 |
18 |
0 |
20 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/gemma3.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gemma3_multimodal.py |
Gemma3MultimodalArchitectureAdapter.__init__ |
15 |
3 |
0 |
2 |
1 |
76% |
| transformer_lens/model_bridge/supported_architectures/gemma3_multimodal.py |
Gemma3MultimodalArchitectureAdapter.setup_hook_compatibility |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gemma3_multimodal.py |
Gemma3MultimodalArchitectureAdapter.setup_component_testing |
22 |
21 |
0 |
22 |
0 |
2% |
| transformer_lens/model_bridge/supported_architectures/gemma3_multimodal.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt2.py |
QKVSplitRearrangeConversion.__init__ |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt2.py |
QKVSplitRearrangeConversion._is_combined_qkv |
7 |
3 |
0 |
4 |
1 |
45% |
| transformer_lens/model_bridge/supported_architectures/gpt2.py |
QKVSplitRearrangeConversion.handle_conversion |
10 |
7 |
0 |
6 |
1 |
25% |
| transformer_lens/model_bridge/supported_architectures/gpt2.py |
QKVSplitRearrangeConversion.revert |
5 |
2 |
0 |
4 |
1 |
44% |
| transformer_lens/model_bridge/supported_architectures/gpt2.py |
GPT2ArchitectureAdapter.__init__ |
12 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt2.py |
(no function) |
14 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt2_lm_head_custom.py |
Gpt2LmHeadCustomArchitectureAdapter.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt2_lm_head_custom.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_bigcode.py |
MQAQKVConversionRule.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_bigcode.py |
MQAQKVConversionRule.handle_conversion |
7 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_bigcode.py |
MQAQKVConversionRule.revert |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_bigcode.py |
GPTBigCodeArchitectureAdapter.__init__ |
17 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_bigcode.py |
GPTBigCodeArchitectureAdapter._split_qkv_matrix |
13 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_bigcode.py |
GPTBigCodeArchitectureAdapter._split_qkv_matrix._make_linear |
5 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_bigcode.py |
(no function) |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_oss.py |
GPTOSSArchitectureAdapter.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_oss.py |
GPTOSSArchitectureAdapter.setup_hook_compatibility |
7 |
1 |
0 |
8 |
3 |
73% |
| transformer_lens/model_bridge/supported_architectures/gpt_oss.py |
GPTOSSArchitectureAdapter.setup_no_processing_hooks |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gpt_oss.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gptj.py |
GptjArchitectureAdapter.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/gptj.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite.py |
GraniteArchitectureAdapter.__init__ |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite.py |
GraniteArchitectureAdapter._setup_common_config |
12 |
0 |
0 |
2 |
1 |
93% |
| transformer_lens/model_bridge/supported_architectures/granite.py |
GraniteArchitectureAdapter._build_attention_bridge |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite.py |
GraniteArchitectureAdapter._build_mlp_bridge |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite.py |
GraniteArchitectureAdapter._build_component_mapping |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite.py |
GraniteArchitectureAdapter.setup_component_testing |
12 |
11 |
0 |
8 |
0 |
5% |
| transformer_lens/model_bridge/supported_architectures/granite.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite_moe.py |
GraniteMoeArchitectureAdapter._build_component_mapping |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite_moe.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite_moe_hybrid.py |
GraniteMoeHybridArchitectureAdapter.__init__ |
8 |
1 |
0 |
2 |
1 |
80% |
| transformer_lens/model_bridge/supported_architectures/granite_moe_hybrid.py |
GraniteMoeHybridArchitectureAdapter._build_mamba_bridge |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/granite_moe_hybrid.py |
GraniteMoeHybridArchitectureAdapter._build_component_mapping |
8 |
1 |
0 |
4 |
2 |
75% |
| transformer_lens/model_bridge/supported_architectures/granite_moe_hybrid.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/hubert.py |
HubertArchitectureAdapter.__init__ |
12 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/hubert.py |
HubertArchitectureAdapter._build_component_mapping |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/hubert.py |
HubertArchitectureAdapter.prepare_loading |
8 |
7 |
0 |
2 |
0 |
10% |
| transformer_lens/model_bridge/supported_architectures/hubert.py |
HubertArchitectureAdapter.prepare_model |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/model_bridge/supported_architectures/hubert.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
_InternLM2AttentionBridge._reconstruct_attention |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
_patch_init_weights_for_internlm2 |
11 |
10 |
0 |
6 |
0 |
6% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
_patch_init_weights_for_internlm2.safe_init_weights |
4 |
4 |
0 |
2 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
InternLM2ArchitectureAdapter.__init__ |
14 |
0 |
0 |
2 |
1 |
94% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
InternLM2ArchitectureAdapter._split_internlm2_wqkv |
23 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
InternLM2ArchitectureAdapter._split_internlm2_wqkv._make_linear |
5 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
InternLM2ArchitectureAdapter.setup_component_testing |
10 |
9 |
0 |
6 |
0 |
6% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
InternLM2ArchitectureAdapter.prepare_loading |
12 |
11 |
0 |
4 |
0 |
6% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
InternLM2ArchitectureAdapter.preprocess_weights |
60 |
3 |
0 |
22 |
4 |
89% |
| transformer_lens/model_bridge/supported_architectures/internlm2.py |
(no function) |
18 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/llama.py |
LlamaArchitectureAdapter.__init__ |
14 |
0 |
0 |
2 |
1 |
94% |
| transformer_lens/model_bridge/supported_architectures/llama.py |
LlamaArchitectureAdapter.setup_component_testing |
7 |
6 |
0 |
6 |
0 |
8% |
| transformer_lens/model_bridge/supported_architectures/llama.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/llava.py |
LlavaArchitectureAdapter.__init__ |
23 |
4 |
0 |
6 |
2 |
79% |
| transformer_lens/model_bridge/supported_architectures/llava.py |
LlavaArchitectureAdapter.setup_component_testing |
16 |
15 |
0 |
16 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/llava.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/llava_next.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/llava_onevision.py |
LlavaOnevisionArchitectureAdapter.prepare_model |
12 |
11 |
0 |
10 |
0 |
5% |
| transformer_lens/model_bridge/supported_architectures/llava_onevision.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mamba.py |
MambaArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mamba.py |
MambaArchitectureAdapter.create_stateful_cache |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mamba.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mamba2.py |
Mamba2ArchitectureAdapter.__init__ |
21 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mamba2.py |
Mamba2ArchitectureAdapter.create_stateful_cache |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mamba2.py |
compute_effective_attention |
12 |
2 |
0 |
8 |
2 |
80% |
| transformer_lens/model_bridge/supported_architectures/mamba2.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mingpt.py |
MingptArchitectureAdapter.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mingpt.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mistral.py |
MistralArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mistral.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mixtral.py |
MixtralArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mixtral.py |
MixtralArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/mixtral.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mpt.py |
MPTArchitectureAdapter.__init__ |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mpt.py |
MPTArchitectureAdapter._split_mpt_qkv |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mpt.py |
MPTArchitectureAdapter._split_mpt_qkv.make_linear |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/mpt.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/nanogpt.py |
NanogptArchitectureAdapter.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/nanogpt.py |
NanogptArchitectureAdapter.convert_weights |
6 |
5 |
0 |
4 |
0 |
10% |
| transformer_lens/model_bridge/supported_architectures/nanogpt.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/neel_solu_old.py |
NeelSoluOldArchitectureAdapter.__init__ |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/neel_solu_old.py |
convert_neel_solu_old_weights |
15 |
14 |
0 |
12 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/neel_solu_old.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/neo.py |
NeoLinearTransposeConversion.__init__ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/neo.py |
NeoLinearTransposeConversion.handle_conversion |
4 |
3 |
0 |
2 |
0 |
17% |
| transformer_lens/model_bridge/supported_architectures/neo.py |
NeoLinearTransposeConversion.revert |
6 |
5 |
0 |
2 |
0 |
12% |
| transformer_lens/model_bridge/supported_architectures/neo.py |
NeoArchitectureAdapter.__init__ |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/neo.py |
(no function) |
13 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/neox.py |
NeoxArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/neox.py |
NeoxArchitectureAdapter.split_qkv_matrix |
27 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/neox.py |
NeoxArchitectureAdapter.setup_component_testing |
5 |
4 |
0 |
6 |
0 |
9% |
| transformer_lens/model_bridge/supported_architectures/neox.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/olmo.py |
OlmoArchitectureAdapter.__init__ |
15 |
0 |
0 |
2 |
1 |
94% |
| transformer_lens/model_bridge/supported_architectures/olmo.py |
OlmoArchitectureAdapter.prepare_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/olmo.py |
OlmoArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/olmo.py |
_patch_olmo_inplace_clamp |
18 |
17 |
0 |
10 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/olmo.py |
_patch_olmo_inplace_clamp._make_patched_forward |
2 |
2 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/olmo.py |
_patch_olmo_inplace_clamp._make_patched_forward.patched_forward |
6 |
6 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/olmo.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/olmo2.py |
Olmo2ArchitectureAdapter.__init__ |
15 |
0 |
0 |
2 |
1 |
94% |
| transformer_lens/model_bridge/supported_architectures/olmo2.py |
Olmo2ArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/olmo2.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/olmo3.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/olmoe.py |
OlmoeArchitectureAdapter.__init__ |
15 |
0 |
0 |
2 |
1 |
94% |
| transformer_lens/model_bridge/supported_architectures/olmoe.py |
OlmoeArchitectureAdapter.prepare_model |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/model_bridge/supported_architectures/olmoe.py |
OlmoeArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/olmoe.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/openelm.py |
OpenElmArchitectureAdapter.__init__ |
16 |
0 |
0 |
2 |
1 |
94% |
| transformer_lens/model_bridge/supported_architectures/openelm.py |
OpenElmArchitectureAdapter.prepare_loading |
26 |
25 |
0 |
12 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/openelm.py |
OpenElmArchitectureAdapter.prepare_loading.safe_compute |
4 |
4 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/openelm.py |
OpenElmArchitectureAdapter.prepare_loading.safe_init_weights |
4 |
4 |
0 |
2 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/openelm.py |
OpenElmArchitectureAdapter.prepare_model |
23 |
22 |
0 |
14 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/openelm.py |
OpenElmArchitectureAdapter.setup_component_testing |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/openelm.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/opt.py |
OptArchitectureAdapter.__init__ |
19 |
3 |
0 |
6 |
3 |
76% |
| transformer_lens/model_bridge/supported_architectures/opt.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/phi.py |
PhiArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/phi.py |
PhiArchitectureAdapter.setup_component_testing |
16 |
15 |
0 |
14 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/phi.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
_SizedSplitConversion.__init__ |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
_SizedSplitConversion.handle_conversion |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
Phi3ArchitectureAdapter.__init__ |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
Phi3ArchitectureAdapter._split_gate_up |
17 |
16 |
0 |
6 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
Phi3ArchitectureAdapter._split_phi3_qkv |
24 |
23 |
0 |
8 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
Phi3ArchitectureAdapter.setup_component_testing |
7 |
6 |
0 |
6 |
0 |
8% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
Phi3ArchitectureAdapter.prepare_loading |
9 |
8 |
0 |
6 |
0 |
7% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
Phi3ArchitectureAdapter.preprocess_weights |
33 |
32 |
0 |
22 |
0 |
2% |
| transformer_lens/model_bridge/supported_architectures/phi3.py |
(no function) |
19 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/pythia.py |
PythiaArchitectureAdapter.__init__ |
6 |
5 |
0 |
0 |
0 |
17% |
| transformer_lens/model_bridge/supported_architectures/pythia.py |
PythiaArchitectureAdapter.split_qkv_matrix |
27 |
26 |
0 |
0 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/pythia.py |
PythiaArchitectureAdapter.setup_component_testing |
7 |
6 |
0 |
6 |
0 |
8% |
| transformer_lens/model_bridge/supported_architectures/pythia.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen.py |
QwenArchitectureAdapter.__init__ |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen.py |
QwenArchitectureAdapter._split_qkv_matrix |
27 |
26 |
0 |
8 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/qwen.py |
QwenArchitectureAdapter._split_qkv_matrix.build_linear |
4 |
4 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/qwen.py |
(no function) |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen2.py |
Qwen2ArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen2.py |
Qwen2ArchitectureAdapter.setup_component_testing |
7 |
6 |
0 |
6 |
0 |
8% |
| transformer_lens/model_bridge/supported_architectures/qwen2.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
Qwen3ArchitectureAdapter.__init__ |
7 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
Qwen3ArchitectureAdapter._setup_qwen3_config |
10 |
0 |
0 |
2 |
1 |
92% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
Qwen3ArchitectureAdapter._build_attention_bridge |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
Qwen3ArchitectureAdapter._build_mlp_bridge |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
Qwen3ArchitectureAdapter._build_linear_attn_bridge |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
Qwen3ArchitectureAdapter._build_component_mapping |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
Qwen3ArchitectureAdapter.setup_component_testing |
19 |
18 |
0 |
16 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
Qwen3ArchitectureAdapter._preprocess_gated_q_proj |
6 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3.py |
(no function) |
16 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3_5.py |
Qwen3_5ArchitectureAdapter.__init__ |
5 |
1 |
0 |
2 |
1 |
71% |
| transformer_lens/model_bridge/supported_architectures/qwen3_5.py |
Qwen3_5ArchitectureAdapter.prepare_loading |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/model_bridge/supported_architectures/qwen3_5.py |
Qwen3_5ArchitectureAdapter.preprocess_weights |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3_5.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3_moe.py |
Qwen3MoeArchitectureAdapter.__init__ |
13 |
0 |
0 |
2 |
1 |
93% |
| transformer_lens/model_bridge/supported_architectures/qwen3_moe.py |
Qwen3MoeArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/qwen3_moe.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3_next.py |
Qwen3NextArchitectureAdapter.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3_next.py |
Qwen3NextArchitectureAdapter._build_mlp_bridge |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3_next.py |
Qwen3NextArchitectureAdapter.preprocess_weights |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/qwen3_next.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/stablelm.py |
StableLmArchitectureAdapter.__init__ |
21 |
0 |
0 |
4 |
2 |
92% |
| transformer_lens/model_bridge/supported_architectures/stablelm.py |
StableLmArchitectureAdapter.setup_hook_compatibility |
24 |
23 |
0 |
10 |
0 |
3% |
| transformer_lens/model_bridge/supported_architectures/stablelm.py |
StableLmArchitectureAdapter.setup_hook_compatibility._make_hooked_forward |
2 |
2 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/stablelm.py |
StableLmArchitectureAdapter.setup_hook_compatibility._make_hooked_forward.hooked_forward |
2 |
2 |
0 |
0 |
0 |
0% |
| transformer_lens/model_bridge/supported_architectures/stablelm.py |
StableLmArchitectureAdapter.setup_component_testing |
13 |
12 |
0 |
14 |
0 |
4% |
| transformer_lens/model_bridge/supported_architectures/stablelm.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/t5.py |
T5ArchitectureAdapter.__init__ |
15 |
2 |
0 |
2 |
1 |
82% |
| transformer_lens/model_bridge/supported_architectures/t5.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/xglm.py |
XGLMArchitectureAdapter.__init__ |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/xglm.py |
XGLMArchitectureAdapter.setup_hook_compatibility |
7 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/xglm.py |
XGLMArchitectureAdapter.setup_hook_compatibility.EmbeddingScaleConversion.__init__ |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/xglm.py |
XGLMArchitectureAdapter.setup_hook_compatibility.EmbeddingScaleConversion.handle_conversion |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/xglm.py |
XGLMArchitectureAdapter.setup_hook_compatibility.EmbeddingScaleConversion.revert |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/supported_architectures/xglm.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/model_bridge/types.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/patching.py |
make_df_from_ranges |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/patching.py |
generic_activation_patch |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/patching.py |
generic_activation_patch |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/patching.py |
generic_activation_patch |
30 |
7 |
0 |
12 |
5 |
71% |
| transformer_lens/patching.py |
generic_activation_patch.patching_hook |
3 |
0 |
0 |
2 |
1 |
80% |
| transformer_lens/patching.py |
layer_pos_patch_setter |
4 |
4 |
0 |
0 |
0 |
0% |
| transformer_lens/patching.py |
layer_pos_head_vector_patch_setter |
4 |
4 |
0 |
0 |
0 |
0% |
| transformer_lens/patching.py |
layer_head_vector_patch_setter |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/patching.py |
layer_head_pattern_patch_setter |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/patching.py |
layer_head_pos_pattern_patch_setter |
4 |
4 |
0 |
0 |
0 |
0% |
| transformer_lens/patching.py |
layer_head_dest_src_pos_pattern_patch_setter |
4 |
4 |
0 |
0 |
0 |
0% |
| transformer_lens/patching.py |
get_act_patch_attn_head_all_pos_every |
9 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/patching.py |
get_act_patch_attn_head_by_pos_every |
10 |
9 |
0 |
0 |
0 |
10% |
| transformer_lens/patching.py |
get_act_patch_block_every |
5 |
4 |
0 |
0 |
0 |
20% |
| transformer_lens/patching.py |
(no function) |
58 |
0 |
32 |
0 |
0 |
100% |
| transformer_lens/pretrained/__init__.py |
(no function) |
0 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/__init__.py |
(no function) |
28 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/apertus.py |
convert_apertus_weights |
62 |
7 |
0 |
14 |
4 |
86% |
| transformer_lens/pretrained/weight_conversions/apertus.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/bert.py |
convert_bert_weights |
34 |
33 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/bert.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/bloom.py |
convert_bloom_weights |
37 |
36 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/bloom.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/coder.py |
convert_coder_weights |
39 |
38 |
0 |
2 |
0 |
2% |
| transformer_lens/pretrained/weight_conversions/coder.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/gemma.py |
convert_gemma_weights |
50 |
2 |
0 |
14 |
3 |
92% |
| transformer_lens/pretrained/weight_conversions/gemma.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/gpt2.py |
convert_gpt2_weights |
39 |
1 |
0 |
4 |
1 |
95% |
| transformer_lens/pretrained/weight_conversions/gpt2.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/gptj.py |
convert_gptj_weights |
32 |
31 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/gptj.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/hubert.py |
convert_hubert_weights |
66 |
11 |
0 |
36 |
12 |
77% |
| transformer_lens/pretrained/weight_conversions/hubert.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/llama.py |
convert_llama_weights |
40 |
39 |
0 |
8 |
0 |
2% |
| transformer_lens/pretrained/weight_conversions/llama.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/mingpt.py |
convert_mingpt_weights |
37 |
36 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/mingpt.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/mistral.py |
convert_mistral_weights |
32 |
31 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/mistral.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/mixtral.py |
convert_mixtral_weights |
33 |
32 |
0 |
4 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/mixtral.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/nanogpt.py |
convert_nanogpt_weights |
59 |
58 |
0 |
12 |
0 |
1% |
| transformer_lens/pretrained/weight_conversions/nanogpt.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/neel_solu_old.py |
convert_neel_solu_old_weights |
15 |
14 |
0 |
12 |
0 |
4% |
| transformer_lens/pretrained/weight_conversions/neel_solu_old.py |
(no function) |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/neo.py |
convert_neo_weights |
36 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/neo.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/neox.py |
convert_neox_weights |
30 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/neox.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/olmo.py |
convert_olmo_weights |
32 |
31 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/olmo.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/olmo2.py |
convert_olmo2_weights |
35 |
34 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/olmo2.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/olmo3.py |
convert_olmo3_weights |
37 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/olmo3.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/olmoe.py |
convert_olmoe_weights |
39 |
38 |
0 |
4 |
0 |
2% |
| transformer_lens/pretrained/weight_conversions/olmoe.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/openai.py |
convert_gpt_oss_weights |
45 |
4 |
0 |
8 |
2 |
89% |
| transformer_lens/pretrained/weight_conversions/openai.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/opt.py |
convert_opt_weights |
36 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/opt.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/phi.py |
convert_phi_weights |
38 |
37 |
0 |
2 |
0 |
2% |
| transformer_lens/pretrained/weight_conversions/phi.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/phi3.py |
convert_phi3_weights |
36 |
35 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/phi3.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/qwen.py |
convert_qwen_weights |
34 |
33 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/qwen.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/qwen2.py |
convert_qwen2_weights |
37 |
36 |
0 |
2 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/qwen2.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/qwen3.py |
convert_qwen3_weights |
38 |
37 |
0 |
4 |
0 |
2% |
| transformer_lens/pretrained/weight_conversions/qwen3.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/pretrained/weight_conversions/t5.py |
convert_t5_weights |
30 |
29 |
0 |
4 |
0 |
3% |
| transformer_lens/pretrained/weight_conversions/t5.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/supported_models.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/__init__.py |
(no function) |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/__init__.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/alias_drift.py |
DriftReport.has_drift |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/alias_drift.py |
DriftReport.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/alias_drift.py |
check_drift |
16 |
1 |
0 |
4 |
1 |
90% |
| transformer_lens/tools/model_registry/alias_drift.py |
print_report |
15 |
14 |
0 |
10 |
0 |
4% |
| transformer_lens/tools/model_registry/alias_drift.py |
main |
11 |
10 |
0 |
4 |
0 |
7% |
| transformer_lens/tools/model_registry/alias_drift.py |
(no function) |
17 |
1 |
0 |
2 |
1 |
89% |
| transformer_lens/tools/model_registry/api.py |
_load_json |
8 |
2 |
0 |
2 |
1 |
70% |
| transformer_lens/tools/model_registry/api.py |
_get_supported_models_report |
8 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/tools/model_registry/api.py |
_get_architecture_gaps_report |
8 |
0 |
0 |
2 |
1 |
90% |
| transformer_lens/tools/model_registry/api.py |
_get_verification_history |
8 |
7 |
0 |
2 |
0 |
10% |
| transformer_lens/tools/model_registry/api.py |
clear_cache |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/api.py |
get_supported_models |
7 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/tools/model_registry/api.py |
get_unsupported_architectures |
7 |
6 |
0 |
4 |
0 |
9% |
| transformer_lens/tools/model_registry/api.py |
is_model_supported |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/api.py |
get_model_architecture |
5 |
4 |
0 |
4 |
0 |
11% |
| transformer_lens/tools/model_registry/api.py |
get_architecture_models |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/api.py |
suggest_similar_model |
12 |
2 |
0 |
4 |
2 |
75% |
| transformer_lens/tools/model_registry/api.py |
suggest_similar_model.score_model |
13 |
3 |
0 |
12 |
4 |
72% |
| transformer_lens/tools/model_registry/api.py |
get_model_info |
6 |
1 |
0 |
4 |
1 |
80% |
| transformer_lens/tools/model_registry/api.py |
get_supported_architectures |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/api.py |
get_all_architectures_with_stats |
18 |
17 |
0 |
12 |
0 |
3% |
| transformer_lens/tools/model_registry/api.py |
is_architecture_supported |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/tools/model_registry/api.py |
get_registry_stats |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/api.py |
(no function) |
28 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/discover_architectures.py |
discover_architectures |
79 |
79 |
0 |
28 |
0 |
0% |
| transformer_lens/tools/model_registry/discover_architectures.py |
main |
5 |
5 |
0 |
0 |
0 |
0% |
| transformer_lens/tools/model_registry/discover_architectures.py |
(no function) |
16 |
16 |
0 |
2 |
0 |
0% |
| transformer_lens/tools/model_registry/exceptions.py |
ModelNotFoundError.__init__ |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/tools/model_registry/exceptions.py |
ArchitectureNotSupportedError.__init__ |
6 |
5 |
0 |
2 |
0 |
12% |
| transformer_lens/tools/model_registry/exceptions.py |
DataNotLoadedError.__init__ |
6 |
5 |
0 |
2 |
0 |
12% |
| transformer_lens/tools/model_registry/exceptions.py |
DataValidationError.__init__ |
8 |
7 |
0 |
4 |
0 |
8% |
| transformer_lens/tools/model_registry/exceptions.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/generate_report.py |
get_architecture_description |
18 |
18 |
0 |
12 |
0 |
0% |
| transformer_lens/tools/model_registry/generate_report.py |
generate_report |
70 |
70 |
0 |
14 |
0 |
0% |
| transformer_lens/tools/model_registry/generate_report.py |
main |
9 |
9 |
0 |
2 |
0 |
0% |
| transformer_lens/tools/model_registry/generate_report.py |
(no function) |
10 |
10 |
0 |
2 |
0 |
0% |
| transformer_lens/tools/model_registry/hf_scraper.py |
_extract_architecture |
6 |
6 |
0 |
4 |
0 |
0% |
| transformer_lens/tools/model_registry/hf_scraper.py |
_extract_param_count |
26 |
26 |
0 |
14 |
0 |
0% |
| transformer_lens/tools/model_registry/hf_scraper.py |
_load_existing_models |
15 |
15 |
0 |
6 |
0 |
0% |
| transformer_lens/tools/model_registry/hf_scraper.py |
_build_model_entry |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/tools/model_registry/hf_scraper.py |
scrape_all_models |
162 |
162 |
0 |
54 |
0 |
0% |
| transformer_lens/tools/model_registry/hf_scraper.py |
_save_checkpoint |
3 |
3 |
0 |
0 |
0 |
0% |
| transformer_lens/tools/model_registry/hf_scraper.py |
main |
10 |
10 |
0 |
0 |
0 |
0% |
| transformer_lens/tools/model_registry/hf_scraper.py |
(no function) |
20 |
20 |
0 |
2 |
0 |
0% |
| transformer_lens/tools/model_registry/registry_io.py |
is_quantized_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/registry_io.py |
load_supported_models_raw |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/registry_io.py |
save_supported_models_raw |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/registry_io.py |
load_verification_history |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/tools/model_registry/registry_io.py |
save_verification_history |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/registry_io.py |
_get_tl_version |
5 |
1 |
0 |
0 |
0 |
80% |
| transformer_lens/tools/model_registry/registry_io.py |
update_model_status |
41 |
5 |
0 |
28 |
2 |
84% |
| transformer_lens/tools/model_registry/registry_io.py |
add_verification_record |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/tools/model_registry/registry_io.py |
(no function) |
24 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/relevancy.py |
_normalize_demand |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/tools/model_registry/relevancy.py |
_normalize_usage |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/tools/model_registry/relevancy.py |
_score_benchmarkability |
13 |
0 |
0 |
12 |
0 |
100% |
| transformer_lens/tools/model_registry/relevancy.py |
compute_relevancy_score |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/relevancy.py |
compute_scores_for_gaps |
6 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/tools/model_registry/relevancy.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
ModelMetadata.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
ModelMetadata.from_dict |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
ModelEntry.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
ModelEntry.from_dict |
12 |
1 |
0 |
8 |
1 |
90% |
| transformer_lens/tools/model_registry/schemas.py |
ArchitectureGap.to_dict |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
ArchitectureGap.from_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
ScanInfo.to_dict |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/tools/model_registry/schemas.py |
ScanInfo.from_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
SupportedModelsReport.to_dict |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
SupportedModelsReport.from_dict |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/tools/model_registry/schemas.py |
ArchitectureGapsReport.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
ArchitectureGapsReport.from_dict |
5 |
0 |
0 |
2 |
1 |
86% |
| transformer_lens/tools/model_registry/schemas.py |
ArchitectureStats.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
ArchitectureAnalysis.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/schemas.py |
(no function) |
83 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/validate.py |
ValidationError.__str__ |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/tools/model_registry/validate.py |
ValidationResult.error_count |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/validate.py |
_validate_string |
9 |
3 |
0 |
8 |
2 |
59% |
| transformer_lens/tools/model_registry/validate.py |
_validate_int |
9 |
2 |
0 |
8 |
3 |
71% |
| transformer_lens/tools/model_registry/validate.py |
_validate_bool |
7 |
3 |
0 |
6 |
2 |
46% |
| transformer_lens/tools/model_registry/validate.py |
_validate_date_string |
11 |
3 |
0 |
6 |
2 |
71% |
| transformer_lens/tools/model_registry/validate.py |
_validate_datetime_string |
11 |
5 |
0 |
6 |
2 |
47% |
| transformer_lens/tools/model_registry/validate.py |
_validate_list |
7 |
3 |
0 |
6 |
2 |
46% |
| transformer_lens/tools/model_registry/validate.py |
_validate_model_metadata |
17 |
16 |
0 |
16 |
0 |
3% |
| transformer_lens/tools/model_registry/validate.py |
_validate_model_entry |
24 |
6 |
0 |
22 |
6 |
70% |
| transformer_lens/tools/model_registry/validate.py |
_validate_architecture_gap |
6 |
1 |
0 |
2 |
1 |
75% |
| transformer_lens/tools/model_registry/validate.py |
_validate_verification_record |
18 |
2 |
0 |
14 |
7 |
72% |
| transformer_lens/tools/model_registry/validate.py |
validate_supported_models_report |
13 |
1 |
0 |
6 |
2 |
84% |
| transformer_lens/tools/model_registry/validate.py |
validate_architecture_gaps_report |
12 |
1 |
0 |
6 |
2 |
83% |
| transformer_lens/tools/model_registry/validate.py |
validate_verification_history |
11 |
1 |
0 |
8 |
3 |
79% |
| transformer_lens/tools/model_registry/validate.py |
validate_json_schema |
19 |
8 |
0 |
14 |
2 |
52% |
| transformer_lens/tools/model_registry/validate.py |
validate_data_directory |
23 |
22 |
0 |
8 |
0 |
3% |
| transformer_lens/tools/model_registry/validate.py |
(no function) |
35 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verification.py |
VerificationRecord.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verification.py |
VerificationRecord.from_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verification.py |
VerificationHistory.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verification.py |
VerificationHistory.from_dict |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/tools/model_registry/verification.py |
VerificationHistory.get_record |
4 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/tools/model_registry/verification.py |
VerificationHistory.is_verified |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verification.py |
VerificationHistory.add_record |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verification.py |
VerificationHistory.invalidate |
7 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/tools/model_registry/verification.py |
(no function) |
27 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verify_models.py |
_handle_sigint |
6 |
6 |
0 |
2 |
0 |
0% |
| transformer_lens/tools/model_registry/verify_models.py |
_sanitize_note |
8 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/tools/model_registry/verify_models.py |
_get_current_model_status |
7 |
6 |
0 |
6 |
0 |
8% |
| transformer_lens/tools/model_registry/verify_models.py |
VerificationProgress.to_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verify_models.py |
VerificationProgress.from_dict |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/tools/model_registry/verify_models.py |
estimate_model_params |
41 |
40 |
0 |
14 |
0 |
2% |
| transformer_lens/tools/model_registry/verify_models.py |
estimate_benchmark_memory_gb |
17 |
16 |
0 |
10 |
0 |
4% |
| transformer_lens/tools/model_registry/verify_models.py |
get_available_memory_gb |
17 |
16 |
0 |
6 |
0 |
4% |
| transformer_lens/tools/model_registry/verify_models.py |
select_models_for_verification |
36 |
35 |
0 |
24 |
0 |
2% |
| transformer_lens/tools/model_registry/verify_models.py |
_extract_phase_scores |
15 |
14 |
0 |
14 |
0 |
3% |
| transformer_lens/tools/model_registry/verify_models.py |
_check_phase_scores |
26 |
25 |
0 |
18 |
0 |
2% |
| transformer_lens/tools/model_registry/verify_models.py |
_build_verified_note |
24 |
23 |
0 |
18 |
0 |
2% |
| transformer_lens/tools/model_registry/verify_models.py |
_clear_hf_cache |
15 |
14 |
0 |
8 |
0 |
4% |
| transformer_lens/tools/model_registry/verify_models.py |
_save_checkpoint |
3 |
2 |
0 |
0 |
0 |
33% |
| transformer_lens/tools/model_registry/verify_models.py |
_load_checkpoint |
8 |
7 |
0 |
2 |
0 |
10% |
| transformer_lens/tools/model_registry/verify_models.py |
verify_models |
235 |
234 |
0 |
128 |
0 |
1% |
| transformer_lens/tools/model_registry/verify_models.py |
_print_dry_run |
25 |
24 |
0 |
8 |
0 |
3% |
| transformer_lens/tools/model_registry/verify_models.py |
_print_summary |
22 |
21 |
0 |
14 |
0 |
3% |
| transformer_lens/tools/model_registry/verify_models.py |
main |
72 |
71 |
0 |
32 |
0 |
1% |
| transformer_lens/tools/model_registry/verify_models.py |
(no function) |
59 |
1 |
0 |
2 |
1 |
97% |
| transformer_lens/train.py |
train |
46 |
45 |
0 |
32 |
0 |
1% |
| transformer_lens/train.py |
(no function) |
29 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/__init__.py |
(no function) |
17 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/activation_functions.py |
gelu_new |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/activation_functions.py |
gelu_fast |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/activation_functions.py |
gelu_pytorch_tanh |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/activation_functions.py |
solu |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/activation_functions.py |
XIELU.__init__ |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/activation_functions.py |
XIELU.forward |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/activation_functions.py |
xielu |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/activation_functions.py |
(no function) |
16 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/addmm.py |
vanilla_addmm |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/addmm.py |
batch_addmm |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/addmm.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/aliases.py |
resolve_alias |
15 |
7 |
0 |
8 |
1 |
57% |
| transformer_lens/utilities/aliases.py |
resolve_alias._resolve_single_target |
14 |
1 |
0 |
10 |
1 |
92% |
| transformer_lens/utilities/aliases.py |
_collect_aliases_from_module |
21 |
0 |
0 |
18 |
0 |
100% |
| transformer_lens/utilities/aliases.py |
collect_aliases_recursive |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/aliases.py |
(no function) |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/architectures.py |
classify_architecture |
9 |
4 |
0 |
8 |
4 |
53% |
| transformer_lens/utilities/architectures.py |
get_architectures_for_config |
6 |
1 |
0 |
4 |
2 |
70% |
| transformer_lens/utilities/architectures.py |
classify_model_config |
7 |
2 |
0 |
6 |
2 |
69% |
| transformer_lens/utilities/architectures.py |
classify_model_name |
9 |
1 |
0 |
2 |
1 |
82% |
| transformer_lens/utilities/architectures.py |
is_masked_lm_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/architectures.py |
is_encoder_decoder_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/architectures.py |
is_multimodal_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/architectures.py |
is_audio_model |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/architectures.py |
(no function) |
14 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/attention.py |
simple_attn_linear |
7 |
2 |
0 |
4 |
2 |
64% |
| transformer_lens/utilities/attention.py |
complex_attn_linear |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/attention.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/attribute_utils.py |
get_nested_attr |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/utilities/attribute_utils.py |
set_nested_attr |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/utilities/attribute_utils.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/bridge_components.py |
collect_all_submodules_of_component |
8 |
1 |
0 |
8 |
2 |
81% |
| transformer_lens/utilities/bridge_components.py |
collect_components_of_block_bridge |
13 |
4 |
0 |
8 |
3 |
67% |
| transformer_lens/utilities/bridge_components.py |
collect_all_components |
6 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/utilities/bridge_components.py |
apply_fn_to_all_components |
6 |
0 |
0 |
4 |
1 |
90% |
| transformer_lens/utilities/bridge_components.py |
(no function) |
8 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/cache.py |
get_pos_offset |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/utilities/cache.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/components_utils.py |
get_act_name |
25 |
3 |
0 |
18 |
3 |
86% |
| transformer_lens/utilities/components_utils.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/defaults_utils.py |
override_or_use_default_value |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/defaults_utils.py |
LocallyOverridenDefaults.__init__ |
5 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/utilities/defaults_utils.py |
LocallyOverridenDefaults.__enter__ |
11 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/utilities/defaults_utils.py |
LocallyOverridenDefaults.__exit__ |
7 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/utilities/defaults_utils.py |
(no function) |
10 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/devices.py |
_torch_version_tuple |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/devices.py |
_torch_mps_has_known_broken_bug |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/devices.py |
get_device |
9 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/utilities/devices.py |
warn_if_mps |
14 |
0 |
0 |
12 |
0 |
100% |
| transformer_lens/utilities/devices.py |
ModelWithCfg.state_dict |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/utilities/devices.py |
ModelWithCfg.to |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/utilities/devices.py |
move_to_and_update_config |
19 |
0 |
0 |
16 |
3 |
91% |
| transformer_lens/utilities/devices.py |
(no function) |
21 |
0 |
3 |
0 |
0 |
100% |
| transformer_lens/utilities/exploratory_utils.py |
test_prompt |
38 |
2 |
0 |
16 |
5 |
87% |
| transformer_lens/utilities/exploratory_utils.py |
(no function) |
10 |
2 |
0 |
0 |
0 |
80% |
| transformer_lens/utilities/gpu_utils.py |
print_gpu_mem |
1 |
1 |
0 |
0 |
0 |
0% |
| transformer_lens/utilities/gpu_utils.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/hf_utils.py |
get_hf_token |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/hf_utils.py |
get_rotary_pct_from_config |
9 |
2 |
0 |
8 |
4 |
65% |
| transformer_lens/utilities/hf_utils.py |
select_compatible_kwargs |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/hf_utils.py |
download_file_from_hf |
7 |
2 |
0 |
4 |
1 |
73% |
| transformer_lens/utilities/hf_utils.py |
clear_huggingface_cache |
11 |
11 |
0 |
4 |
0 |
0% |
| transformer_lens/utilities/hf_utils.py |
clear_huggingface_cache.handle_remove_readonly |
14 |
14 |
0 |
8 |
0 |
0% |
| transformer_lens/utilities/hf_utils.py |
keep_single_column |
4 |
1 |
0 |
4 |
1 |
75% |
| transformer_lens/utilities/hf_utils.py |
get_dataset |
5 |
4 |
0 |
2 |
0 |
14% |
| transformer_lens/utilities/hf_utils.py |
(no function) |
21 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/initialization_utils.py |
calc_fan_in_and_fan_out |
14 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/utilities/initialization_utils.py |
init_xavier_uniform_ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/initialization_utils.py |
init_xavier_normal_ |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/initialization_utils.py |
init_kaiming_uniform_ |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/initialization_utils.py |
init_kaiming_normal_ |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/initialization_utils.py |
(no function) |
11 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/library_utils.py |
is_library_available |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/library_utils.py |
(no function) |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/lm_utils.py |
lm_cross_entropy_loss |
10 |
1 |
0 |
4 |
1 |
86% |
| transformer_lens/utilities/lm_utils.py |
lm_accuracy |
5 |
4 |
0 |
2 |
0 |
14% |
| transformer_lens/utilities/lm_utils.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/logits_utils.py |
_apply_repetition_penalty |
6 |
5 |
0 |
2 |
0 |
12% |
| transformer_lens/utilities/logits_utils.py |
sample_logits |
28 |
14 |
0 |
14 |
4 |
52% |
| transformer_lens/utilities/logits_utils.py |
(no function) |
6 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/matrix.py |
composition_scores |
14 |
0 |
0 |
6 |
1 |
95% |
| transformer_lens/utilities/matrix.py |
get_matrix_corner |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/utilities/matrix.py |
(no function) |
7 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/multi_gpu.py |
calculate_available_device_cuda_memory |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/multi_gpu.py |
determine_available_memory_for_available_devices |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/utilities/multi_gpu.py |
sort_devices_based_on_available_memory |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/multi_gpu.py |
get_best_available_cuda_device |
6 |
5 |
0 |
2 |
0 |
12% |
| transformer_lens/utilities/multi_gpu.py |
get_best_available_device |
5 |
1 |
0 |
2 |
1 |
71% |
| transformer_lens/utilities/multi_gpu.py |
get_device_for_block_index |
9 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/utilities/multi_gpu.py |
_validate_device_map_values |
6 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/utilities/multi_gpu.py |
resolve_device_map |
13 |
3 |
0 |
10 |
1 |
74% |
| transformer_lens/utilities/multi_gpu.py |
find_embedding_device |
19 |
4 |
0 |
8 |
0 |
85% |
| transformer_lens/utilities/multi_gpu.py |
count_unique_devices |
4 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/utilities/multi_gpu.py |
(no function) |
18 |
0 |
2 |
0 |
0 |
100% |
| transformer_lens/utilities/slice.py |
Slice.__init__ |
16 |
2 |
0 |
10 |
2 |
85% |
| transformer_lens/utilities/slice.py |
Slice.apply |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/slice.py |
Slice.indices |
5 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/utilities/slice.py |
Slice.__repr__ |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/slice.py |
Slice.unwrap |
5 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/utilities/slice.py |
(no function) |
15 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/tensors.py |
to_numpy |
10 |
3 |
0 |
8 |
1 |
67% |
| transformer_lens/utilities/tensors.py |
get_corner |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/tensors.py |
remove_batch_dim |
3 |
2 |
0 |
2 |
0 |
20% |
| transformer_lens/utilities/tensors.py |
transpose |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/tensors.py |
is_square |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/tensors.py |
is_lower_triangular |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/utilities/tensors.py |
check_structure |
33 |
32 |
0 |
22 |
0 |
2% |
| transformer_lens/utilities/tensors.py |
get_offset_position_ids |
3 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/tensors.py |
get_cumsum_along_dim |
6 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/utilities/tensors.py |
repeat_along_head_dimension |
4 |
0 |
0 |
2 |
1 |
83% |
| transformer_lens/utilities/tensors.py |
filter_dict_by_prefix |
2 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/tensors.py |
(no function) |
17 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utilities/tokenize_utils.py |
tokenize_and_concatenate |
17 |
0 |
0 |
6 |
2 |
91% |
| transformer_lens/utilities/tokenize_utils.py |
tokenize_and_concatenate.tokenize_function |
28 |
2 |
0 |
14 |
3 |
88% |
| transformer_lens/utilities/tokenize_utils.py |
get_tokenizer_with_bos |
12 |
0 |
0 |
6 |
0 |
100% |
| transformer_lens/utilities/tokenize_utils.py |
get_input_with_manually_prepended_bos |
4 |
1 |
0 |
2 |
1 |
67% |
| transformer_lens/utilities/tokenize_utils.py |
get_tokens_with_bos_removed |
11 |
0 |
0 |
4 |
0 |
100% |
| transformer_lens/utilities/tokenize_utils.py |
get_attention_mask |
13 |
1 |
0 |
6 |
1 |
89% |
| transformer_lens/utilities/tokenize_utils.py |
(no function) |
16 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/utils.py |
(no function) |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights._get_param_key |
3 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights._prepare_component_path |
8 |
1 |
0 |
4 |
2 |
75% |
| transformer_lens/weight_processing.py |
ProcessWeights._resolve_state_dict_key |
11 |
1 |
0 |
8 |
2 |
84% |
| transformer_lens/weight_processing.py |
ProcessWeights._safe_get_tensor |
2 |
1 |
0 |
0 |
0 |
50% |
| transformer_lens/weight_processing.py |
ProcessWeights.fold_layer_norm_bias_single |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.fold_layer_norm_weight_single |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.center_weight_single |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.fold_layer_norm_biases |
5 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.fold_layer_norm_biases._zero_bias |
1 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.fold_layer_norm_weights |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.center_attention_weights |
4 |
0 |
0 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.extract_attention_tensors_for_folding |
40 |
0 |
0 |
14 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.extract_attention_tensors_for_folding._reshape_bias_if_needed |
7 |
4 |
0 |
6 |
1 |
46% |
| transformer_lens/weight_processing.py |
ProcessWeights._fold_layer |
59 |
5 |
0 |
32 |
11 |
82% |
| transformer_lens/weight_processing.py |
ProcessWeights._fold_mlp_layer_norm |
126 |
27 |
0 |
72 |
22 |
71% |
| transformer_lens/weight_processing.py |
ProcessWeights._store_processed_attention_tensors |
20 |
2 |
0 |
10 |
5 |
77% |
| transformer_lens/weight_processing.py |
ProcessWeights._fold_unembed_layer_norm |
37 |
6 |
0 |
18 |
7 |
76% |
| transformer_lens/weight_processing.py |
ProcessWeights._fold_final_rms_bias |
26 |
4 |
0 |
12 |
5 |
76% |
| transformer_lens/weight_processing.py |
ProcessWeights.fold_layer_norm |
7 |
0 |
0 |
2 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.center_writing_weights |
93 |
32 |
0 |
44 |
13 |
66% |
| transformer_lens/weight_processing.py |
ProcessWeights.center_unembed |
21 |
1 |
0 |
10 |
3 |
87% |
| transformer_lens/weight_processing.py |
ProcessWeights.fold_value_biases |
88 |
53 |
0 |
38 |
7 |
40% |
| transformer_lens/weight_processing.py |
ProcessWeights.process_weights |
35 |
6 |
0 |
34 |
6 |
80% |
| transformer_lens/weight_processing.py |
ProcessWeights.refactor_factored_attn_matrices |
51 |
0 |
0 |
8 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.convert_tensor_to_tl_format |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.convert_tensor_to_tl_format |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.convert_tensor_to_tl_format |
36 |
9 |
0 |
28 |
5 |
72% |
| transformer_lens/weight_processing.py |
ProcessWeights.convert_tensor_to_hf_format |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.convert_tensor_to_hf_format |
0 |
0 |
1 |
0 |
0 |
100% |
| transformer_lens/weight_processing.py |
ProcessWeights.convert_tensor_to_hf_format |
32 |
9 |
0 |
24 |
5 |
68% |
| transformer_lens/weight_processing.py |
ProcessWeights.distribute_weights_to_components |
43 |
22 |
0 |
30 |
8 |
51% |
| transformer_lens/weight_processing.py |
(no function) |
60 |
0 |
42 |
0 |
0 |
100% |