|
24 | 24 |
|
25 | 25 |
|
26 | 26 | TENSOR_NAMES = loading_utils.ModelLoader.TensorNames(
|
27 |
| - ff_up_proj="vision_tower.vision_model.encoder.layers.{}.mlp.fc1", |
28 |
| - ff_down_proj="vision_tower.vision_model.encoder.layers.{}.mlp.fc2", |
| 27 | + ff_up_proj="siglip_vision_model.encoder_blocks.{}.mlp.fc1", |
| 28 | + ff_down_proj="siglip_vision_model.encoder_blocks.{}.mlp.fc2", |
29 | 29 | attn_query_proj=(
|
30 |
| - "vision_tower.vision_model.encoder.layers.{}.self_attn.q_proj" |
| 30 | + "siglip_vision_model.encoder_blocks.{}.self_attn.q_proj" |
31 | 31 | ),
|
32 | 32 | attn_key_proj=(
|
33 |
| - "vision_tower.vision_model.encoder.layers.{}.self_attn.k_proj" |
| 33 | + "siglip_vision_model.encoder_blocks.{}.self_attn.k_proj" |
34 | 34 | ),
|
35 | 35 | attn_value_proj=(
|
36 |
| - "vision_tower.vision_model.encoder.layers.{}.self_attn.v_proj" |
| 36 | + "siglip_vision_model.encoder_blocks.{}.self_attn.v_proj" |
37 | 37 | ),
|
38 | 38 | attn_output_proj=(
|
39 |
| - "vision_tower.vision_model.encoder.layers.{}.self_attn.out_proj" |
| 39 | + "siglip_vision_model.encoder_blocks.{}.self_attn.o_proj" |
40 | 40 | ),
|
41 |
| - pre_attn_norm="vision_tower.vision_model.encoder.layers.{}.layer_norm1", |
42 |
| - embedding="vision_tower.vision_model.embeddings.patch_embedding", |
| 41 | + pre_attn_norm="siglip_vision_model.encoder_blocks.{}.layer_norm1", |
| 42 | + pre_ff_norm="siglip_vision_model.encoder_blocks.{}.layer_norm2", |
| 43 | + embedding="siglip_vision_model.patch_embedding", |
43 | 44 | embedding_position=(
|
44 |
| - "vision_tower.vision_model.embeddings.position_embedding.weight" |
| 45 | + "siglip_vision_model.position_embedding.weight" |
45 | 46 | ),
|
46 |
| - final_norm="vision_tower.vision_model.post_layernorm", |
| 47 | + final_norm="siglip_vision_model.final_norm", |
47 | 48 | )
|
48 | 49 |
|
49 | 50 |
|
|
0 commit comments