albertvillanova HF Staff commited on
Commit
cb570ef
·
verified ·
1 Parent(s): b0b2701

Upload Qwen2_5_VLForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +10 -8
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -7,22 +7,21 @@
7
  "dtype": "bfloat16",
8
  "eos_token_id": 151645,
9
  "hidden_act": "silu",
10
- "hidden_size": 2048,
11
  "image_token_id": 151655,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 11008,
14
  "max_position_embeddings": 128000,
15
  "max_window_layers": 70,
16
  "model_type": "qwen2_5_vl",
17
- "num_attention_heads": 16,
18
- "num_hidden_layers": 36,
19
  "num_key_value_heads": 2,
20
  "rms_norm_eps": 1e-06,
21
  "rope_scaling": {
22
  "mrope_section": [
23
- 16,
24
- 24,
25
- 24
26
  ],
27
  "rope_type": "default",
28
  "type": "default"
@@ -30,7 +29,6 @@
30
  "rope_theta": 1000000.0,
31
  "sliding_window": 32768,
32
  "text_config": {
33
- "_name_or_path": "Qwen/Qwen2.5-VL-3B-Instruct",
34
  "architectures": [
35
  "Qwen2_5_VLForConditionalGeneration"
36
  ],
@@ -40,6 +38,7 @@
40
  "eos_token_id": 151645,
41
  "hidden_act": "silu",
42
  "hidden_size": 16,
 
43
  "initializer_range": 0.02,
44
  "intermediate_size": 11008,
45
  "layer_types": [
@@ -66,10 +65,13 @@
66
  "tie_word_embeddings": true,
67
  "use_cache": true,
68
  "use_sliding_window": false,
 
 
 
69
  "vision_token_id": 151654,
70
  "vocab_size": 151936
71
  },
72
- "transformers_version": "4.57.1",
73
  "use_cache": true,
74
  "use_sliding_window": false,
75
  "video_token_id": 151656,
 
7
  "dtype": "bfloat16",
8
  "eos_token_id": 151645,
9
  "hidden_act": "silu",
10
+ "hidden_size": 16,
11
  "image_token_id": 151655,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 11008,
14
  "max_position_embeddings": 128000,
15
  "max_window_layers": 70,
16
  "model_type": "qwen2_5_vl",
17
+ "num_attention_heads": 4,
18
+ "num_hidden_layers": 2,
19
  "num_key_value_heads": 2,
20
  "rms_norm_eps": 1e-06,
21
  "rope_scaling": {
22
  "mrope_section": [
23
+ 1,
24
+ 1
 
25
  ],
26
  "rope_type": "default",
27
  "type": "default"
 
29
  "rope_theta": 1000000.0,
30
  "sliding_window": 32768,
31
  "text_config": {
 
32
  "architectures": [
33
  "Qwen2_5_VLForConditionalGeneration"
34
  ],
 
38
  "eos_token_id": 151645,
39
  "hidden_act": "silu",
40
  "hidden_size": 16,
41
+ "image_token_id": null,
42
  "initializer_range": 0.02,
43
  "intermediate_size": 11008,
44
  "layer_types": [
 
65
  "tie_word_embeddings": true,
66
  "use_cache": true,
67
  "use_sliding_window": false,
68
+ "video_token_id": null,
69
+ "vision_end_token_id": 151653,
70
+ "vision_start_token_id": 151652,
71
  "vision_token_id": 151654,
72
  "vocab_size": 151936
73
  },
74
+ "transformers_version": "4.56.1",
75
  "use_cache": true,
76
  "use_sliding_window": false,
77
  "video_token_id": 151656,
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 151643,
4
  "eos_token_id": 151645,
5
- "transformers_version": "4.57.1"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 151643,
4
  "eos_token_id": 151645,
5
+ "transformers_version": "4.56.1"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9ee6de012cbabb328744133306ae85256f9d1d994145c50a0a341918c7af35c
3
  size 18086192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1adaecbcf061f2bbc1cc4e00dcfdd13424e9d25abb3c9d6b014f423e3b6d1bc8
3
  size 18086192