dacorvo HF Staff commited on
Commit
497b238
·
verified ·
1 Parent(s): 275ba81

Synchronizing local compiler cache.

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +18 -0
  2. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.3.dev1/llama4_text/meta-llama/Llama-4-Maverick-17B-128E-Instruct/0782cbf4b3ed6eea7218.json +190 -0
  3. neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.3.dev1/llama4_text/meta-llama/Llama-4-Maverick-17B-128E-Instruct/9fc13314975962cba5a6.json +190 -0
  4. neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/compile_flags.json +1 -0
  5. neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/model.done +0 -0
  6. neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/model.hlo_module.pb +3 -0
  7. neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/model.neff +3 -0
  8. neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/wrapped_neff.hlo +3 -0
  9. neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/compile_flags.json +1 -0
  10. neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/model.done +0 -0
  11. neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/model.hlo_module.pb +3 -0
  12. neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/model.neff +3 -0
  13. neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/compile_flags.json +1 -0
  14. neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/model.done +0 -0
  15. neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/model.hlo_module.pb +3 -0
  16. neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/model.neff +3 -0
  17. neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/wrapped_neff.hlo +3 -0
  18. neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/compile_flags.json +1 -0
  19. neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/model.done +0 -0
  20. neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/model.hlo_module.pb +3 -0
  21. neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/model.neff +3 -0
  22. neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/compile_flags.json +1 -0
  23. neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/model.done +0 -0
  24. neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/model.hlo_module.pb +3 -0
  25. neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/model.neff +3 -0
  26. neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/compile_flags.json +1 -0
  27. neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/model.done +0 -0
  28. neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/model.hlo_module.pb +3 -0
  29. neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/model.neff +3 -0
  30. neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/compile_flags.json +1 -0
  31. neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/model.done +0 -0
  32. neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/model.hlo_module.pb +3 -0
  33. neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/model.neff +3 -0
  34. neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/wrapped_neff.hlo +3 -0
  35. neuronxcc-2.21.18209.0+043b1bf7/MODULE_674332a679782ee04898+a32116a7/compile_flags.json +1 -0
  36. neuronxcc-2.21.18209.0+043b1bf7/MODULE_674332a679782ee04898+a32116a7/model.hlo_module.pb +3 -0
  37. neuronxcc-2.21.18209.0+043b1bf7/MODULE_674332a679782ee04898+a32116a7/model.log +3 -0
  38. neuronxcc-2.21.18209.0+043b1bf7/MODULE_8325de44f605a823c436+a32116a7/compile_flags.json +1 -0
  39. neuronxcc-2.21.18209.0+043b1bf7/MODULE_8325de44f605a823c436+a32116a7/model.hlo_module.pb +3 -0
  40. neuronxcc-2.21.18209.0+043b1bf7/MODULE_8325de44f605a823c436+a32116a7/model.log +3 -0
  41. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/compile_flags.json +1 -0
  42. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/model.done +0 -0
  43. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/model.hlo_module.pb +3 -0
  44. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/model.neff +3 -0
  45. neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/wrapped_neff.hlo +3 -0
  46. neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/compile_flags.json +1 -0
  47. neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/model.done +0 -0
  48. neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/model.hlo_module.pb +3 -0
  49. neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/model.neff +3 -0
  50. neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/wrapped_neff.hlo +3 -0
.gitattributes CHANGED
@@ -13312,3 +13312,21 @@ neuronxcc-2.21.18209.0+043b1bf7/MODULE_e6150818c39bca2677d1+8eeafb64/model.neff
13312
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_e6150818c39bca2677d1+8eeafb64/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13313
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_f2f5d7eea0d7ac74facb+26bb721a/model.neff filter=lfs diff=lfs merge=lfs -text
13314
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_f2f5d7eea0d7ac74facb+26bb721a/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13312
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_e6150818c39bca2677d1+8eeafb64/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13313
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_f2f5d7eea0d7ac74facb+26bb721a/model.neff filter=lfs diff=lfs merge=lfs -text
13314
  neuronxcc-2.21.18209.0+043b1bf7/MODULE_f2f5d7eea0d7ac74facb+26bb721a/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13315
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/model.neff filter=lfs diff=lfs merge=lfs -text
13316
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13317
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/model.neff filter=lfs diff=lfs merge=lfs -text
13318
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/model.neff filter=lfs diff=lfs merge=lfs -text
13319
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13320
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/model.neff filter=lfs diff=lfs merge=lfs -text
13321
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/model.neff filter=lfs diff=lfs merge=lfs -text
13322
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/model.neff filter=lfs diff=lfs merge=lfs -text
13323
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/model.neff filter=lfs diff=lfs merge=lfs -text
13324
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13325
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/model.neff filter=lfs diff=lfs merge=lfs -text
13326
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13327
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/model.neff filter=lfs diff=lfs merge=lfs -text
13328
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13329
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_ec5bfc4ef4ea5c5e3798+ac10809c/model.neff filter=lfs diff=lfs merge=lfs -text
13330
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_ec5bfc4ef4ea5c5e3798+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
13331
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_f850227f2a268ed6719b+ac10809c/model.neff filter=lfs diff=lfs merge=lfs -text
13332
+ neuronxcc-2.21.18209.0+043b1bf7/MODULE_f850227f2a268ed6719b+ac10809c/wrapped_neff.hlo filter=lfs diff=lfs merge=lfs -text
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.3.dev1/llama4_text/meta-llama/Llama-4-Maverick-17B-128E-Instruct/0782cbf4b3ed6eea7218.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "meta-llama/Llama-4-Maverick-17B-128E-Instruct",
4
+ "_task": "text-generation",
5
+ "attention_bias": false,
6
+ "attention_chunk_size": 8192,
7
+ "attention_dropout": 0.0,
8
+ "attn_scale": 0.1,
9
+ "attn_temperature_tuning": true,
10
+ "dtype": "bfloat16",
11
+ "floor_scale": 8192,
12
+ "for_llm_compressor": false,
13
+ "head_dim": 128,
14
+ "hidden_act": "silu",
15
+ "hidden_size": 5120,
16
+ "initializer_range": 0.02,
17
+ "interleave_moe_layer_step": 2,
18
+ "intermediate_size": 8192,
19
+ "intermediate_size_mlp": 16384,
20
+ "layer_types": [
21
+ "chunked_attention",
22
+ "chunked_attention",
23
+ "chunked_attention",
24
+ "full_attention",
25
+ "chunked_attention",
26
+ "chunked_attention",
27
+ "chunked_attention",
28
+ "full_attention",
29
+ "chunked_attention",
30
+ "chunked_attention",
31
+ "chunked_attention",
32
+ "full_attention",
33
+ "chunked_attention",
34
+ "chunked_attention",
35
+ "chunked_attention",
36
+ "full_attention",
37
+ "chunked_attention",
38
+ "chunked_attention",
39
+ "chunked_attention",
40
+ "full_attention",
41
+ "chunked_attention",
42
+ "chunked_attention",
43
+ "chunked_attention",
44
+ "full_attention",
45
+ "chunked_attention",
46
+ "chunked_attention",
47
+ "chunked_attention",
48
+ "full_attention",
49
+ "chunked_attention",
50
+ "chunked_attention",
51
+ "chunked_attention",
52
+ "full_attention",
53
+ "chunked_attention",
54
+ "chunked_attention",
55
+ "chunked_attention",
56
+ "full_attention",
57
+ "chunked_attention",
58
+ "chunked_attention",
59
+ "chunked_attention",
60
+ "full_attention",
61
+ "chunked_attention",
62
+ "chunked_attention",
63
+ "chunked_attention",
64
+ "full_attention",
65
+ "chunked_attention",
66
+ "chunked_attention",
67
+ "chunked_attention",
68
+ "full_attention"
69
+ ],
70
+ "max_position_embeddings": 1048576,
71
+ "model_type": "llama4_text",
72
+ "moe_layers": [
73
+ 1,
74
+ 3,
75
+ 5,
76
+ 7,
77
+ 9,
78
+ 11,
79
+ 13,
80
+ 15,
81
+ 17,
82
+ 19,
83
+ 21,
84
+ 23,
85
+ 25,
86
+ 27,
87
+ 29,
88
+ 31,
89
+ 33,
90
+ 35,
91
+ 37,
92
+ 39,
93
+ 41,
94
+ 43,
95
+ 45,
96
+ 47
97
+ ],
98
+ "neuron": {
99
+ "_serialized_key": "NxDNeuronConfig",
100
+ "batch_size": 1,
101
+ "capacity_factor": null,
102
+ "checkpoint_id": "meta-llama/Llama-4-Maverick-17B-128E-Instruct",
103
+ "checkpoint_revision": "73d14711bcc77c16df3470856949c3764056b617",
104
+ "continuous_batching": false,
105
+ "ep_degree": 1,
106
+ "fused_qkv": false,
107
+ "glu_mlp": true,
108
+ "local_ranks_size": 64,
109
+ "max_batch_size": 1,
110
+ "max_context_length": 4096,
111
+ "max_topk": 256,
112
+ "n_active_tokens": 4096,
113
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
114
+ "on_device_sampling": true,
115
+ "optimum_neuron_version": "0.4.3.dev1",
116
+ "output_logits": false,
117
+ "pp_degree": 1,
118
+ "sequence_length": 4096,
119
+ "speculation_length": 0,
120
+ "start_rank_id": 0,
121
+ "target": "trn2",
122
+ "torch_dtype": "bfloat16",
123
+ "tp_degree": 64
124
+ },
125
+ "no_rope_layers": [
126
+ 1,
127
+ 1,
128
+ 1,
129
+ 0,
130
+ 1,
131
+ 1,
132
+ 1,
133
+ 0,
134
+ 1,
135
+ 1,
136
+ 1,
137
+ 0,
138
+ 1,
139
+ 1,
140
+ 1,
141
+ 0,
142
+ 1,
143
+ 1,
144
+ 1,
145
+ 0,
146
+ 1,
147
+ 1,
148
+ 1,
149
+ 0,
150
+ 1,
151
+ 1,
152
+ 1,
153
+ 0,
154
+ 1,
155
+ 1,
156
+ 1,
157
+ 0,
158
+ 1,
159
+ 1,
160
+ 1,
161
+ 0,
162
+ 1,
163
+ 1,
164
+ 1,
165
+ 0,
166
+ 1,
167
+ 1,
168
+ 1,
169
+ 0,
170
+ 1,
171
+ 1,
172
+ 1,
173
+ 0
174
+ ],
175
+ "num_attention_heads": 40,
176
+ "num_experts_per_tok": 1,
177
+ "num_hidden_layers": 48,
178
+ "num_key_value_heads": 8,
179
+ "num_local_experts": 128,
180
+ "output_router_logits": false,
181
+ "rms_norm_eps": 1e-05,
182
+ "rope_scaling": null,
183
+ "rope_theta": 500000.0,
184
+ "router_aux_loss_coef": 0.001,
185
+ "router_jitter_noise": 0.0,
186
+ "tie_word_embeddings": false,
187
+ "use_cache": true,
188
+ "use_qk_norm": false,
189
+ "vocab_size": 202048
190
+ }
neuronxcc-2.21.18209.0+043b1bf7/0_REGISTRY/0.4.3.dev1/llama4_text/meta-llama/Llama-4-Maverick-17B-128E-Instruct/9fc13314975962cba5a6.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_entry_class": "SingleModelCacheEntry",
3
+ "_model_id": "meta-llama/Llama-4-Maverick-17B-128E-Instruct",
4
+ "_task": "text-generation",
5
+ "attention_bias": false,
6
+ "attention_chunk_size": 8192,
7
+ "attention_dropout": 0.0,
8
+ "attn_scale": 0.1,
9
+ "attn_temperature_tuning": true,
10
+ "dtype": "bfloat16",
11
+ "floor_scale": 8192,
12
+ "for_llm_compressor": false,
13
+ "head_dim": 128,
14
+ "hidden_act": "silu",
15
+ "hidden_size": 5120,
16
+ "initializer_range": 0.02,
17
+ "interleave_moe_layer_step": 2,
18
+ "intermediate_size": 8192,
19
+ "intermediate_size_mlp": 16384,
20
+ "layer_types": [
21
+ "chunked_attention",
22
+ "chunked_attention",
23
+ "chunked_attention",
24
+ "full_attention",
25
+ "chunked_attention",
26
+ "chunked_attention",
27
+ "chunked_attention",
28
+ "full_attention",
29
+ "chunked_attention",
30
+ "chunked_attention",
31
+ "chunked_attention",
32
+ "full_attention",
33
+ "chunked_attention",
34
+ "chunked_attention",
35
+ "chunked_attention",
36
+ "full_attention",
37
+ "chunked_attention",
38
+ "chunked_attention",
39
+ "chunked_attention",
40
+ "full_attention",
41
+ "chunked_attention",
42
+ "chunked_attention",
43
+ "chunked_attention",
44
+ "full_attention",
45
+ "chunked_attention",
46
+ "chunked_attention",
47
+ "chunked_attention",
48
+ "full_attention",
49
+ "chunked_attention",
50
+ "chunked_attention",
51
+ "chunked_attention",
52
+ "full_attention",
53
+ "chunked_attention",
54
+ "chunked_attention",
55
+ "chunked_attention",
56
+ "full_attention",
57
+ "chunked_attention",
58
+ "chunked_attention",
59
+ "chunked_attention",
60
+ "full_attention",
61
+ "chunked_attention",
62
+ "chunked_attention",
63
+ "chunked_attention",
64
+ "full_attention",
65
+ "chunked_attention",
66
+ "chunked_attention",
67
+ "chunked_attention",
68
+ "full_attention"
69
+ ],
70
+ "max_position_embeddings": 1048576,
71
+ "model_type": "llama4_text",
72
+ "moe_layers": [
73
+ 1,
74
+ 3,
75
+ 5,
76
+ 7,
77
+ 9,
78
+ 11,
79
+ 13,
80
+ 15,
81
+ 17,
82
+ 19,
83
+ 21,
84
+ 23,
85
+ 25,
86
+ 27,
87
+ 29,
88
+ 31,
89
+ 33,
90
+ 35,
91
+ 37,
92
+ 39,
93
+ 41,
94
+ 43,
95
+ 45,
96
+ 47
97
+ ],
98
+ "neuron": {
99
+ "_serialized_key": "NxDNeuronConfig",
100
+ "batch_size": 1,
101
+ "capacity_factor": null,
102
+ "checkpoint_id": "meta-llama/Llama-4-Maverick-17B-128E-Instruct",
103
+ "checkpoint_revision": "73d14711bcc77c16df3470856949c3764056b617",
104
+ "continuous_batching": false,
105
+ "ep_degree": 1,
106
+ "fused_qkv": false,
107
+ "glu_mlp": true,
108
+ "local_ranks_size": 64,
109
+ "max_batch_size": 1,
110
+ "max_context_length": 8192,
111
+ "max_topk": 256,
112
+ "n_active_tokens": 8192,
113
+ "neuronxcc_version": "2.21.18209.0+043b1bf7",
114
+ "on_device_sampling": true,
115
+ "optimum_neuron_version": "0.4.3.dev1",
116
+ "output_logits": false,
117
+ "pp_degree": 1,
118
+ "sequence_length": 8192,
119
+ "speculation_length": 0,
120
+ "start_rank_id": 0,
121
+ "target": "trn2",
122
+ "torch_dtype": "bfloat16",
123
+ "tp_degree": 64
124
+ },
125
+ "no_rope_layers": [
126
+ 1,
127
+ 1,
128
+ 1,
129
+ 0,
130
+ 1,
131
+ 1,
132
+ 1,
133
+ 0,
134
+ 1,
135
+ 1,
136
+ 1,
137
+ 0,
138
+ 1,
139
+ 1,
140
+ 1,
141
+ 0,
142
+ 1,
143
+ 1,
144
+ 1,
145
+ 0,
146
+ 1,
147
+ 1,
148
+ 1,
149
+ 0,
150
+ 1,
151
+ 1,
152
+ 1,
153
+ 0,
154
+ 1,
155
+ 1,
156
+ 1,
157
+ 0,
158
+ 1,
159
+ 1,
160
+ 1,
161
+ 0,
162
+ 1,
163
+ 1,
164
+ 1,
165
+ 0,
166
+ 1,
167
+ 1,
168
+ 1,
169
+ 0,
170
+ 1,
171
+ 1,
172
+ 1,
173
+ 0
174
+ ],
175
+ "num_attention_heads": 40,
176
+ "num_experts_per_tok": 1,
177
+ "num_hidden_layers": 48,
178
+ "num_key_value_heads": 8,
179
+ "num_local_experts": 128,
180
+ "output_router_logits": false,
181
+ "rms_norm_eps": 1e-05,
182
+ "rope_scaling": null,
183
+ "rope_theta": 500000.0,
184
+ "router_aux_loss_coef": 0.001,
185
+ "router_jitter_noise": 0.0,
186
+ "tie_word_embeddings": false,
187
+ "use_cache": true,
188
+ "use_qk_norm": false,
189
+ "vocab_size": 202048
190
+ }
neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/token_generation/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:024adf15e89a8497d981f4e938a3324f26464df0317c371a54b04a9af78bbbab
3
+ size 203660169
neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:308e47f230df190897ed9e5282bf56ec98ed13a3d7a92ec4b1b666d845d1d289
3
+ size 11254784
neuronxcc-2.21.18209.0+043b1bf7/MODULE_05be74483bc93c8757df+ac10809c/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a81d483014f7a46d5e5ef1038c202899a516f7925b20a4b24698c050fe1e0ff
3
+ size 11564149
neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcaea021293d4d4472d7afda78070bec3b2760e9f748cd882859201d1a26917b
3
+ size 107351189
neuronxcc-2.21.18209.0+043b1bf7/MODULE_30b7f3226f615e011f33+a32116a7/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c059eda8abaeb544f662db22082ce4b294a6af28bc592db36e86c4a8ca6ebad5
3
+ size 39937024
neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/token_generation/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d996d1b6b51ba153eff193f4ee2995efe72f50f85f41dc33347c7a3f3b57fe4
3
+ size 204269521
neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba38c5f2f4aa3906bfab6e329ad29bc4eef48551c1b39f710c70cb42d9709c58
3
+ size 11316224
neuronxcc-2.21.18209.0+043b1bf7/MODULE_3b867487803e851b46f0+ac10809c/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34ce2961eab1a7f057e1d2a01fa3066a66c28ed8b5478a78b7c0f867f1ac69af
3
+ size 11625573
neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e540c4f1e31e36c9b251c7099e31307fc0d1e6008b967542cd4239534e1c9feb
3
+ size 208063973
neuronxcc-2.21.18209.0+043b1bf7/MODULE_416e0fa403c0921df6c3+a32116a7/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a4082b934fd93cfc3969a5b641b273afaa8002499e848790bad7912a07a4ad01
3
+ size 77569024
neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f6f942527d5e73f3ee38027454e1c6f85c9b3073d9acb209127e2b68f7e7ead
3
+ size 208356581
neuronxcc-2.21.18209.0+043b1bf7/MODULE_4200960674e1921e9aec+a32116a7/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7bd52b12ed63676b9008ab140fee0705f1afefa7b2be860962fcf9902f84e5d
3
+ size 77579264
neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90f287de9e03fb9e81b450ba63b209b2cedc3906bbdb3de4ab291532276186ec
3
+ size 207847637
neuronxcc-2.21.18209.0+043b1bf7/MODULE_4ec3b67fd9a78f7cbdb4+a32116a7/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e43e6fe7e06e1f1bc0a66df6b299ded92c4a335a23d3c984a6a8544da3b080e
3
+ size 77569024
neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/token_generation/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:833c44de5240e3c4e4dc6986a3d3429c0b1e6ed7a75c4b505c28216f9d85a76c
3
+ size 204263377
neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0bc0ef0a08a71504b94db3e2733a14e9307a354b66e5b499ee2118db26807de
3
+ size 11070464
neuronxcc-2.21.18209.0+043b1bf7/MODULE_65e34d6afbea63ef1564+ac10809c/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac9a303d035b0d4a4a595ee9dfc6211e558bf2943c767ee3af6a4d46f2b55292
3
+ size 11379813
neuronxcc-2.21.18209.0+043b1bf7/MODULE_674332a679782ee04898+a32116a7/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_674332a679782ee04898+a32116a7/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dce21659a8523d02dcd06c067d00863c54f39f5e180742c35e54fedb64d5680
3
+ size 233570789
neuronxcc-2.21.18209.0+043b1bf7/MODULE_674332a679782ee04898+a32116a7/model.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ Failed compilation with ['neuronx-cc', 'compile', '--framework=XLA', '/tmp/nxd_model/context_encoding/_tp0_bk0/model.MODULE_674332a679782ee04898+a32116a7.hlo_module.pb', '--output', '/tmp/nxd_model/context_encoding/_tp0_bk0/model.MODULE_674332a679782ee04898+a32116a7.neff', '--target=trn2', '--auto-cast=none', '--model-type=transformer', '--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ', '-O2', '--lnc=2', '--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt', '--verbose=35']: [XCG815] Estimated peak HBM usage (26.537GB) exceeds 24GB. Neff might be unable to load on chip. If you believe this estimation to be inaccurate, you can disable the check using: `--internal-backend-options=' --disable-hbm-usage-check '` - Please open a support ticket at https://github.com/aws-neuron/aws-neuron-sdk/issues/new. You may also be able to obtain more information using the 'XLA_IR_DEBUG' and 'XLA_HLO_DEBUG' environment variables.
2
+ 2025-11-27T15:03:49Z Non-signal exit. Backend exited with code 1 and stderr: [XCG815] Estimated peak HBM usage (26.537GB) exceeds 24GB. Neff might be unable to load on chip. If you believe this estimation to be inaccurate, you can disable the check using: `--internal-backend-options=' --disable-hbm-usage-check '` - Please open a support ticket at https://github.com/aws-neuron/aws-neuron-sdk/issues/new. You may also be able to obtain more information using the 'XLA_IR_DEBUG' and 'XLA_HLO_DEBUG' environment variables.
3
+
neuronxcc-2.21.18209.0+043b1bf7/MODULE_8325de44f605a823c436+a32116a7/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_8325de44f605a823c436+a32116a7/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a3e441a854ee41a2276c1518609fd4d76b0243a9324305bc07d8cd8d64d6fa1
3
+ size 309218022
neuronxcc-2.21.18209.0+043b1bf7/MODULE_8325de44f605a823c436+a32116a7/model.log ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ Failed compilation with ['neuronx-cc', 'compile', '--framework=XLA', '/tmp/nxd_model/context_encoding/_tp0_bk0/model.MODULE_8325de44f605a823c436+a32116a7.hlo_module.pb', '--output', '/tmp/nxd_model/context_encoding/_tp0_bk0/model.MODULE_8325de44f605a823c436+a32116a7.neff', '--target=trn2', '--auto-cast=none', '--model-type=transformer', '--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ', '-O2', '--lnc=2', '--logfile=/tmp/nxd_model/context_encoding/_tp0_bk0/log-neuron-cc.txt', '--verbose=35']: [XCG815] Estimated peak HBM usage (29.349GB) exceeds 24GB. Neff might be unable to load on chip. If you believe this estimation to be inaccurate, you can disable the check using: `--internal-backend-options=' --disable-hbm-usage-check '` - Please open a support ticket at https://github.com/aws-neuron/aws-neuron-sdk/issues/new. You may also be able to obtain more information using the 'XLA_IR_DEBUG' and 'XLA_HLO_DEBUG' environment variables.
2
+ 2025-11-27T12:18:44Z Non-signal exit. Backend exited with code 1 and stderr: [XCG815] Estimated peak HBM usage (29.349GB) exceeds 24GB. Neff might be unable to load on chip. If you believe this estimation to be inaccurate, you can disable the check using: `--internal-backend-options=' --disable-hbm-usage-check '` - Please open a support ticket at https://github.com/aws-neuron/aws-neuron-sdk/issues/new. You may also be able to obtain more information using the 'XLA_IR_DEBUG' and 'XLA_HLO_DEBUG' environment variables.
3
+
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/token_generation/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:608a242d7b2e8f63eb7fc01a96237daf8141e54c737fa3029ceaab5aaf2b66ec
3
+ size 304936914
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84fba4d0937a4e0d98003e8309115635db8e7521dbde620a3095507f85d1edc1
3
+ size 15391744
neuronxcc-2.21.18209.0+043b1bf7/MODULE_a42f0ea98535622da192+ac10809c/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70b13c9403176a2f84027601bf46e9f78dd3528b386784e9e054515646492279
3
+ size 15701093
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/compile_flags.json ADDED
@@ -0,0 +1 @@
 
 
1
+ ["--target=trn2", "--auto-cast=none", "--model-type=transformer", "--tensorizer-options=--enable-ccop-compute-overlap --cc-pipeline-tiling-factor=2 --vectorize-strided-dma ", "-O2", "--lnc=2", "--logfile=/tmp/nxd_model/token_generation/_tp0_bk0/log-neuron-cc.txt", "--enable-internal-neff-wrapper"]
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/model.done ADDED
File without changes
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/model.hlo_module.pb ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09751a30b4f4a733f9b2e3ea959d6e13a475daa18e9b00c87827261fa81d928c
3
+ size 103598913
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/model.neff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d59380a9125d65a7774945e905d0ad89cff759e65ac2e5f479fff373ad9e0edc
3
+ size 7169024
neuronxcc-2.21.18209.0+043b1bf7/MODULE_b50fd34438465c143456+ac10809c/wrapped_neff.hlo ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:292d6ce7151d25e424cb3eb49a31e99b915adacc4129a08d870c6744a6012c77
3
+ size 7478373