diff --git "a/model.safetensors.index.json" "b/model.safetensors.index.json" --- "a/model.safetensors.index.json" +++ "b/model.safetensors.index.json" @@ -1,725 +1,728 @@ { - "lm_head.weight": "model-00001-of-00723.safetensors", - "model.embed_tokens.weight": "model-00002-of-00723.safetensors", - "model.layers.0.input_layernorm.weight": "model-00003-of-00723.safetensors", - "model.layers.0.mlp.down_proj.weight": "model-00004-of-00723.safetensors", - "model.layers.0.mlp.gate_proj.weight": "model-00005-of-00723.safetensors", - "model.layers.0.mlp.up_proj.weight": "model-00006-of-00723.safetensors", - "model.layers.0.post_attention_layernorm.weight": "model-00007-of-00723.safetensors", - "model.layers.0.self_attn.k_proj.weight": "model-00008-of-00723.safetensors", - "model.layers.0.self_attn.o_proj.weight": "model-00009-of-00723.safetensors", - "model.layers.0.self_attn.q_proj.weight": "model-00010-of-00723.safetensors", - "model.layers.0.self_attn.v_proj.weight": "model-00011-of-00723.safetensors", - "model.layers.1.input_layernorm.weight": "model-00012-of-00723.safetensors", - "model.layers.1.mlp.down_proj.weight": "model-00013-of-00723.safetensors", - "model.layers.1.mlp.gate_proj.weight": "model-00014-of-00723.safetensors", - "model.layers.1.mlp.up_proj.weight": "model-00015-of-00723.safetensors", - "model.layers.1.post_attention_layernorm.weight": "model-00016-of-00723.safetensors", - "model.layers.1.self_attn.k_proj.weight": "model-00017-of-00723.safetensors", - "model.layers.1.self_attn.o_proj.weight": "model-00018-of-00723.safetensors", - "model.layers.1.self_attn.q_proj.weight": "model-00019-of-00723.safetensors", - "model.layers.1.self_attn.v_proj.weight": "model-00020-of-00723.safetensors", - "model.layers.2.input_layernorm.weight": "model-00021-of-00723.safetensors", - "model.layers.2.mlp.down_proj.weight": "model-00022-of-00723.safetensors", - "model.layers.2.mlp.gate_proj.weight": "model-00023-of-00723.safetensors", - "model.layers.2.mlp.up_proj.weight": "model-00024-of-00723.safetensors", - "model.layers.2.post_attention_layernorm.weight": "model-00025-of-00723.safetensors", - "model.layers.2.self_attn.k_proj.weight": "model-00026-of-00723.safetensors", - "model.layers.2.self_attn.o_proj.weight": "model-00027-of-00723.safetensors", - "model.layers.2.self_attn.q_proj.weight": "model-00028-of-00723.safetensors", - "model.layers.2.self_attn.v_proj.weight": "model-00029-of-00723.safetensors", - "model.layers.3.input_layernorm.weight": "model-00030-of-00723.safetensors", - "model.layers.3.mlp.down_proj.weight": "model-00031-of-00723.safetensors", - "model.layers.3.mlp.gate_proj.weight": "model-00032-of-00723.safetensors", - "model.layers.3.mlp.up_proj.weight": "model-00033-of-00723.safetensors", - "model.layers.3.post_attention_layernorm.weight": "model-00034-of-00723.safetensors", - "model.layers.3.self_attn.k_proj.weight": "model-00035-of-00723.safetensors", - "model.layers.3.self_attn.o_proj.weight": "model-00036-of-00723.safetensors", - "model.layers.3.self_attn.q_proj.weight": "model-00037-of-00723.safetensors", - "model.layers.3.self_attn.v_proj.weight": "model-00038-of-00723.safetensors", - "model.layers.4.input_layernorm.weight": "model-00039-of-00723.safetensors", - "model.layers.4.mlp.down_proj.weight": "model-00040-of-00723.safetensors", - "model.layers.4.mlp.gate_proj.weight": "model-00041-of-00723.safetensors", - "model.layers.4.mlp.up_proj.weight": "model-00042-of-00723.safetensors", - "model.layers.4.post_attention_layernorm.weight": "model-00043-of-00723.safetensors", - "model.layers.4.self_attn.k_proj.weight": "model-00044-of-00723.safetensors", - "model.layers.4.self_attn.o_proj.weight": "model-00045-of-00723.safetensors", - "model.layers.4.self_attn.q_proj.weight": "model-00046-of-00723.safetensors", - "model.layers.4.self_attn.v_proj.weight": "model-00047-of-00723.safetensors", - "model.layers.5.input_layernorm.weight": "model-00048-of-00723.safetensors", - "model.layers.5.mlp.down_proj.weight": "model-00049-of-00723.safetensors", - "model.layers.5.mlp.gate_proj.weight": "model-00050-of-00723.safetensors", - "model.layers.5.mlp.up_proj.weight": "model-00051-of-00723.safetensors", - "model.layers.5.post_attention_layernorm.weight": "model-00052-of-00723.safetensors", - "model.layers.5.self_attn.k_proj.weight": "model-00053-of-00723.safetensors", - "model.layers.5.self_attn.o_proj.weight": "model-00054-of-00723.safetensors", - "model.layers.5.self_attn.q_proj.weight": "model-00055-of-00723.safetensors", - "model.layers.5.self_attn.v_proj.weight": "model-00056-of-00723.safetensors", - "model.layers.6.input_layernorm.weight": "model-00057-of-00723.safetensors", - "model.layers.6.mlp.down_proj.weight": "model-00058-of-00723.safetensors", - "model.layers.6.mlp.gate_proj.weight": "model-00059-of-00723.safetensors", - "model.layers.6.mlp.up_proj.weight": "model-00060-of-00723.safetensors", - "model.layers.6.post_attention_layernorm.weight": "model-00061-of-00723.safetensors", - "model.layers.6.self_attn.k_proj.weight": "model-00062-of-00723.safetensors", - "model.layers.6.self_attn.o_proj.weight": "model-00063-of-00723.safetensors", - "model.layers.6.self_attn.q_proj.weight": "model-00064-of-00723.safetensors", - "model.layers.6.self_attn.v_proj.weight": "model-00065-of-00723.safetensors", - "model.layers.7.input_layernorm.weight": "model-00066-of-00723.safetensors", - "model.layers.7.mlp.down_proj.weight": "model-00067-of-00723.safetensors", - "model.layers.7.mlp.gate_proj.weight": "model-00068-of-00723.safetensors", - "model.layers.7.mlp.up_proj.weight": "model-00069-of-00723.safetensors", - "model.layers.7.post_attention_layernorm.weight": "model-00070-of-00723.safetensors", - "model.layers.7.self_attn.k_proj.weight": "model-00071-of-00723.safetensors", - "model.layers.7.self_attn.o_proj.weight": "model-00072-of-00723.safetensors", - "model.layers.7.self_attn.q_proj.weight": "model-00073-of-00723.safetensors", - "model.layers.7.self_attn.v_proj.weight": "model-00074-of-00723.safetensors", - "model.layers.8.input_layernorm.weight": "model-00075-of-00723.safetensors", - "model.layers.8.mlp.down_proj.weight": "model-00076-of-00723.safetensors", - "model.layers.8.mlp.gate_proj.weight": "model-00077-of-00723.safetensors", - "model.layers.8.mlp.up_proj.weight": "model-00078-of-00723.safetensors", - "model.layers.8.post_attention_layernorm.weight": "model-00079-of-00723.safetensors", - "model.layers.8.self_attn.k_proj.weight": "model-00080-of-00723.safetensors", - "model.layers.8.self_attn.o_proj.weight": "model-00081-of-00723.safetensors", - "model.layers.8.self_attn.q_proj.weight": "model-00082-of-00723.safetensors", - "model.layers.8.self_attn.v_proj.weight": "model-00083-of-00723.safetensors", - "model.layers.9.input_layernorm.weight": "model-00084-of-00723.safetensors", - "model.layers.9.mlp.down_proj.weight": "model-00085-of-00723.safetensors", - "model.layers.9.mlp.gate_proj.weight": "model-00086-of-00723.safetensors", - "model.layers.9.mlp.up_proj.weight": "model-00087-of-00723.safetensors", - "model.layers.9.post_attention_layernorm.weight": "model-00088-of-00723.safetensors", - "model.layers.9.self_attn.k_proj.weight": "model-00089-of-00723.safetensors", - "model.layers.9.self_attn.o_proj.weight": "model-00090-of-00723.safetensors", - "model.layers.9.self_attn.q_proj.weight": "model-00091-of-00723.safetensors", - "model.layers.9.self_attn.v_proj.weight": "model-00092-of-00723.safetensors", - "model.layers.10.input_layernorm.weight": "model-00093-of-00723.safetensors", - "model.layers.10.mlp.down_proj.weight": "model-00094-of-00723.safetensors", - "model.layers.10.mlp.gate_proj.weight": "model-00095-of-00723.safetensors", - "model.layers.10.mlp.up_proj.weight": "model-00096-of-00723.safetensors", - "model.layers.10.post_attention_layernorm.weight": "model-00097-of-00723.safetensors", - "model.layers.10.self_attn.k_proj.weight": "model-00098-of-00723.safetensors", - "model.layers.10.self_attn.o_proj.weight": "model-00099-of-00723.safetensors", - "model.layers.10.self_attn.q_proj.weight": "model-00100-of-00723.safetensors", - "model.layers.10.self_attn.v_proj.weight": "model-00101-of-00723.safetensors", - "model.layers.11.input_layernorm.weight": "model-00102-of-00723.safetensors", - "model.layers.11.mlp.down_proj.weight": "model-00103-of-00723.safetensors", - "model.layers.11.mlp.gate_proj.weight": "model-00104-of-00723.safetensors", - "model.layers.11.mlp.up_proj.weight": "model-00105-of-00723.safetensors", - "model.layers.11.post_attention_layernorm.weight": "model-00106-of-00723.safetensors", - "model.layers.11.self_attn.k_proj.weight": "model-00107-of-00723.safetensors", - "model.layers.11.self_attn.o_proj.weight": "model-00108-of-00723.safetensors", - "model.layers.11.self_attn.q_proj.weight": "model-00109-of-00723.safetensors", - "model.layers.11.self_attn.v_proj.weight": "model-00110-of-00723.safetensors", - "model.layers.12.input_layernorm.weight": "model-00111-of-00723.safetensors", - "model.layers.12.mlp.down_proj.weight": "model-00112-of-00723.safetensors", - "model.layers.12.mlp.gate_proj.weight": "model-00113-of-00723.safetensors", - "model.layers.12.mlp.up_proj.weight": "model-00114-of-00723.safetensors", - "model.layers.12.post_attention_layernorm.weight": "model-00115-of-00723.safetensors", - "model.layers.12.self_attn.k_proj.weight": "model-00116-of-00723.safetensors", - "model.layers.12.self_attn.o_proj.weight": "model-00117-of-00723.safetensors", - "model.layers.12.self_attn.q_proj.weight": "model-00118-of-00723.safetensors", - "model.layers.12.self_attn.v_proj.weight": "model-00119-of-00723.safetensors", - "model.layers.13.input_layernorm.weight": "model-00120-of-00723.safetensors", - "model.layers.13.mlp.down_proj.weight": "model-00121-of-00723.safetensors", - "model.layers.13.mlp.gate_proj.weight": "model-00122-of-00723.safetensors", - "model.layers.13.mlp.up_proj.weight": "model-00123-of-00723.safetensors", - "model.layers.13.post_attention_layernorm.weight": "model-00124-of-00723.safetensors", - "model.layers.13.self_attn.k_proj.weight": "model-00125-of-00723.safetensors", - "model.layers.13.self_attn.o_proj.weight": "model-00126-of-00723.safetensors", - "model.layers.13.self_attn.q_proj.weight": "model-00127-of-00723.safetensors", - "model.layers.13.self_attn.v_proj.weight": "model-00128-of-00723.safetensors", - "model.layers.14.input_layernorm.weight": "model-00129-of-00723.safetensors", - "model.layers.14.mlp.down_proj.weight": "model-00130-of-00723.safetensors", - "model.layers.14.mlp.gate_proj.weight": "model-00131-of-00723.safetensors", - "model.layers.14.mlp.up_proj.weight": "model-00132-of-00723.safetensors", - "model.layers.14.post_attention_layernorm.weight": "model-00133-of-00723.safetensors", - "model.layers.14.self_attn.k_proj.weight": "model-00134-of-00723.safetensors", - "model.layers.14.self_attn.o_proj.weight": "model-00135-of-00723.safetensors", - "model.layers.14.self_attn.q_proj.weight": "model-00136-of-00723.safetensors", - "model.layers.14.self_attn.v_proj.weight": "model-00137-of-00723.safetensors", - "model.layers.15.input_layernorm.weight": "model-00138-of-00723.safetensors", - "model.layers.15.mlp.down_proj.weight": "model-00139-of-00723.safetensors", - "model.layers.15.mlp.gate_proj.weight": "model-00140-of-00723.safetensors", - "model.layers.15.mlp.up_proj.weight": "model-00141-of-00723.safetensors", - "model.layers.15.post_attention_layernorm.weight": "model-00142-of-00723.safetensors", - "model.layers.15.self_attn.k_proj.weight": "model-00143-of-00723.safetensors", - "model.layers.15.self_attn.o_proj.weight": "model-00144-of-00723.safetensors", - "model.layers.15.self_attn.q_proj.weight": "model-00145-of-00723.safetensors", - "model.layers.15.self_attn.v_proj.weight": "model-00146-of-00723.safetensors", - "model.layers.16.input_layernorm.weight": "model-00147-of-00723.safetensors", - "model.layers.16.mlp.down_proj.weight": "model-00148-of-00723.safetensors", - "model.layers.16.mlp.gate_proj.weight": "model-00149-of-00723.safetensors", - "model.layers.16.mlp.up_proj.weight": "model-00150-of-00723.safetensors", - "model.layers.16.post_attention_layernorm.weight": "model-00151-of-00723.safetensors", - "model.layers.16.self_attn.k_proj.weight": "model-00152-of-00723.safetensors", - "model.layers.16.self_attn.o_proj.weight": "model-00153-of-00723.safetensors", - "model.layers.16.self_attn.q_proj.weight": "model-00154-of-00723.safetensors", - "model.layers.16.self_attn.v_proj.weight": "model-00155-of-00723.safetensors", - "model.layers.17.input_layernorm.weight": "model-00156-of-00723.safetensors", - "model.layers.17.mlp.down_proj.weight": "model-00157-of-00723.safetensors", - "model.layers.17.mlp.gate_proj.weight": "model-00158-of-00723.safetensors", - "model.layers.17.mlp.up_proj.weight": "model-00159-of-00723.safetensors", - "model.layers.17.post_attention_layernorm.weight": "model-00160-of-00723.safetensors", - "model.layers.17.self_attn.k_proj.weight": "model-00161-of-00723.safetensors", - "model.layers.17.self_attn.o_proj.weight": "model-00162-of-00723.safetensors", - "model.layers.17.self_attn.q_proj.weight": "model-00163-of-00723.safetensors", - "model.layers.17.self_attn.v_proj.weight": "model-00164-of-00723.safetensors", - "model.layers.18.input_layernorm.weight": "model-00165-of-00723.safetensors", - "model.layers.18.mlp.down_proj.weight": "model-00166-of-00723.safetensors", - "model.layers.18.mlp.gate_proj.weight": "model-00167-of-00723.safetensors", - "model.layers.18.mlp.up_proj.weight": "model-00168-of-00723.safetensors", - "model.layers.18.post_attention_layernorm.weight": "model-00169-of-00723.safetensors", - "model.layers.18.self_attn.k_proj.weight": "model-00170-of-00723.safetensors", - "model.layers.18.self_attn.o_proj.weight": "model-00171-of-00723.safetensors", - "model.layers.18.self_attn.q_proj.weight": "model-00172-of-00723.safetensors", - "model.layers.18.self_attn.v_proj.weight": "model-00173-of-00723.safetensors", - "model.layers.19.input_layernorm.weight": "model-00174-of-00723.safetensors", - "model.layers.19.mlp.down_proj.weight": "model-00175-of-00723.safetensors", - "model.layers.19.mlp.gate_proj.weight": "model-00176-of-00723.safetensors", - "model.layers.19.mlp.up_proj.weight": "model-00177-of-00723.safetensors", - "model.layers.19.post_attention_layernorm.weight": "model-00178-of-00723.safetensors", - "model.layers.19.self_attn.k_proj.weight": "model-00179-of-00723.safetensors", - "model.layers.19.self_attn.o_proj.weight": "model-00180-of-00723.safetensors", - "model.layers.19.self_attn.q_proj.weight": "model-00181-of-00723.safetensors", - "model.layers.19.self_attn.v_proj.weight": "model-00182-of-00723.safetensors", - "model.layers.20.input_layernorm.weight": "model-00183-of-00723.safetensors", - "model.layers.20.mlp.down_proj.weight": "model-00184-of-00723.safetensors", - "model.layers.20.mlp.gate_proj.weight": "model-00185-of-00723.safetensors", - "model.layers.20.mlp.up_proj.weight": "model-00186-of-00723.safetensors", - "model.layers.20.post_attention_layernorm.weight": "model-00187-of-00723.safetensors", - "model.layers.20.self_attn.k_proj.weight": "model-00188-of-00723.safetensors", - "model.layers.20.self_attn.o_proj.weight": "model-00189-of-00723.safetensors", - "model.layers.20.self_attn.q_proj.weight": "model-00190-of-00723.safetensors", - "model.layers.20.self_attn.v_proj.weight": "model-00191-of-00723.safetensors", - "model.layers.21.input_layernorm.weight": "model-00192-of-00723.safetensors", - "model.layers.21.mlp.down_proj.weight": "model-00193-of-00723.safetensors", - "model.layers.21.mlp.gate_proj.weight": "model-00194-of-00723.safetensors", - "model.layers.21.mlp.up_proj.weight": "model-00195-of-00723.safetensors", - "model.layers.21.post_attention_layernorm.weight": "model-00196-of-00723.safetensors", - "model.layers.21.self_attn.k_proj.weight": "model-00197-of-00723.safetensors", - "model.layers.21.self_attn.o_proj.weight": "model-00198-of-00723.safetensors", - "model.layers.21.self_attn.q_proj.weight": "model-00199-of-00723.safetensors", - "model.layers.21.self_attn.v_proj.weight": "model-00200-of-00723.safetensors", - "model.layers.22.input_layernorm.weight": "model-00201-of-00723.safetensors", - "model.layers.22.mlp.down_proj.weight": "model-00202-of-00723.safetensors", - "model.layers.22.mlp.gate_proj.weight": "model-00203-of-00723.safetensors", - "model.layers.22.mlp.up_proj.weight": "model-00204-of-00723.safetensors", - "model.layers.22.post_attention_layernorm.weight": "model-00205-of-00723.safetensors", - "model.layers.22.self_attn.k_proj.weight": "model-00206-of-00723.safetensors", - "model.layers.22.self_attn.o_proj.weight": "model-00207-of-00723.safetensors", - "model.layers.22.self_attn.q_proj.weight": "model-00208-of-00723.safetensors", - "model.layers.22.self_attn.v_proj.weight": "model-00209-of-00723.safetensors", - "model.layers.23.input_layernorm.weight": "model-00210-of-00723.safetensors", - "model.layers.23.mlp.down_proj.weight": "model-00211-of-00723.safetensors", - "model.layers.23.mlp.gate_proj.weight": "model-00212-of-00723.safetensors", - "model.layers.23.mlp.up_proj.weight": "model-00213-of-00723.safetensors", - "model.layers.23.post_attention_layernorm.weight": "model-00214-of-00723.safetensors", - "model.layers.23.self_attn.k_proj.weight": "model-00215-of-00723.safetensors", - "model.layers.23.self_attn.o_proj.weight": "model-00216-of-00723.safetensors", - "model.layers.23.self_attn.q_proj.weight": "model-00217-of-00723.safetensors", - "model.layers.23.self_attn.v_proj.weight": "model-00218-of-00723.safetensors", - "model.layers.24.input_layernorm.weight": "model-00219-of-00723.safetensors", - "model.layers.24.mlp.down_proj.weight": "model-00220-of-00723.safetensors", - "model.layers.24.mlp.gate_proj.weight": "model-00221-of-00723.safetensors", - "model.layers.24.mlp.up_proj.weight": "model-00222-of-00723.safetensors", - "model.layers.24.post_attention_layernorm.weight": "model-00223-of-00723.safetensors", - "model.layers.24.self_attn.k_proj.weight": "model-00224-of-00723.safetensors", - "model.layers.24.self_attn.o_proj.weight": "model-00225-of-00723.safetensors", - "model.layers.24.self_attn.q_proj.weight": "model-00226-of-00723.safetensors", - "model.layers.24.self_attn.v_proj.weight": "model-00227-of-00723.safetensors", - "model.layers.25.input_layernorm.weight": "model-00228-of-00723.safetensors", - "model.layers.25.mlp.down_proj.weight": "model-00229-of-00723.safetensors", - "model.layers.25.mlp.gate_proj.weight": "model-00230-of-00723.safetensors", - "model.layers.25.mlp.up_proj.weight": "model-00231-of-00723.safetensors", - "model.layers.25.post_attention_layernorm.weight": "model-00232-of-00723.safetensors", - "model.layers.25.self_attn.k_proj.weight": "model-00233-of-00723.safetensors", - "model.layers.25.self_attn.o_proj.weight": "model-00234-of-00723.safetensors", - "model.layers.25.self_attn.q_proj.weight": "model-00235-of-00723.safetensors", - "model.layers.25.self_attn.v_proj.weight": "model-00236-of-00723.safetensors", - "model.layers.26.input_layernorm.weight": "model-00237-of-00723.safetensors", - "model.layers.26.mlp.down_proj.weight": "model-00238-of-00723.safetensors", - "model.layers.26.mlp.gate_proj.weight": "model-00239-of-00723.safetensors", - "model.layers.26.mlp.up_proj.weight": "model-00240-of-00723.safetensors", - "model.layers.26.post_attention_layernorm.weight": "model-00241-of-00723.safetensors", - "model.layers.26.self_attn.k_proj.weight": "model-00242-of-00723.safetensors", - "model.layers.26.self_attn.o_proj.weight": "model-00243-of-00723.safetensors", - "model.layers.26.self_attn.q_proj.weight": "model-00244-of-00723.safetensors", - "model.layers.26.self_attn.v_proj.weight": "model-00245-of-00723.safetensors", - "model.layers.27.input_layernorm.weight": "model-00246-of-00723.safetensors", - "model.layers.27.mlp.down_proj.weight": "model-00247-of-00723.safetensors", - "model.layers.27.mlp.gate_proj.weight": "model-00248-of-00723.safetensors", - "model.layers.27.mlp.up_proj.weight": "model-00249-of-00723.safetensors", - "model.layers.27.post_attention_layernorm.weight": "model-00250-of-00723.safetensors", - "model.layers.27.self_attn.k_proj.weight": "model-00251-of-00723.safetensors", - "model.layers.27.self_attn.o_proj.weight": "model-00252-of-00723.safetensors", - "model.layers.27.self_attn.q_proj.weight": "model-00253-of-00723.safetensors", - "model.layers.27.self_attn.v_proj.weight": "model-00254-of-00723.safetensors", - "model.layers.28.input_layernorm.weight": "model-00255-of-00723.safetensors", - "model.layers.28.mlp.down_proj.weight": "model-00256-of-00723.safetensors", - "model.layers.28.mlp.gate_proj.weight": "model-00257-of-00723.safetensors", - "model.layers.28.mlp.up_proj.weight": "model-00258-of-00723.safetensors", - "model.layers.28.post_attention_layernorm.weight": "model-00259-of-00723.safetensors", - "model.layers.28.self_attn.k_proj.weight": "model-00260-of-00723.safetensors", - "model.layers.28.self_attn.o_proj.weight": "model-00261-of-00723.safetensors", - "model.layers.28.self_attn.q_proj.weight": "model-00262-of-00723.safetensors", - "model.layers.28.self_attn.v_proj.weight": "model-00263-of-00723.safetensors", - "model.layers.29.input_layernorm.weight": "model-00264-of-00723.safetensors", - "model.layers.29.mlp.down_proj.weight": "model-00265-of-00723.safetensors", - "model.layers.29.mlp.gate_proj.weight": "model-00266-of-00723.safetensors", - "model.layers.29.mlp.up_proj.weight": "model-00267-of-00723.safetensors", - "model.layers.29.post_attention_layernorm.weight": "model-00268-of-00723.safetensors", - "model.layers.29.self_attn.k_proj.weight": "model-00269-of-00723.safetensors", - "model.layers.29.self_attn.o_proj.weight": "model-00270-of-00723.safetensors", - "model.layers.29.self_attn.q_proj.weight": "model-00271-of-00723.safetensors", - "model.layers.29.self_attn.v_proj.weight": "model-00272-of-00723.safetensors", - "model.layers.30.input_layernorm.weight": "model-00273-of-00723.safetensors", - "model.layers.30.mlp.down_proj.weight": "model-00274-of-00723.safetensors", - "model.layers.30.mlp.gate_proj.weight": "model-00275-of-00723.safetensors", - "model.layers.30.mlp.up_proj.weight": "model-00276-of-00723.safetensors", - "model.layers.30.post_attention_layernorm.weight": "model-00277-of-00723.safetensors", - "model.layers.30.self_attn.k_proj.weight": "model-00278-of-00723.safetensors", - "model.layers.30.self_attn.o_proj.weight": "model-00279-of-00723.safetensors", - "model.layers.30.self_attn.q_proj.weight": "model-00280-of-00723.safetensors", - "model.layers.30.self_attn.v_proj.weight": "model-00281-of-00723.safetensors", - "model.layers.31.input_layernorm.weight": "model-00282-of-00723.safetensors", - "model.layers.31.mlp.down_proj.weight": "model-00283-of-00723.safetensors", - "model.layers.31.mlp.gate_proj.weight": "model-00284-of-00723.safetensors", - "model.layers.31.mlp.up_proj.weight": "model-00285-of-00723.safetensors", - "model.layers.31.post_attention_layernorm.weight": "model-00286-of-00723.safetensors", - "model.layers.31.self_attn.k_proj.weight": "model-00287-of-00723.safetensors", - "model.layers.31.self_attn.o_proj.weight": "model-00288-of-00723.safetensors", - "model.layers.31.self_attn.q_proj.weight": "model-00289-of-00723.safetensors", - "model.layers.31.self_attn.v_proj.weight": "model-00290-of-00723.safetensors", - "model.layers.32.input_layernorm.weight": "model-00291-of-00723.safetensors", - "model.layers.32.mlp.down_proj.weight": "model-00292-of-00723.safetensors", - "model.layers.32.mlp.gate_proj.weight": "model-00293-of-00723.safetensors", - "model.layers.32.mlp.up_proj.weight": "model-00294-of-00723.safetensors", - "model.layers.32.post_attention_layernorm.weight": "model-00295-of-00723.safetensors", - "model.layers.32.self_attn.k_proj.weight": "model-00296-of-00723.safetensors", - "model.layers.32.self_attn.o_proj.weight": "model-00297-of-00723.safetensors", - "model.layers.32.self_attn.q_proj.weight": "model-00298-of-00723.safetensors", - "model.layers.32.self_attn.v_proj.weight": "model-00299-of-00723.safetensors", - "model.layers.33.input_layernorm.weight": "model-00300-of-00723.safetensors", - "model.layers.33.mlp.down_proj.weight": "model-00301-of-00723.safetensors", - "model.layers.33.mlp.gate_proj.weight": "model-00302-of-00723.safetensors", - "model.layers.33.mlp.up_proj.weight": "model-00303-of-00723.safetensors", - "model.layers.33.post_attention_layernorm.weight": "model-00304-of-00723.safetensors", - "model.layers.33.self_attn.k_proj.weight": "model-00305-of-00723.safetensors", - "model.layers.33.self_attn.o_proj.weight": "model-00306-of-00723.safetensors", - "model.layers.33.self_attn.q_proj.weight": "model-00307-of-00723.safetensors", - "model.layers.33.self_attn.v_proj.weight": "model-00308-of-00723.safetensors", - "model.layers.34.input_layernorm.weight": "model-00309-of-00723.safetensors", - "model.layers.34.mlp.down_proj.weight": "model-00310-of-00723.safetensors", - "model.layers.34.mlp.gate_proj.weight": "model-00311-of-00723.safetensors", - "model.layers.34.mlp.up_proj.weight": "model-00312-of-00723.safetensors", - "model.layers.34.post_attention_layernorm.weight": "model-00313-of-00723.safetensors", - "model.layers.34.self_attn.k_proj.weight": "model-00314-of-00723.safetensors", - "model.layers.34.self_attn.o_proj.weight": "model-00315-of-00723.safetensors", - "model.layers.34.self_attn.q_proj.weight": "model-00316-of-00723.safetensors", - "model.layers.34.self_attn.v_proj.weight": "model-00317-of-00723.safetensors", - "model.layers.35.input_layernorm.weight": "model-00318-of-00723.safetensors", - "model.layers.35.mlp.down_proj.weight": "model-00319-of-00723.safetensors", - "model.layers.35.mlp.gate_proj.weight": "model-00320-of-00723.safetensors", - "model.layers.35.mlp.up_proj.weight": "model-00321-of-00723.safetensors", - "model.layers.35.post_attention_layernorm.weight": "model-00322-of-00723.safetensors", - "model.layers.35.self_attn.k_proj.weight": "model-00323-of-00723.safetensors", - "model.layers.35.self_attn.o_proj.weight": "model-00324-of-00723.safetensors", - "model.layers.35.self_attn.q_proj.weight": "model-00325-of-00723.safetensors", - "model.layers.35.self_attn.v_proj.weight": "model-00326-of-00723.safetensors", - "model.layers.36.input_layernorm.weight": "model-00327-of-00723.safetensors", - "model.layers.36.mlp.down_proj.weight": "model-00328-of-00723.safetensors", - "model.layers.36.mlp.gate_proj.weight": "model-00329-of-00723.safetensors", - "model.layers.36.mlp.up_proj.weight": "model-00330-of-00723.safetensors", - "model.layers.36.post_attention_layernorm.weight": "model-00331-of-00723.safetensors", - "model.layers.36.self_attn.k_proj.weight": "model-00332-of-00723.safetensors", - "model.layers.36.self_attn.o_proj.weight": "model-00333-of-00723.safetensors", - "model.layers.36.self_attn.q_proj.weight": "model-00334-of-00723.safetensors", - "model.layers.36.self_attn.v_proj.weight": "model-00335-of-00723.safetensors", - "model.layers.37.input_layernorm.weight": "model-00336-of-00723.safetensors", - "model.layers.37.mlp.down_proj.weight": "model-00337-of-00723.safetensors", - "model.layers.37.mlp.gate_proj.weight": "model-00338-of-00723.safetensors", - "model.layers.37.mlp.up_proj.weight": "model-00339-of-00723.safetensors", - "model.layers.37.post_attention_layernorm.weight": "model-00340-of-00723.safetensors", - "model.layers.37.self_attn.k_proj.weight": "model-00341-of-00723.safetensors", - "model.layers.37.self_attn.o_proj.weight": "model-00342-of-00723.safetensors", - "model.layers.37.self_attn.q_proj.weight": "model-00343-of-00723.safetensors", - "model.layers.37.self_attn.v_proj.weight": "model-00344-of-00723.safetensors", - "model.layers.38.input_layernorm.weight": "model-00345-of-00723.safetensors", - "model.layers.38.mlp.down_proj.weight": "model-00346-of-00723.safetensors", - "model.layers.38.mlp.gate_proj.weight": "model-00347-of-00723.safetensors", - "model.layers.38.mlp.up_proj.weight": "model-00348-of-00723.safetensors", - "model.layers.38.post_attention_layernorm.weight": "model-00349-of-00723.safetensors", - "model.layers.38.self_attn.k_proj.weight": "model-00350-of-00723.safetensors", - "model.layers.38.self_attn.o_proj.weight": "model-00351-of-00723.safetensors", - "model.layers.38.self_attn.q_proj.weight": "model-00352-of-00723.safetensors", - "model.layers.38.self_attn.v_proj.weight": "model-00353-of-00723.safetensors", - "model.layers.39.input_layernorm.weight": "model-00354-of-00723.safetensors", - "model.layers.39.mlp.down_proj.weight": "model-00355-of-00723.safetensors", - "model.layers.39.mlp.gate_proj.weight": "model-00356-of-00723.safetensors", - "model.layers.39.mlp.up_proj.weight": "model-00357-of-00723.safetensors", - "model.layers.39.post_attention_layernorm.weight": "model-00358-of-00723.safetensors", - "model.layers.39.self_attn.k_proj.weight": "model-00359-of-00723.safetensors", - "model.layers.39.self_attn.o_proj.weight": "model-00360-of-00723.safetensors", - "model.layers.39.self_attn.q_proj.weight": "model-00361-of-00723.safetensors", - "model.layers.39.self_attn.v_proj.weight": "model-00362-of-00723.safetensors", - "model.layers.40.input_layernorm.weight": "model-00363-of-00723.safetensors", - "model.layers.40.mlp.down_proj.weight": "model-00364-of-00723.safetensors", - "model.layers.40.mlp.gate_proj.weight": "model-00365-of-00723.safetensors", - "model.layers.40.mlp.up_proj.weight": "model-00366-of-00723.safetensors", - "model.layers.40.post_attention_layernorm.weight": "model-00367-of-00723.safetensors", - "model.layers.40.self_attn.k_proj.weight": "model-00368-of-00723.safetensors", - "model.layers.40.self_attn.o_proj.weight": "model-00369-of-00723.safetensors", - "model.layers.40.self_attn.q_proj.weight": "model-00370-of-00723.safetensors", - "model.layers.40.self_attn.v_proj.weight": "model-00371-of-00723.safetensors", - "model.layers.41.input_layernorm.weight": "model-00372-of-00723.safetensors", - "model.layers.41.mlp.down_proj.weight": "model-00373-of-00723.safetensors", - "model.layers.41.mlp.gate_proj.weight": "model-00374-of-00723.safetensors", - "model.layers.41.mlp.up_proj.weight": "model-00375-of-00723.safetensors", - "model.layers.41.post_attention_layernorm.weight": "model-00376-of-00723.safetensors", - "model.layers.41.self_attn.k_proj.weight": "model-00377-of-00723.safetensors", - "model.layers.41.self_attn.o_proj.weight": "model-00378-of-00723.safetensors", - "model.layers.41.self_attn.q_proj.weight": "model-00379-of-00723.safetensors", - "model.layers.41.self_attn.v_proj.weight": "model-00380-of-00723.safetensors", - "model.layers.42.input_layernorm.weight": "model-00381-of-00723.safetensors", - "model.layers.42.mlp.down_proj.weight": "model-00382-of-00723.safetensors", - "model.layers.42.mlp.gate_proj.weight": "model-00383-of-00723.safetensors", - "model.layers.42.mlp.up_proj.weight": "model-00384-of-00723.safetensors", - "model.layers.42.post_attention_layernorm.weight": "model-00385-of-00723.safetensors", - "model.layers.42.self_attn.k_proj.weight": "model-00386-of-00723.safetensors", - "model.layers.42.self_attn.o_proj.weight": "model-00387-of-00723.safetensors", - "model.layers.42.self_attn.q_proj.weight": "model-00388-of-00723.safetensors", - "model.layers.42.self_attn.v_proj.weight": "model-00389-of-00723.safetensors", - "model.layers.43.input_layernorm.weight": "model-00390-of-00723.safetensors", - "model.layers.43.mlp.down_proj.weight": "model-00391-of-00723.safetensors", - "model.layers.43.mlp.gate_proj.weight": "model-00392-of-00723.safetensors", - "model.layers.43.mlp.up_proj.weight": "model-00393-of-00723.safetensors", - "model.layers.43.post_attention_layernorm.weight": "model-00394-of-00723.safetensors", - "model.layers.43.self_attn.k_proj.weight": "model-00395-of-00723.safetensors", - "model.layers.43.self_attn.o_proj.weight": "model-00396-of-00723.safetensors", - "model.layers.43.self_attn.q_proj.weight": "model-00397-of-00723.safetensors", - "model.layers.43.self_attn.v_proj.weight": "model-00398-of-00723.safetensors", - "model.layers.44.input_layernorm.weight": "model-00399-of-00723.safetensors", - "model.layers.44.mlp.down_proj.weight": "model-00400-of-00723.safetensors", - "model.layers.44.mlp.gate_proj.weight": "model-00401-of-00723.safetensors", - "model.layers.44.mlp.up_proj.weight": "model-00402-of-00723.safetensors", - "model.layers.44.post_attention_layernorm.weight": "model-00403-of-00723.safetensors", - "model.layers.44.self_attn.k_proj.weight": "model-00404-of-00723.safetensors", - "model.layers.44.self_attn.o_proj.weight": "model-00405-of-00723.safetensors", - "model.layers.44.self_attn.q_proj.weight": "model-00406-of-00723.safetensors", - "model.layers.44.self_attn.v_proj.weight": "model-00407-of-00723.safetensors", - "model.layers.45.input_layernorm.weight": "model-00408-of-00723.safetensors", - "model.layers.45.mlp.down_proj.weight": "model-00409-of-00723.safetensors", - "model.layers.45.mlp.gate_proj.weight": "model-00410-of-00723.safetensors", - "model.layers.45.mlp.up_proj.weight": "model-00411-of-00723.safetensors", - "model.layers.45.post_attention_layernorm.weight": "model-00412-of-00723.safetensors", - "model.layers.45.self_attn.k_proj.weight": "model-00413-of-00723.safetensors", - "model.layers.45.self_attn.o_proj.weight": "model-00414-of-00723.safetensors", - "model.layers.45.self_attn.q_proj.weight": "model-00415-of-00723.safetensors", - "model.layers.45.self_attn.v_proj.weight": "model-00416-of-00723.safetensors", - "model.layers.46.input_layernorm.weight": "model-00417-of-00723.safetensors", - "model.layers.46.mlp.down_proj.weight": "model-00418-of-00723.safetensors", - "model.layers.46.mlp.gate_proj.weight": "model-00419-of-00723.safetensors", - "model.layers.46.mlp.up_proj.weight": "model-00420-of-00723.safetensors", - "model.layers.46.post_attention_layernorm.weight": "model-00421-of-00723.safetensors", - "model.layers.46.self_attn.k_proj.weight": "model-00422-of-00723.safetensors", - "model.layers.46.self_attn.o_proj.weight": "model-00423-of-00723.safetensors", - "model.layers.46.self_attn.q_proj.weight": "model-00424-of-00723.safetensors", - "model.layers.46.self_attn.v_proj.weight": "model-00425-of-00723.safetensors", - "model.layers.47.input_layernorm.weight": "model-00426-of-00723.safetensors", - "model.layers.47.mlp.down_proj.weight": "model-00427-of-00723.safetensors", - "model.layers.47.mlp.gate_proj.weight": "model-00428-of-00723.safetensors", - "model.layers.47.mlp.up_proj.weight": "model-00429-of-00723.safetensors", - "model.layers.47.post_attention_layernorm.weight": "model-00430-of-00723.safetensors", - "model.layers.47.self_attn.k_proj.weight": "model-00431-of-00723.safetensors", - "model.layers.47.self_attn.o_proj.weight": "model-00432-of-00723.safetensors", - "model.layers.47.self_attn.q_proj.weight": "model-00433-of-00723.safetensors", - "model.layers.47.self_attn.v_proj.weight": "model-00434-of-00723.safetensors", - "model.layers.48.input_layernorm.weight": "model-00435-of-00723.safetensors", - "model.layers.48.mlp.down_proj.weight": "model-00436-of-00723.safetensors", - "model.layers.48.mlp.gate_proj.weight": "model-00437-of-00723.safetensors", - "model.layers.48.mlp.up_proj.weight": "model-00438-of-00723.safetensors", - "model.layers.48.post_attention_layernorm.weight": "model-00439-of-00723.safetensors", - "model.layers.48.self_attn.k_proj.weight": "model-00440-of-00723.safetensors", - "model.layers.48.self_attn.o_proj.weight": "model-00441-of-00723.safetensors", - "model.layers.48.self_attn.q_proj.weight": "model-00442-of-00723.safetensors", - "model.layers.48.self_attn.v_proj.weight": "model-00443-of-00723.safetensors", - "model.layers.49.input_layernorm.weight": "model-00444-of-00723.safetensors", - "model.layers.49.mlp.down_proj.weight": "model-00445-of-00723.safetensors", - "model.layers.49.mlp.gate_proj.weight": "model-00446-of-00723.safetensors", - "model.layers.49.mlp.up_proj.weight": "model-00447-of-00723.safetensors", - "model.layers.49.post_attention_layernorm.weight": "model-00448-of-00723.safetensors", - "model.layers.49.self_attn.k_proj.weight": "model-00449-of-00723.safetensors", - "model.layers.49.self_attn.o_proj.weight": "model-00450-of-00723.safetensors", - "model.layers.49.self_attn.q_proj.weight": "model-00451-of-00723.safetensors", - "model.layers.49.self_attn.v_proj.weight": "model-00452-of-00723.safetensors", - "model.layers.50.input_layernorm.weight": "model-00453-of-00723.safetensors", - "model.layers.50.mlp.down_proj.weight": "model-00454-of-00723.safetensors", - "model.layers.50.mlp.gate_proj.weight": "model-00455-of-00723.safetensors", - "model.layers.50.mlp.up_proj.weight": "model-00456-of-00723.safetensors", - "model.layers.50.post_attention_layernorm.weight": "model-00457-of-00723.safetensors", - "model.layers.50.self_attn.k_proj.weight": "model-00458-of-00723.safetensors", - "model.layers.50.self_attn.o_proj.weight": "model-00459-of-00723.safetensors", - "model.layers.50.self_attn.q_proj.weight": "model-00460-of-00723.safetensors", - "model.layers.50.self_attn.v_proj.weight": "model-00461-of-00723.safetensors", - "model.layers.51.input_layernorm.weight": "model-00462-of-00723.safetensors", - "model.layers.51.mlp.down_proj.weight": "model-00463-of-00723.safetensors", - "model.layers.51.mlp.gate_proj.weight": "model-00464-of-00723.safetensors", - "model.layers.51.mlp.up_proj.weight": "model-00465-of-00723.safetensors", - "model.layers.51.post_attention_layernorm.weight": "model-00466-of-00723.safetensors", - "model.layers.51.self_attn.k_proj.weight": "model-00467-of-00723.safetensors", - "model.layers.51.self_attn.o_proj.weight": "model-00468-of-00723.safetensors", - "model.layers.51.self_attn.q_proj.weight": "model-00469-of-00723.safetensors", - "model.layers.51.self_attn.v_proj.weight": "model-00470-of-00723.safetensors", - "model.layers.52.input_layernorm.weight": "model-00471-of-00723.safetensors", - "model.layers.52.mlp.down_proj.weight": "model-00472-of-00723.safetensors", - "model.layers.52.mlp.gate_proj.weight": "model-00473-of-00723.safetensors", - "model.layers.52.mlp.up_proj.weight": "model-00474-of-00723.safetensors", - "model.layers.52.post_attention_layernorm.weight": "model-00475-of-00723.safetensors", - "model.layers.52.self_attn.k_proj.weight": "model-00476-of-00723.safetensors", - "model.layers.52.self_attn.o_proj.weight": "model-00477-of-00723.safetensors", - "model.layers.52.self_attn.q_proj.weight": "model-00478-of-00723.safetensors", - "model.layers.52.self_attn.v_proj.weight": "model-00479-of-00723.safetensors", - "model.layers.53.input_layernorm.weight": "model-00480-of-00723.safetensors", - "model.layers.53.mlp.down_proj.weight": "model-00481-of-00723.safetensors", - "model.layers.53.mlp.gate_proj.weight": "model-00482-of-00723.safetensors", - "model.layers.53.mlp.up_proj.weight": "model-00483-of-00723.safetensors", - "model.layers.53.post_attention_layernorm.weight": "model-00484-of-00723.safetensors", - "model.layers.53.self_attn.k_proj.weight": "model-00485-of-00723.safetensors", - "model.layers.53.self_attn.o_proj.weight": "model-00486-of-00723.safetensors", - "model.layers.53.self_attn.q_proj.weight": "model-00487-of-00723.safetensors", - "model.layers.53.self_attn.v_proj.weight": "model-00488-of-00723.safetensors", - "model.layers.54.input_layernorm.weight": "model-00489-of-00723.safetensors", - "model.layers.54.mlp.down_proj.weight": "model-00490-of-00723.safetensors", - "model.layers.54.mlp.gate_proj.weight": "model-00491-of-00723.safetensors", - "model.layers.54.mlp.up_proj.weight": "model-00492-of-00723.safetensors", - "model.layers.54.post_attention_layernorm.weight": "model-00493-of-00723.safetensors", - "model.layers.54.self_attn.k_proj.weight": "model-00494-of-00723.safetensors", - "model.layers.54.self_attn.o_proj.weight": "model-00495-of-00723.safetensors", - "model.layers.54.self_attn.q_proj.weight": "model-00496-of-00723.safetensors", - "model.layers.54.self_attn.v_proj.weight": "model-00497-of-00723.safetensors", - "model.layers.55.input_layernorm.weight": "model-00498-of-00723.safetensors", - "model.layers.55.mlp.down_proj.weight": "model-00499-of-00723.safetensors", - "model.layers.55.mlp.gate_proj.weight": "model-00500-of-00723.safetensors", - "model.layers.55.mlp.up_proj.weight": "model-00501-of-00723.safetensors", - "model.layers.55.post_attention_layernorm.weight": "model-00502-of-00723.safetensors", - "model.layers.55.self_attn.k_proj.weight": "model-00503-of-00723.safetensors", - "model.layers.55.self_attn.o_proj.weight": "model-00504-of-00723.safetensors", - "model.layers.55.self_attn.q_proj.weight": "model-00505-of-00723.safetensors", - "model.layers.55.self_attn.v_proj.weight": "model-00506-of-00723.safetensors", - "model.layers.56.input_layernorm.weight": "model-00507-of-00723.safetensors", - "model.layers.56.mlp.down_proj.weight": "model-00508-of-00723.safetensors", - "model.layers.56.mlp.gate_proj.weight": "model-00509-of-00723.safetensors", - "model.layers.56.mlp.up_proj.weight": "model-00510-of-00723.safetensors", - "model.layers.56.post_attention_layernorm.weight": "model-00511-of-00723.safetensors", - "model.layers.56.self_attn.k_proj.weight": "model-00512-of-00723.safetensors", - "model.layers.56.self_attn.o_proj.weight": "model-00513-of-00723.safetensors", - "model.layers.56.self_attn.q_proj.weight": "model-00514-of-00723.safetensors", - "model.layers.56.self_attn.v_proj.weight": "model-00515-of-00723.safetensors", - "model.layers.57.input_layernorm.weight": "model-00516-of-00723.safetensors", - "model.layers.57.mlp.down_proj.weight": "model-00517-of-00723.safetensors", - "model.layers.57.mlp.gate_proj.weight": "model-00518-of-00723.safetensors", - "model.layers.57.mlp.up_proj.weight": "model-00519-of-00723.safetensors", - "model.layers.57.post_attention_layernorm.weight": "model-00520-of-00723.safetensors", - "model.layers.57.self_attn.k_proj.weight": "model-00521-of-00723.safetensors", - "model.layers.57.self_attn.o_proj.weight": "model-00522-of-00723.safetensors", - "model.layers.57.self_attn.q_proj.weight": "model-00523-of-00723.safetensors", - "model.layers.57.self_attn.v_proj.weight": "model-00524-of-00723.safetensors", - "model.layers.58.input_layernorm.weight": "model-00525-of-00723.safetensors", - "model.layers.58.mlp.down_proj.weight": "model-00526-of-00723.safetensors", - "model.layers.58.mlp.gate_proj.weight": "model-00527-of-00723.safetensors", - "model.layers.58.mlp.up_proj.weight": "model-00528-of-00723.safetensors", - "model.layers.58.post_attention_layernorm.weight": "model-00529-of-00723.safetensors", - "model.layers.58.self_attn.k_proj.weight": "model-00530-of-00723.safetensors", - "model.layers.58.self_attn.o_proj.weight": "model-00531-of-00723.safetensors", - "model.layers.58.self_attn.q_proj.weight": "model-00532-of-00723.safetensors", - "model.layers.58.self_attn.v_proj.weight": "model-00533-of-00723.safetensors", - "model.layers.59.input_layernorm.weight": "model-00534-of-00723.safetensors", - "model.layers.59.mlp.down_proj.weight": "model-00535-of-00723.safetensors", - "model.layers.59.mlp.gate_proj.weight": "model-00536-of-00723.safetensors", - "model.layers.59.mlp.up_proj.weight": "model-00537-of-00723.safetensors", - "model.layers.59.post_attention_layernorm.weight": "model-00538-of-00723.safetensors", - "model.layers.59.self_attn.k_proj.weight": "model-00539-of-00723.safetensors", - "model.layers.59.self_attn.o_proj.weight": "model-00540-of-00723.safetensors", - "model.layers.59.self_attn.q_proj.weight": "model-00541-of-00723.safetensors", - "model.layers.59.self_attn.v_proj.weight": "model-00542-of-00723.safetensors", - "model.layers.60.input_layernorm.weight": "model-00543-of-00723.safetensors", - "model.layers.60.mlp.down_proj.weight": "model-00544-of-00723.safetensors", - "model.layers.60.mlp.gate_proj.weight": "model-00545-of-00723.safetensors", - "model.layers.60.mlp.up_proj.weight": "model-00546-of-00723.safetensors", - "model.layers.60.post_attention_layernorm.weight": "model-00547-of-00723.safetensors", - "model.layers.60.self_attn.k_proj.weight": "model-00548-of-00723.safetensors", - "model.layers.60.self_attn.o_proj.weight": "model-00549-of-00723.safetensors", - "model.layers.60.self_attn.q_proj.weight": "model-00550-of-00723.safetensors", - "model.layers.60.self_attn.v_proj.weight": "model-00551-of-00723.safetensors", - "model.layers.61.input_layernorm.weight": "model-00552-of-00723.safetensors", - "model.layers.61.mlp.down_proj.weight": "model-00553-of-00723.safetensors", - "model.layers.61.mlp.gate_proj.weight": "model-00554-of-00723.safetensors", - "model.layers.61.mlp.up_proj.weight": "model-00555-of-00723.safetensors", - "model.layers.61.post_attention_layernorm.weight": "model-00556-of-00723.safetensors", - "model.layers.61.self_attn.k_proj.weight": "model-00557-of-00723.safetensors", - "model.layers.61.self_attn.o_proj.weight": "model-00558-of-00723.safetensors", - "model.layers.61.self_attn.q_proj.weight": "model-00559-of-00723.safetensors", - "model.layers.61.self_attn.v_proj.weight": "model-00560-of-00723.safetensors", - "model.layers.62.input_layernorm.weight": "model-00561-of-00723.safetensors", - "model.layers.62.mlp.down_proj.weight": "model-00562-of-00723.safetensors", - "model.layers.62.mlp.gate_proj.weight": "model-00563-of-00723.safetensors", - "model.layers.62.mlp.up_proj.weight": "model-00564-of-00723.safetensors", - "model.layers.62.post_attention_layernorm.weight": "model-00565-of-00723.safetensors", - "model.layers.62.self_attn.k_proj.weight": "model-00566-of-00723.safetensors", - "model.layers.62.self_attn.o_proj.weight": "model-00567-of-00723.safetensors", - "model.layers.62.self_attn.q_proj.weight": "model-00568-of-00723.safetensors", - "model.layers.62.self_attn.v_proj.weight": "model-00569-of-00723.safetensors", - "model.layers.63.input_layernorm.weight": "model-00570-of-00723.safetensors", - "model.layers.63.mlp.down_proj.weight": "model-00571-of-00723.safetensors", - "model.layers.63.mlp.gate_proj.weight": "model-00572-of-00723.safetensors", - "model.layers.63.mlp.up_proj.weight": "model-00573-of-00723.safetensors", - "model.layers.63.post_attention_layernorm.weight": "model-00574-of-00723.safetensors", - "model.layers.63.self_attn.k_proj.weight": "model-00575-of-00723.safetensors", - "model.layers.63.self_attn.o_proj.weight": "model-00576-of-00723.safetensors", - "model.layers.63.self_attn.q_proj.weight": "model-00577-of-00723.safetensors", - "model.layers.63.self_attn.v_proj.weight": "model-00578-of-00723.safetensors", - "model.layers.64.input_layernorm.weight": "model-00579-of-00723.safetensors", - "model.layers.64.mlp.down_proj.weight": "model-00580-of-00723.safetensors", - "model.layers.64.mlp.gate_proj.weight": "model-00581-of-00723.safetensors", - "model.layers.64.mlp.up_proj.weight": "model-00582-of-00723.safetensors", - "model.layers.64.post_attention_layernorm.weight": "model-00583-of-00723.safetensors", - "model.layers.64.self_attn.k_proj.weight": "model-00584-of-00723.safetensors", - "model.layers.64.self_attn.o_proj.weight": "model-00585-of-00723.safetensors", - "model.layers.64.self_attn.q_proj.weight": "model-00586-of-00723.safetensors", - "model.layers.64.self_attn.v_proj.weight": "model-00587-of-00723.safetensors", - "model.layers.65.input_layernorm.weight": "model-00588-of-00723.safetensors", - "model.layers.65.mlp.down_proj.weight": "model-00589-of-00723.safetensors", - "model.layers.65.mlp.gate_proj.weight": "model-00590-of-00723.safetensors", - "model.layers.65.mlp.up_proj.weight": "model-00591-of-00723.safetensors", - "model.layers.65.post_attention_layernorm.weight": "model-00592-of-00723.safetensors", - "model.layers.65.self_attn.k_proj.weight": "model-00593-of-00723.safetensors", - "model.layers.65.self_attn.o_proj.weight": "model-00594-of-00723.safetensors", - "model.layers.65.self_attn.q_proj.weight": "model-00595-of-00723.safetensors", - "model.layers.65.self_attn.v_proj.weight": "model-00596-of-00723.safetensors", - "model.layers.66.input_layernorm.weight": "model-00597-of-00723.safetensors", - "model.layers.66.mlp.down_proj.weight": "model-00598-of-00723.safetensors", - "model.layers.66.mlp.gate_proj.weight": "model-00599-of-00723.safetensors", - "model.layers.66.mlp.up_proj.weight": "model-00600-of-00723.safetensors", - "model.layers.66.post_attention_layernorm.weight": "model-00601-of-00723.safetensors", - "model.layers.66.self_attn.k_proj.weight": "model-00602-of-00723.safetensors", - "model.layers.66.self_attn.o_proj.weight": "model-00603-of-00723.safetensors", - "model.layers.66.self_attn.q_proj.weight": "model-00604-of-00723.safetensors", - "model.layers.66.self_attn.v_proj.weight": "model-00605-of-00723.safetensors", - "model.layers.67.input_layernorm.weight": "model-00606-of-00723.safetensors", - "model.layers.67.mlp.down_proj.weight": "model-00607-of-00723.safetensors", - "model.layers.67.mlp.gate_proj.weight": "model-00608-of-00723.safetensors", - "model.layers.67.mlp.up_proj.weight": "model-00609-of-00723.safetensors", - "model.layers.67.post_attention_layernorm.weight": "model-00610-of-00723.safetensors", - "model.layers.67.self_attn.k_proj.weight": "model-00611-of-00723.safetensors", - "model.layers.67.self_attn.o_proj.weight": "model-00612-of-00723.safetensors", - "model.layers.67.self_attn.q_proj.weight": "model-00613-of-00723.safetensors", - "model.layers.67.self_attn.v_proj.weight": "model-00614-of-00723.safetensors", - "model.layers.68.input_layernorm.weight": "model-00615-of-00723.safetensors", - "model.layers.68.mlp.down_proj.weight": "model-00616-of-00723.safetensors", - "model.layers.68.mlp.gate_proj.weight": "model-00617-of-00723.safetensors", - "model.layers.68.mlp.up_proj.weight": "model-00618-of-00723.safetensors", - "model.layers.68.post_attention_layernorm.weight": "model-00619-of-00723.safetensors", - "model.layers.68.self_attn.k_proj.weight": "model-00620-of-00723.safetensors", - "model.layers.68.self_attn.o_proj.weight": "model-00621-of-00723.safetensors", - "model.layers.68.self_attn.q_proj.weight": "model-00622-of-00723.safetensors", - "model.layers.68.self_attn.v_proj.weight": "model-00623-of-00723.safetensors", - "model.layers.69.input_layernorm.weight": "model-00624-of-00723.safetensors", - "model.layers.69.mlp.down_proj.weight": "model-00625-of-00723.safetensors", - "model.layers.69.mlp.gate_proj.weight": "model-00626-of-00723.safetensors", - "model.layers.69.mlp.up_proj.weight": "model-00627-of-00723.safetensors", - "model.layers.69.post_attention_layernorm.weight": "model-00628-of-00723.safetensors", - "model.layers.69.self_attn.k_proj.weight": "model-00629-of-00723.safetensors", - "model.layers.69.self_attn.o_proj.weight": "model-00630-of-00723.safetensors", - "model.layers.69.self_attn.q_proj.weight": "model-00631-of-00723.safetensors", - "model.layers.69.self_attn.v_proj.weight": "model-00632-of-00723.safetensors", - "model.layers.70.input_layernorm.weight": "model-00633-of-00723.safetensors", - "model.layers.70.mlp.down_proj.weight": "model-00634-of-00723.safetensors", - "model.layers.70.mlp.gate_proj.weight": "model-00635-of-00723.safetensors", - "model.layers.70.mlp.up_proj.weight": "model-00636-of-00723.safetensors", - "model.layers.70.post_attention_layernorm.weight": "model-00637-of-00723.safetensors", - "model.layers.70.self_attn.k_proj.weight": "model-00638-of-00723.safetensors", - "model.layers.70.self_attn.o_proj.weight": "model-00639-of-00723.safetensors", - "model.layers.70.self_attn.q_proj.weight": "model-00640-of-00723.safetensors", - "model.layers.70.self_attn.v_proj.weight": "model-00641-of-00723.safetensors", - "model.layers.71.input_layernorm.weight": "model-00642-of-00723.safetensors", - "model.layers.71.mlp.down_proj.weight": "model-00643-of-00723.safetensors", - "model.layers.71.mlp.gate_proj.weight": "model-00644-of-00723.safetensors", - "model.layers.71.mlp.up_proj.weight": "model-00645-of-00723.safetensors", - "model.layers.71.post_attention_layernorm.weight": "model-00646-of-00723.safetensors", - "model.layers.71.self_attn.k_proj.weight": "model-00647-of-00723.safetensors", - "model.layers.71.self_attn.o_proj.weight": "model-00648-of-00723.safetensors", - "model.layers.71.self_attn.q_proj.weight": "model-00649-of-00723.safetensors", - "model.layers.71.self_attn.v_proj.weight": "model-00650-of-00723.safetensors", - "model.layers.72.input_layernorm.weight": "model-00651-of-00723.safetensors", - "model.layers.72.mlp.down_proj.weight": "model-00652-of-00723.safetensors", - "model.layers.72.mlp.gate_proj.weight": "model-00653-of-00723.safetensors", - "model.layers.72.mlp.up_proj.weight": "model-00654-of-00723.safetensors", - "model.layers.72.post_attention_layernorm.weight": "model-00655-of-00723.safetensors", - "model.layers.72.self_attn.k_proj.weight": "model-00656-of-00723.safetensors", - "model.layers.72.self_attn.o_proj.weight": "model-00657-of-00723.safetensors", - "model.layers.72.self_attn.q_proj.weight": "model-00658-of-00723.safetensors", - "model.layers.72.self_attn.v_proj.weight": "model-00659-of-00723.safetensors", - "model.layers.73.input_layernorm.weight": "model-00660-of-00723.safetensors", - "model.layers.73.mlp.down_proj.weight": "model-00661-of-00723.safetensors", - "model.layers.73.mlp.gate_proj.weight": "model-00662-of-00723.safetensors", - "model.layers.73.mlp.up_proj.weight": "model-00663-of-00723.safetensors", - "model.layers.73.post_attention_layernorm.weight": "model-00664-of-00723.safetensors", - "model.layers.73.self_attn.k_proj.weight": "model-00665-of-00723.safetensors", - "model.layers.73.self_attn.o_proj.weight": "model-00666-of-00723.safetensors", - "model.layers.73.self_attn.q_proj.weight": "model-00667-of-00723.safetensors", - "model.layers.73.self_attn.v_proj.weight": "model-00668-of-00723.safetensors", - "model.layers.74.input_layernorm.weight": "model-00669-of-00723.safetensors", - "model.layers.74.mlp.down_proj.weight": "model-00670-of-00723.safetensors", - "model.layers.74.mlp.gate_proj.weight": "model-00671-of-00723.safetensors", - "model.layers.74.mlp.up_proj.weight": "model-00672-of-00723.safetensors", - "model.layers.74.post_attention_layernorm.weight": "model-00673-of-00723.safetensors", - "model.layers.74.self_attn.k_proj.weight": "model-00674-of-00723.safetensors", - "model.layers.74.self_attn.o_proj.weight": "model-00675-of-00723.safetensors", - "model.layers.74.self_attn.q_proj.weight": "model-00676-of-00723.safetensors", - "model.layers.74.self_attn.v_proj.weight": "model-00677-of-00723.safetensors", - "model.layers.75.input_layernorm.weight": "model-00678-of-00723.safetensors", - "model.layers.75.mlp.down_proj.weight": "model-00679-of-00723.safetensors", - "model.layers.75.mlp.gate_proj.weight": "model-00680-of-00723.safetensors", - "model.layers.75.mlp.up_proj.weight": "model-00681-of-00723.safetensors", - "model.layers.75.post_attention_layernorm.weight": "model-00682-of-00723.safetensors", - "model.layers.75.self_attn.k_proj.weight": "model-00683-of-00723.safetensors", - "model.layers.75.self_attn.o_proj.weight": "model-00684-of-00723.safetensors", - "model.layers.75.self_attn.q_proj.weight": "model-00685-of-00723.safetensors", - "model.layers.75.self_attn.v_proj.weight": "model-00686-of-00723.safetensors", - "model.layers.76.input_layernorm.weight": "model-00687-of-00723.safetensors", - "model.layers.76.mlp.down_proj.weight": "model-00688-of-00723.safetensors", - "model.layers.76.mlp.gate_proj.weight": "model-00689-of-00723.safetensors", - "model.layers.76.mlp.up_proj.weight": "model-00690-of-00723.safetensors", - "model.layers.76.post_attention_layernorm.weight": "model-00691-of-00723.safetensors", - "model.layers.76.self_attn.k_proj.weight": "model-00692-of-00723.safetensors", - "model.layers.76.self_attn.o_proj.weight": "model-00693-of-00723.safetensors", - "model.layers.76.self_attn.q_proj.weight": "model-00694-of-00723.safetensors", - "model.layers.76.self_attn.v_proj.weight": "model-00695-of-00723.safetensors", - "model.layers.77.input_layernorm.weight": "model-00696-of-00723.safetensors", - "model.layers.77.mlp.down_proj.weight": "model-00697-of-00723.safetensors", - "model.layers.77.mlp.gate_proj.weight": "model-00698-of-00723.safetensors", - "model.layers.77.mlp.up_proj.weight": "model-00699-of-00723.safetensors", - "model.layers.77.post_attention_layernorm.weight": "model-00700-of-00723.safetensors", - "model.layers.77.self_attn.k_proj.weight": "model-00701-of-00723.safetensors", - "model.layers.77.self_attn.o_proj.weight": "model-00702-of-00723.safetensors", - "model.layers.77.self_attn.q_proj.weight": "model-00703-of-00723.safetensors", - "model.layers.77.self_attn.v_proj.weight": "model-00704-of-00723.safetensors", - "model.layers.78.input_layernorm.weight": "model-00705-of-00723.safetensors", - "model.layers.78.mlp.down_proj.weight": "model-00706-of-00723.safetensors", - "model.layers.78.mlp.gate_proj.weight": "model-00707-of-00723.safetensors", - "model.layers.78.mlp.up_proj.weight": "model-00708-of-00723.safetensors", - "model.layers.78.post_attention_layernorm.weight": "model-00709-of-00723.safetensors", - "model.layers.78.self_attn.k_proj.weight": "model-00710-of-00723.safetensors", - "model.layers.78.self_attn.o_proj.weight": "model-00711-of-00723.safetensors", - "model.layers.78.self_attn.q_proj.weight": "model-00712-of-00723.safetensors", - "model.layers.78.self_attn.v_proj.weight": "model-00713-of-00723.safetensors", - "model.layers.79.input_layernorm.weight": "model-00714-of-00723.safetensors", - "model.layers.79.mlp.down_proj.weight": "model-00715-of-00723.safetensors", - "model.layers.79.mlp.gate_proj.weight": "model-00716-of-00723.safetensors", - "model.layers.79.mlp.up_proj.weight": "model-00717-of-00723.safetensors", - "model.layers.79.post_attention_layernorm.weight": "model-00718-of-00723.safetensors", - "model.layers.79.self_attn.k_proj.weight": "model-00719-of-00723.safetensors", - "model.layers.79.self_attn.o_proj.weight": "model-00720-of-00723.safetensors", - "model.layers.79.self_attn.q_proj.weight": "model-00721-of-00723.safetensors", - "model.layers.79.self_attn.v_proj.weight": "model-00722-of-00723.safetensors", - "model.norm.weight": "model-00723-of-00723.safetensors" + "metadata": "", + "weight_map": { + "lm_head.weight": "model-00001-of-00723.safetensors", + "model.embed_tokens.weight": "model-00002-of-00723.safetensors", + "model.layers.0.input_layernorm.weight": "model-00003-of-00723.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00004-of-00723.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00005-of-00723.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00006-of-00723.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00007-of-00723.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00008-of-00723.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00009-of-00723.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00010-of-00723.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00011-of-00723.safetensors", + "model.layers.1.input_layernorm.weight": "model-00012-of-00723.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00013-of-00723.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00014-of-00723.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00015-of-00723.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00016-of-00723.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00017-of-00723.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00018-of-00723.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00019-of-00723.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00020-of-00723.safetensors", + "model.layers.2.input_layernorm.weight": "model-00021-of-00723.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00022-of-00723.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00023-of-00723.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00024-of-00723.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00025-of-00723.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00026-of-00723.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00027-of-00723.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00028-of-00723.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00029-of-00723.safetensors", + "model.layers.3.input_layernorm.weight": "model-00030-of-00723.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00031-of-00723.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00032-of-00723.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00033-of-00723.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00034-of-00723.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00035-of-00723.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00036-of-00723.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00037-of-00723.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00038-of-00723.safetensors", + "model.layers.4.input_layernorm.weight": "model-00039-of-00723.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00040-of-00723.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00041-of-00723.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00042-of-00723.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00043-of-00723.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00044-of-00723.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00045-of-00723.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00046-of-00723.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00047-of-00723.safetensors", + "model.layers.5.input_layernorm.weight": "model-00048-of-00723.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00049-of-00723.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00050-of-00723.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00051-of-00723.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00052-of-00723.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00053-of-00723.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00054-of-00723.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00055-of-00723.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00056-of-00723.safetensors", + "model.layers.6.input_layernorm.weight": "model-00057-of-00723.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00058-of-00723.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00059-of-00723.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00060-of-00723.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00061-of-00723.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00062-of-00723.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00063-of-00723.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00064-of-00723.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00065-of-00723.safetensors", + "model.layers.7.input_layernorm.weight": "model-00066-of-00723.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00067-of-00723.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00068-of-00723.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00069-of-00723.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00070-of-00723.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00071-of-00723.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00072-of-00723.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00073-of-00723.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00074-of-00723.safetensors", + "model.layers.8.input_layernorm.weight": "model-00075-of-00723.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00076-of-00723.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00077-of-00723.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00078-of-00723.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00079-of-00723.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00080-of-00723.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00081-of-00723.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00082-of-00723.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00083-of-00723.safetensors", + "model.layers.9.input_layernorm.weight": "model-00084-of-00723.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00085-of-00723.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00086-of-00723.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00087-of-00723.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00088-of-00723.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00089-of-00723.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00090-of-00723.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00091-of-00723.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00092-of-00723.safetensors", + "model.layers.10.input_layernorm.weight": "model-00093-of-00723.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00094-of-00723.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00095-of-00723.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00096-of-00723.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00097-of-00723.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00098-of-00723.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00099-of-00723.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00100-of-00723.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00101-of-00723.safetensors", + "model.layers.11.input_layernorm.weight": "model-00102-of-00723.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00103-of-00723.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00104-of-00723.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00105-of-00723.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00106-of-00723.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00107-of-00723.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00108-of-00723.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00109-of-00723.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00110-of-00723.safetensors", + "model.layers.12.input_layernorm.weight": "model-00111-of-00723.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00112-of-00723.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00113-of-00723.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00114-of-00723.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00115-of-00723.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00116-of-00723.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00117-of-00723.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00118-of-00723.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00119-of-00723.safetensors", + "model.layers.13.input_layernorm.weight": "model-00120-of-00723.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00121-of-00723.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00122-of-00723.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00123-of-00723.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00124-of-00723.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00125-of-00723.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00126-of-00723.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00127-of-00723.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00128-of-00723.safetensors", + "model.layers.14.input_layernorm.weight": "model-00129-of-00723.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00130-of-00723.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00131-of-00723.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00132-of-00723.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00133-of-00723.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00134-of-00723.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00135-of-00723.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00136-of-00723.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00137-of-00723.safetensors", + "model.layers.15.input_layernorm.weight": "model-00138-of-00723.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00139-of-00723.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00140-of-00723.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00141-of-00723.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00142-of-00723.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00143-of-00723.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00144-of-00723.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00145-of-00723.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00146-of-00723.safetensors", + "model.layers.16.input_layernorm.weight": "model-00147-of-00723.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00148-of-00723.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00149-of-00723.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00150-of-00723.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00151-of-00723.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00152-of-00723.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00153-of-00723.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00154-of-00723.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00155-of-00723.safetensors", + "model.layers.17.input_layernorm.weight": "model-00156-of-00723.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00157-of-00723.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00158-of-00723.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00159-of-00723.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00160-of-00723.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00161-of-00723.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00162-of-00723.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00163-of-00723.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00164-of-00723.safetensors", + "model.layers.18.input_layernorm.weight": "model-00165-of-00723.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00166-of-00723.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00167-of-00723.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00168-of-00723.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00169-of-00723.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00170-of-00723.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00171-of-00723.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00172-of-00723.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00173-of-00723.safetensors", + "model.layers.19.input_layernorm.weight": "model-00174-of-00723.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00175-of-00723.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00176-of-00723.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00177-of-00723.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00178-of-00723.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00179-of-00723.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00180-of-00723.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00181-of-00723.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00182-of-00723.safetensors", + "model.layers.20.input_layernorm.weight": "model-00183-of-00723.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00184-of-00723.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00185-of-00723.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00186-of-00723.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00187-of-00723.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00188-of-00723.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00189-of-00723.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00190-of-00723.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00191-of-00723.safetensors", + "model.layers.21.input_layernorm.weight": "model-00192-of-00723.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00193-of-00723.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00194-of-00723.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00195-of-00723.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00196-of-00723.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00197-of-00723.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00198-of-00723.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00199-of-00723.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00200-of-00723.safetensors", + "model.layers.22.input_layernorm.weight": "model-00201-of-00723.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00202-of-00723.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00203-of-00723.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00204-of-00723.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00205-of-00723.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00206-of-00723.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00207-of-00723.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00208-of-00723.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00209-of-00723.safetensors", + "model.layers.23.input_layernorm.weight": "model-00210-of-00723.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00211-of-00723.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00212-of-00723.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00213-of-00723.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00214-of-00723.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00215-of-00723.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00216-of-00723.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00217-of-00723.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00218-of-00723.safetensors", + "model.layers.24.input_layernorm.weight": "model-00219-of-00723.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00220-of-00723.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00221-of-00723.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00222-of-00723.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00223-of-00723.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00224-of-00723.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00225-of-00723.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00226-of-00723.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00227-of-00723.safetensors", + "model.layers.25.input_layernorm.weight": "model-00228-of-00723.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00229-of-00723.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00230-of-00723.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00231-of-00723.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00232-of-00723.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00233-of-00723.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00234-of-00723.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00235-of-00723.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00236-of-00723.safetensors", + "model.layers.26.input_layernorm.weight": "model-00237-of-00723.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00238-of-00723.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00239-of-00723.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00240-of-00723.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00241-of-00723.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00242-of-00723.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00243-of-00723.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00244-of-00723.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00245-of-00723.safetensors", + "model.layers.27.input_layernorm.weight": "model-00246-of-00723.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00247-of-00723.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00248-of-00723.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00249-of-00723.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00250-of-00723.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00251-of-00723.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00252-of-00723.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00253-of-00723.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00254-of-00723.safetensors", + "model.layers.28.input_layernorm.weight": "model-00255-of-00723.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00256-of-00723.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00257-of-00723.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00258-of-00723.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00259-of-00723.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00260-of-00723.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00261-of-00723.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00262-of-00723.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00263-of-00723.safetensors", + "model.layers.29.input_layernorm.weight": "model-00264-of-00723.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00265-of-00723.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00266-of-00723.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00267-of-00723.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00268-of-00723.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00269-of-00723.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00270-of-00723.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00271-of-00723.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00272-of-00723.safetensors", + "model.layers.30.input_layernorm.weight": "model-00273-of-00723.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00274-of-00723.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00275-of-00723.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00276-of-00723.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00277-of-00723.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00278-of-00723.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00279-of-00723.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00280-of-00723.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00281-of-00723.safetensors", + "model.layers.31.input_layernorm.weight": "model-00282-of-00723.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00283-of-00723.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00284-of-00723.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00285-of-00723.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00286-of-00723.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00287-of-00723.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00288-of-00723.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00289-of-00723.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00290-of-00723.safetensors", + "model.layers.32.input_layernorm.weight": "model-00291-of-00723.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00292-of-00723.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00293-of-00723.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00294-of-00723.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00295-of-00723.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00296-of-00723.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00297-of-00723.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00298-of-00723.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00299-of-00723.safetensors", + "model.layers.33.input_layernorm.weight": "model-00300-of-00723.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00301-of-00723.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00302-of-00723.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00303-of-00723.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00304-of-00723.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00305-of-00723.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00306-of-00723.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00307-of-00723.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00308-of-00723.safetensors", + "model.layers.34.input_layernorm.weight": "model-00309-of-00723.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00310-of-00723.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00311-of-00723.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00312-of-00723.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00313-of-00723.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00314-of-00723.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00315-of-00723.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00316-of-00723.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00317-of-00723.safetensors", + "model.layers.35.input_layernorm.weight": "model-00318-of-00723.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00319-of-00723.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00320-of-00723.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00321-of-00723.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00322-of-00723.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00323-of-00723.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00324-of-00723.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00325-of-00723.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00326-of-00723.safetensors", + "model.layers.36.input_layernorm.weight": "model-00327-of-00723.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00328-of-00723.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00329-of-00723.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00330-of-00723.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00331-of-00723.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00332-of-00723.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00333-of-00723.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00334-of-00723.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00335-of-00723.safetensors", + "model.layers.37.input_layernorm.weight": "model-00336-of-00723.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00337-of-00723.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00338-of-00723.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00339-of-00723.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00340-of-00723.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00341-of-00723.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00342-of-00723.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00343-of-00723.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00344-of-00723.safetensors", + "model.layers.38.input_layernorm.weight": "model-00345-of-00723.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00346-of-00723.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00347-of-00723.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00348-of-00723.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00349-of-00723.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00350-of-00723.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00351-of-00723.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00352-of-00723.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00353-of-00723.safetensors", + "model.layers.39.input_layernorm.weight": "model-00354-of-00723.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00355-of-00723.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00356-of-00723.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00357-of-00723.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00358-of-00723.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00359-of-00723.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00360-of-00723.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00361-of-00723.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00362-of-00723.safetensors", + "model.layers.40.input_layernorm.weight": "model-00363-of-00723.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00364-of-00723.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00365-of-00723.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00366-of-00723.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00367-of-00723.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00368-of-00723.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00369-of-00723.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00370-of-00723.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00371-of-00723.safetensors", + "model.layers.41.input_layernorm.weight": "model-00372-of-00723.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00373-of-00723.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00374-of-00723.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00375-of-00723.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00376-of-00723.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00377-of-00723.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00378-of-00723.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00379-of-00723.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00380-of-00723.safetensors", + "model.layers.42.input_layernorm.weight": "model-00381-of-00723.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00382-of-00723.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00383-of-00723.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00384-of-00723.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00385-of-00723.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00386-of-00723.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00387-of-00723.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00388-of-00723.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00389-of-00723.safetensors", + "model.layers.43.input_layernorm.weight": "model-00390-of-00723.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00391-of-00723.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00392-of-00723.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00393-of-00723.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00394-of-00723.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00395-of-00723.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00396-of-00723.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00397-of-00723.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00398-of-00723.safetensors", + "model.layers.44.input_layernorm.weight": "model-00399-of-00723.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00400-of-00723.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00401-of-00723.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00402-of-00723.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00403-of-00723.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00404-of-00723.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00405-of-00723.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00406-of-00723.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00407-of-00723.safetensors", + "model.layers.45.input_layernorm.weight": "model-00408-of-00723.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00409-of-00723.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00410-of-00723.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00411-of-00723.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00412-of-00723.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00413-of-00723.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00414-of-00723.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00415-of-00723.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00416-of-00723.safetensors", + "model.layers.46.input_layernorm.weight": "model-00417-of-00723.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00418-of-00723.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00419-of-00723.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00420-of-00723.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00421-of-00723.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00422-of-00723.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00423-of-00723.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00424-of-00723.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00425-of-00723.safetensors", + "model.layers.47.input_layernorm.weight": "model-00426-of-00723.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00427-of-00723.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00428-of-00723.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00429-of-00723.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00430-of-00723.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00431-of-00723.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00432-of-00723.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00433-of-00723.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00434-of-00723.safetensors", + "model.layers.48.input_layernorm.weight": "model-00435-of-00723.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00436-of-00723.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00437-of-00723.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00438-of-00723.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00439-of-00723.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00440-of-00723.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00441-of-00723.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00442-of-00723.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00443-of-00723.safetensors", + "model.layers.49.input_layernorm.weight": "model-00444-of-00723.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00445-of-00723.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00446-of-00723.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00447-of-00723.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00448-of-00723.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00449-of-00723.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00450-of-00723.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00451-of-00723.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00452-of-00723.safetensors", + "model.layers.50.input_layernorm.weight": "model-00453-of-00723.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00454-of-00723.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00455-of-00723.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00456-of-00723.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00457-of-00723.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00458-of-00723.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00459-of-00723.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00460-of-00723.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00461-of-00723.safetensors", + "model.layers.51.input_layernorm.weight": "model-00462-of-00723.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00463-of-00723.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00464-of-00723.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00465-of-00723.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00466-of-00723.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00467-of-00723.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00468-of-00723.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00469-of-00723.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00470-of-00723.safetensors", + "model.layers.52.input_layernorm.weight": "model-00471-of-00723.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00472-of-00723.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00473-of-00723.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00474-of-00723.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00475-of-00723.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00476-of-00723.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00477-of-00723.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00478-of-00723.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00479-of-00723.safetensors", + "model.layers.53.input_layernorm.weight": "model-00480-of-00723.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00481-of-00723.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00482-of-00723.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00483-of-00723.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00484-of-00723.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00485-of-00723.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00486-of-00723.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00487-of-00723.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00488-of-00723.safetensors", + "model.layers.54.input_layernorm.weight": "model-00489-of-00723.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00490-of-00723.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00491-of-00723.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00492-of-00723.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00493-of-00723.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00494-of-00723.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00495-of-00723.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00496-of-00723.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00497-of-00723.safetensors", + "model.layers.55.input_layernorm.weight": "model-00498-of-00723.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00499-of-00723.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00500-of-00723.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00501-of-00723.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00502-of-00723.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00503-of-00723.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00504-of-00723.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00505-of-00723.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00506-of-00723.safetensors", + "model.layers.56.input_layernorm.weight": "model-00507-of-00723.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00508-of-00723.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00509-of-00723.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00510-of-00723.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00511-of-00723.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00512-of-00723.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00513-of-00723.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00514-of-00723.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00515-of-00723.safetensors", + "model.layers.57.input_layernorm.weight": "model-00516-of-00723.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00517-of-00723.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00518-of-00723.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00519-of-00723.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00520-of-00723.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00521-of-00723.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00522-of-00723.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00523-of-00723.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00524-of-00723.safetensors", + "model.layers.58.input_layernorm.weight": "model-00525-of-00723.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00526-of-00723.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00527-of-00723.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00528-of-00723.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00529-of-00723.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00530-of-00723.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00531-of-00723.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00532-of-00723.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00533-of-00723.safetensors", + "model.layers.59.input_layernorm.weight": "model-00534-of-00723.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00535-of-00723.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00536-of-00723.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00537-of-00723.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00538-of-00723.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00539-of-00723.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00540-of-00723.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00541-of-00723.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00542-of-00723.safetensors", + "model.layers.60.input_layernorm.weight": "model-00543-of-00723.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00544-of-00723.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00545-of-00723.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00546-of-00723.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00547-of-00723.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00548-of-00723.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00549-of-00723.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00550-of-00723.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00551-of-00723.safetensors", + "model.layers.61.input_layernorm.weight": "model-00552-of-00723.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00553-of-00723.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00554-of-00723.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00555-of-00723.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00556-of-00723.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00557-of-00723.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00558-of-00723.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00559-of-00723.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00560-of-00723.safetensors", + "model.layers.62.input_layernorm.weight": "model-00561-of-00723.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00562-of-00723.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00563-of-00723.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00564-of-00723.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00565-of-00723.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00566-of-00723.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00567-of-00723.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00568-of-00723.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00569-of-00723.safetensors", + "model.layers.63.input_layernorm.weight": "model-00570-of-00723.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00571-of-00723.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00572-of-00723.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00573-of-00723.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00574-of-00723.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00575-of-00723.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00576-of-00723.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00577-of-00723.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00578-of-00723.safetensors", + "model.layers.64.input_layernorm.weight": "model-00579-of-00723.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00580-of-00723.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00581-of-00723.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00582-of-00723.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00583-of-00723.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00584-of-00723.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00585-of-00723.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00586-of-00723.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00587-of-00723.safetensors", + "model.layers.65.input_layernorm.weight": "model-00588-of-00723.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00589-of-00723.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00590-of-00723.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00591-of-00723.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00592-of-00723.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00593-of-00723.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00594-of-00723.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00595-of-00723.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00596-of-00723.safetensors", + "model.layers.66.input_layernorm.weight": "model-00597-of-00723.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00598-of-00723.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00599-of-00723.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00600-of-00723.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00601-of-00723.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00602-of-00723.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00603-of-00723.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00604-of-00723.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00605-of-00723.safetensors", + "model.layers.67.input_layernorm.weight": "model-00606-of-00723.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00607-of-00723.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00608-of-00723.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00609-of-00723.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00610-of-00723.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00611-of-00723.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00612-of-00723.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00613-of-00723.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00614-of-00723.safetensors", + "model.layers.68.input_layernorm.weight": "model-00615-of-00723.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00616-of-00723.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00617-of-00723.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00618-of-00723.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00619-of-00723.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00620-of-00723.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00621-of-00723.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00622-of-00723.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00623-of-00723.safetensors", + "model.layers.69.input_layernorm.weight": "model-00624-of-00723.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00625-of-00723.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00626-of-00723.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00627-of-00723.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00628-of-00723.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00629-of-00723.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00630-of-00723.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00631-of-00723.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00632-of-00723.safetensors", + "model.layers.70.input_layernorm.weight": "model-00633-of-00723.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00634-of-00723.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00635-of-00723.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00636-of-00723.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00637-of-00723.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00638-of-00723.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00639-of-00723.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00640-of-00723.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00641-of-00723.safetensors", + "model.layers.71.input_layernorm.weight": "model-00642-of-00723.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00643-of-00723.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00644-of-00723.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00645-of-00723.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00646-of-00723.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00647-of-00723.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00648-of-00723.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00649-of-00723.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00650-of-00723.safetensors", + "model.layers.72.input_layernorm.weight": "model-00651-of-00723.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00652-of-00723.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00653-of-00723.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00654-of-00723.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00655-of-00723.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00656-of-00723.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00657-of-00723.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00658-of-00723.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00659-of-00723.safetensors", + "model.layers.73.input_layernorm.weight": "model-00660-of-00723.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00661-of-00723.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00662-of-00723.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00663-of-00723.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00664-of-00723.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00665-of-00723.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00666-of-00723.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00667-of-00723.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00668-of-00723.safetensors", + "model.layers.74.input_layernorm.weight": "model-00669-of-00723.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00670-of-00723.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00671-of-00723.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00672-of-00723.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00673-of-00723.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00674-of-00723.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00675-of-00723.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00676-of-00723.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00677-of-00723.safetensors", + "model.layers.75.input_layernorm.weight": "model-00678-of-00723.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00679-of-00723.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00680-of-00723.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00681-of-00723.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00682-of-00723.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00683-of-00723.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00684-of-00723.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00685-of-00723.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00686-of-00723.safetensors", + "model.layers.76.input_layernorm.weight": "model-00687-of-00723.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00688-of-00723.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00689-of-00723.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00690-of-00723.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00691-of-00723.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00692-of-00723.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00693-of-00723.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00694-of-00723.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00695-of-00723.safetensors", + "model.layers.77.input_layernorm.weight": "model-00696-of-00723.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00697-of-00723.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00698-of-00723.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00699-of-00723.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00700-of-00723.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00701-of-00723.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00702-of-00723.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00703-of-00723.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00704-of-00723.safetensors", + "model.layers.78.input_layernorm.weight": "model-00705-of-00723.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00706-of-00723.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00707-of-00723.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00708-of-00723.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00709-of-00723.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00710-of-00723.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00711-of-00723.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00712-of-00723.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00713-of-00723.safetensors", + "model.layers.79.input_layernorm.weight": "model-00714-of-00723.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00715-of-00723.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00716-of-00723.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00717-of-00723.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00718-of-00723.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00719-of-00723.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00720-of-00723.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00721-of-00723.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00722-of-00723.safetensors", + "model.norm.weight": "model-00723-of-00723.safetensors" + } } \ No newline at end of file